Search results for: data streams
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24617

Search results for: data streams

24317 The Utilization of Big Data in Knowledge Management Creation

Authors: Daniel Brian Thompson, Subarmaniam Kannan

Abstract:

The huge weightage of knowledge in this world and within the repository of organizations has already reached immense capacity and is constantly increasing as time goes by. To accommodate these constraints, Big Data implementation and algorithms are utilized to obtain new or enhanced knowledge for decision-making. With the transition from data to knowledge provides the transformational changes which will provide tangible benefits to the individual implementing these practices. Today, various organization would derive knowledge from observations and intuitions where this information or data will be translated into best practices for knowledge acquisition, generation and sharing. Through the widespread usage of Big Data, the main intention is to provide information that has been cleaned and analyzed to nurture tangible insights for an organization to apply to their knowledge-creation practices based on facts and figures. The translation of data into knowledge will generate value for an organization to make decisive decisions to proceed with the transition of best practices. Without a strong foundation of knowledge and Big Data, businesses are not able to grow and be enhanced within the competitive environment.

Keywords: big data, knowledge management, data driven, knowledge creation

Procedia PDF Downloads 93
24316 Adsorption of Cd(II) and Pb(II) from Aqueous Solutions by Using Pods of Acacia Karoo

Authors: Gulshan Kumar Jawa, Sandeep Mohan Ahuja

Abstract:

With the increase in industrialization, the presence of heavy metals in wastewater streams has turned into a serious concern for the ecosystem. The metals diffuse through the food chains, causing various health hazards. Conventional methods used to remove these heavy metals from water have some limitations, such as cost, secondary pollution due to sludge formation, recovery of metal, economic viability at low metal concentrations, etc. Many of the biomaterials have been investigated by researchers for the adsorption of heavy metals from water solutions as an alternative technique for the last two decades and have found promising results. In this paper, the batch study on the use of pods of acacia karoo for the adsorption of Cd(II) and Pb(II) from aqueous solutions has been reported. The effect of various parameters on the removal of metal ions, such as pH, contact time, stirring speed, initial metal ion concentration, adsorbent dose, and temperature, have been established to find the optimum parameters through one parameter optimization. Further, kinetic, equilibrium, and thermodynamic studies have been conducted. The pods of acacia karoo have shown great potential for adsorption of Cd(II) and Pb(II) from aqueous solutions and have proven to be a better and more economical alternative for the purpose.

Keywords: adsorption, heavy metals, biomaterials, Cadmium(II), Lead(II), pods of acacia karoo

Procedia PDF Downloads 20
24315 Survey on Data Security Issues Through Cloud Computing Amongst Sme’s in Nairobi County, Kenya

Authors: Masese Chuma Benard, Martin Onsiro Ronald

Abstract:

Businesses have been using cloud computing more frequently recently because they wish to take advantage of its advantages. However, employing cloud computing also introduces new security concerns, particularly with regard to data security, potential risks and weaknesses that could be exploited by attackers, and various tactics and strategies that could be used to lessen these risks. This study examines data security issues on cloud computing amongst sme’s in Nairobi county, Kenya. The study used the sample size of 48, the research approach was mixed methods, The findings show that data owner has no control over the cloud merchant's data management procedures, there is no way to ensure that data is handled legally. This implies that you will lose control over the data stored in the cloud. Data and information stored in the cloud may face a range of availability issues due to internet outages; this can represent a significant risk to data kept in shared clouds. Integrity, availability, and secrecy are all mentioned.

Keywords: data security, cloud computing, information, information security, small and medium-sized firms (SMEs)

Procedia PDF Downloads 65
24314 Cloud Design for Storing Large Amount of Data

Authors: M. Strémy, P. Závacký, P. Cuninka, M. Juhás

Abstract:

Main goal of this paper is to introduce our design of private cloud for storing large amount of data, especially pictures, and to provide good technological backend for data analysis based on parallel processing and business intelligence. We have tested hypervisors, cloud management tools, storage for storing all data and Hadoop to provide data analysis on unstructured data. Providing high availability, virtual network management, logical separation of projects and also rapid deployment of physical servers to our environment was also needed.

Keywords: cloud, glusterfs, hadoop, juju, kvm, maas, openstack, virtualization

Procedia PDF Downloads 339
24313 Estimation of Missing Values in Aggregate Level Spatial Data

Authors: Amitha Puranik, V. S. Binu, Seena Biju

Abstract:

Missing data is a common problem in spatial analysis especially at the aggregate level. Missing can either occur in covariate or in response variable or in both in a given location. Many missing data techniques are available to estimate the missing data values but not all of these methods can be applied on spatial data since the data are autocorrelated. Hence there is a need to develop a method that estimates the missing values in both response variable and covariates in spatial data by taking account of the spatial autocorrelation. The present study aims to develop a model to estimate the missing data points at the aggregate level in spatial data by accounting for (a) Spatial autocorrelation of the response variable (b) Spatial autocorrelation of covariates and (c) Correlation between covariates and the response variable. Estimating the missing values of spatial data requires a model that explicitly account for the spatial autocorrelation. The proposed model not only accounts for spatial autocorrelation but also utilizes the correlation that exists between covariates, within covariates and between a response variable and covariates. The precise estimation of the missing data points in spatial data will result in an increased precision of the estimated effects of independent variables on the response variable in spatial regression analysis.

Keywords: spatial regression, missing data estimation, spatial autocorrelation, simulation analysis

Procedia PDF Downloads 360
24312 Waste Bone Based Catalyst: Characterization and Esterification Application

Authors: Amit Keshav

Abstract:

Waste bone, produced in large quantity (8-10 kg./day) from a slaughterhouse, could be a cheap (cost $0.20 per kg) substitute for commercial catalysts. In the present work, catalyst for esterification reaction was prepared from waste bone and characterized by various techniques. Bone was deoiled and then sulfonated. Fourier-transform infrared spectroscopy (FTIR) spectra of prepared catalyst predicted –OH vibration at 3416 and 1630 cm⁻¹, S-O stretching at 1124 cm⁻¹ and intense bands of hydroxypatite in a region between 500 and 700 cm⁻¹. X-ray diffraction (XRD) predicts peaks of hydroxyapatite, CaO, and tricalcium phosphate. Scanning electron microscope (SEM) was employed to reveal the presence of non-uniformity deposited fine particles on the catalyst surface that represents active acidic sites. The prepared catalyst was employed to study its performance on esterification reaction between acrylic acid and ethanol in a molar ratio of 1:1 at a set temperature of 60 °C. Results show an equilibrium conversion of 49% which is matched to the commercial catalysts employed in literature. Thus waste bone could be a good catalyst for acrylic acid removal from waste industrial streams via the process of esterification.Keywords— Heterogeneous catalyst, characterization, esterification, equilibrium conversion

Keywords: heterogeneous catalyst, characterization, esterification, equilibrium conversion

Procedia PDF Downloads 129
24311 Association Rules Mining and NOSQL Oriented Document in Big Data

Authors: Sarra Senhadji, Imene Benzeguimi, Zohra Yagoub

Abstract:

Big Data represents the recent technology of manipulating voluminous and unstructured data sets over multiple sources. Therefore, NOSQL appears to handle the problem of unstructured data. Association rules mining is one of the popular techniques of data mining to extract hidden relationship from transactional databases. The algorithm for finding association dependencies is well-solved with Map Reduce. The goal of our work is to reduce the time of generating of frequent itemsets by using Map Reduce and NOSQL database oriented document. A comparative study is given to evaluate the performances of our algorithm with the classical algorithm Apriori.

Keywords: Apriori, Association rules mining, Big Data, Data Mining, Hadoop, MapReduce, MongoDB, NoSQL

Procedia PDF Downloads 143
24310 Immunization-Data-Quality in Public Health Facilities in the Pastoralist Communities: A Comparative Study Evidence from Afar and Somali Regional States, Ethiopia

Authors: Melaku Tsehay

Abstract:

The Consortium of Christian Relief and Development Associations (CCRDA), and the CORE Group Polio Partners (CGPP) Secretariat have been working with Global Alliance for Vac-cines and Immunization (GAVI) to improve the immunization data quality in Afar and Somali Regional States. The main aim of this study was to compare the quality of immunization data before and after the above interventions in health facilities in the pastoralist communities in Ethiopia. To this end, a comparative-cross-sectional study was conducted on 51 health facilities. The baseline data was collected in May 2019, while the end line data in August 2021. The WHO data quality self-assessment tool (DQS) was used to collect data. A significant improvment was seen in the accuracy of the pentavalent vaccine (PT)1 (p = 0.012) data at the health posts (HP), while PT3 (p = 0.010), and Measles (p = 0.020) at the health centers (HC). Besides, a highly sig-nificant improvment was observed in the accuracy of tetanus toxoid (TT)2 data at HP (p < 0.001). The level of over- or under-reporting was found to be < 8%, at the HP, and < 10% at the HC for PT3. The data completeness was also increased from 72.09% to 88.89% at the HC. Nearly 74% of the health facilities timely reported their respective immunization data, which is much better than the baseline (7.1%) (p < 0.001). These findings may provide some hints for the policies and pro-grams targetting on improving immunization data qaulity in the pastoralist communities.

Keywords: data quality, immunization, verification factor, pastoralist region

Procedia PDF Downloads 78
24309 The English Classroom: Scope and Space for Motivation

Authors: Madhavi Godavarthy

Abstract:

The globalized world has been witnessing the ubiquity of the English language and has made it mandatory that students be equipped with the required Communication and soft skills. For students and especially for students studying in technical streams, gaining command over the English language is only a part of the bigger challenges they will face in the future. Linguistic capabilities if blended with the right attitude and a positive personality would deliver better results in the present environment of the digitalized world. An English classroom has that ‘space’; a space if utilized well by the teacher can pay rich dividends. The prescribed syllabus for English in the process of adapting itself to the challenges of a more and more technical world has meted out an indifferent treatment in including ‘literary’ material in their curriculum. A debate has always existed regarding the same and diversified opinions have been given. When the student is motivated to reach Literature through intrinsic motivation, it may contribute to his/her personality-development. In the present paper, the element of focus is on the scope and space to motivate students by creating a specific space for herself/himself amidst the schedules of the teaching-learning processes by taking into consideration a few literary excerpts for the purpose.

Keywords: English language, teaching and learning process, reader response theory, intrinsic motivation, literary texts

Procedia PDF Downloads 593
24308 Augmenting Cultural Heritage Through 4.0 Technologies: A Research on the Archival Jewelry of the Gianfranco Ferré Research Center

Authors: Greta Rizzi, Ashley Gallitto, Federica Vacca

Abstract:

Looking at design artifacts as bearers and disseminators of material knowledge and intangible socio-cultural meanings, the significance of archival jewelry was investigated following digital cultural heritage research streams. The application of the reverse engineering concept guided the research path: starting with the study of Gianfranco Ferré's archival jewelry and analyzing its technical heritage and symbolic value, the digitalization, dematerialization, and rematerialization of the artifact were carried out. According to that, the proposed paper results from research conducted within the residency program between the Gianfranco Ferré Research Center (GFRC) and Massachusetts Institute of Technology (MIT), involving both the Design and Mechanical Engineering Departments of Politecnico di Milano. The paper will discuss the analysis of traditional design manufacturing techniques, re-imagined through 3D scanning, 3D modeling, and 3D printing technical knowledge while emphasizing the significance of the designer's role as an explorer of socio-cultural meanings and technological mediators in the analog-digital-analog transition.

Keywords: Archival jewelry, cultural heritage, rematerialization, reverse engineering.

Procedia PDF Downloads 30
24307 Identifying Critical Success Factors for Data Quality Management through a Delphi Study

Authors: Maria Paula Santos, Ana Lucas

Abstract:

Organizations support their operations and decision making on the data they have at their disposal, so the quality of these data is remarkably important and Data Quality (DQ) is currently a relevant issue, the literature being unanimous in pointing out that poor DQ can result in large costs for organizations. The literature review identified and described 24 Critical Success Factors (CSF) for Data Quality Management (DQM) that were presented to a panel of experts, who ordered them according to their degree of importance, using the Delphi method with the Q-sort technique, based on an online questionnaire. The study shows that the five most important CSF for DQM are: definition of appropriate policies and standards, control of inputs, definition of a strategic plan for DQ, organizational culture focused on quality of the data and obtaining top management commitment and support.

Keywords: critical success factors, data quality, data quality management, Delphi, Q-Sort

Procedia PDF Downloads 197
24306 Application of Random Forest Model in The Prediction of River Water Quality

Authors: Turuganti Venkateswarlu, Jagadeesh Anmala

Abstract:

Excessive runoffs from various non-point source land uses, and other point sources are rapidly contaminating the water quality of streams in the Upper Green River watershed, Kentucky, USA. It is essential to maintain the stream water quality as the river basin is one of the major freshwater sources in this province. It is also important to understand the water quality parameters (WQPs) quantitatively and qualitatively along with their important features as stream water is sensitive to climatic events and land-use practices. In this paper, a model was developed for predicting one of the significant WQPs, Fecal Coliform (FC) from precipitation, temperature, urban land use factor (ULUF), agricultural land use factor (ALUF), and forest land-use factor (FLUF) using Random Forest (RF) algorithm. The RF model, a novel ensemble learning algorithm, can even find out advanced feature importance characteristics from the given model inputs for different combinations. This model’s outcomes showed a good correlation between FC and climate events and land use factors (R2 = 0.94) and precipitation and temperature are the primary influencing factors for FC.

Keywords: water quality, land use factors, random forest, fecal coliform

Procedia PDF Downloads 180
24305 Experimental Study of Local Scour Downstream of Cylindrical Bridge Piers

Authors: Mohammed Traeq Shukri

Abstract:

Scour is a natural phenomenon caused by the erosive action of flowing stream on alluvial beds, which removes the sediment around or near structures located in flowing water. It means the lowering of the riverbed level by water erosions such that there is a tendency to expose the foundations of a structure. It is the result of the erosive action of flowing water, excavating and carrying away material from the bed and banks of streams and from around the piers of bridges. The failure of bridges due to excessive local scour during floods poses a challenging problem to hydraulic engineers. The failure of bridges piers is due to many reasons such as localized scour combined with general riverbed degradation. In this paper, we try to estimate the temporal variation of scour depth at non-uniform cylindrical bridge pier, by experimental work in civil engineering hydraulic laboratories of Gaziantep University on a channel have dimensions of 8.3m length, 0.8m width and 0.9m depth. The experiments will be carried on 20 cm depth of sediment layer having d50=0.4 mm. Three bridge pier shapes having different scaled models will be constructed in a 1.5m of test section in the channel.

Keywords: scour, local scour, bridge piers, scour depth, vortex, horseshoe vortex

Procedia PDF Downloads 152
24304 Data Mining in Medicine Domain Using Decision Trees and Vector Support Machine

Authors: Djamila Benhaddouche, Abdelkader Benyettou

Abstract:

In this paper, we used data mining to extract biomedical knowledge. In general, complex biomedical data collected in studies of populations are treated by statistical methods, although they are robust, they are not sufficient in themselves to harness the potential wealth of data. For that you used in step two learning algorithms: the Decision Trees and Support Vector Machine (SVM). These supervised classification methods are used to make the diagnosis of thyroid disease. In this context, we propose to promote the study and use of symbolic data mining techniques.

Keywords: biomedical data, learning, classifier, algorithms decision tree, knowledge extraction

Procedia PDF Downloads 532
24303 Analysis of Different Classification Techniques Using WEKA for Diabetic Disease

Authors: Usama Ahmed

Abstract:

Data mining is the process of analyze data which are used to predict helpful information. It is the field of research which solve various type of problem. In data mining, classification is an important technique to classify different kind of data. Diabetes is most common disease. This paper implements different classification technique using Waikato Environment for Knowledge Analysis (WEKA) on diabetes dataset and find which algorithm is suitable for working. The best classification algorithm based on diabetic data is Naïve Bayes. The accuracy of Naïve Bayes is 76.31% and take 0.06 seconds to build the model.

Keywords: data mining, classification, diabetes, WEKA

Procedia PDF Downloads 132
24302 Comprehensive Study of Data Science

Authors: Asifa Amara, Prachi Singh, Kanishka, Debargho Pathak, Akshat Kumar, Jayakumar Eravelly

Abstract:

Today's generation is totally dependent on technology that uses data as its fuel. The present study is all about innovations and developments in data science and gives an idea about how efficiently to use the data provided. This study will help to understand the core concepts of data science. The concept of artificial intelligence was introduced by Alan Turing in which the main principle was to create an artificial system that can run independently of human-given programs and can function with the help of analyzing data to understand the requirements of the users. Data science comprises business understanding, analyzing data, ethical concerns, understanding programming languages, various fields and sources of data, skills, etc. The usage of data science has evolved over the years. In this review article, we have covered a part of data science, i.e., machine learning. Machine learning uses data science for its work. Machines learn through their experience, which helps them to do any work more efficiently. This article includes a comparative study image between human understanding and machine understanding, advantages, applications, and real-time examples of machine learning. Data science is an important game changer in the life of human beings. Since the advent of data science, we have found its benefits and how it leads to a better understanding of people, and how it cherishes individual needs. It has improved business strategies, services provided by them, forecasting, the ability to attend sustainable developments, etc. This study also focuses on a better understanding of data science which will help us to create a better world.

Keywords: data science, machine learning, data analytics, artificial intelligence

Procedia PDF Downloads 59
24301 Application of Artificial Neural Network Technique for Diagnosing Asthma

Authors: Azadeh Bashiri

Abstract:

Introduction: Lack of proper diagnosis and inadequate treatment of asthma leads to physical and financial complications. This study aimed to use data mining techniques and creating a neural network intelligent system for diagnosis of asthma. Methods: The study population is the patients who had visited one of the Lung Clinics in Tehran. Data were analyzed using the SPSS statistical tool and the chi-square Pearson's coefficient was the basis of decision making for data ranking. The considered neural network is trained using back propagation learning technique. Results: According to the analysis performed by means of SPSS to select the top factors, 13 effective factors were selected, in different performances, data was mixed in various forms, so the different models were made for training the data and testing networks and in all different modes, the network was able to predict correctly 100% of all cases. Conclusion: Using data mining methods before the design structure of system, aimed to reduce the data dimension and the optimum choice of the data, will lead to a more accurate system. Therefore, considering the data mining approaches due to the nature of medical data is necessary.

Keywords: asthma, data mining, Artificial Neural Network, intelligent system

Procedia PDF Downloads 256
24300 Interpreting Privacy Harms from a Non-Economic Perspective

Authors: Christopher Muhawe, Masooda Bashir

Abstract:

With increased Internet Communication Technology(ICT), the virtual world has become the new normal. At the same time, there is an unprecedented collection of massive amounts of data by both private and public entities. Unfortunately, this increase in data collection has been in tandem with an increase in data misuse and data breach. Regrettably, the majority of data breach and data misuse claims have been unsuccessful in the United States courts for the failure of proof of direct injury to physical or economic interests. The requirement to express data privacy harms from an economic or physical stance negates the fact that not all data harms are physical or economic in nature. The challenge is compounded by the fact that data breach harms and risks do not attach immediately. This research will use a descriptive and normative approach to show that not all data harms can be expressed in economic or physical terms. Expressing privacy harms purely from an economic or physical harm perspective negates the fact that data insecurity may result into harms which run counter the functions of privacy in our lives. The promotion of liberty, selfhood, autonomy, promotion of human social relations and the furtherance of the existence of a free society. There is no economic value that can be placed on these functions of privacy. The proposed approach addresses data harms from a psychological and social perspective.

Keywords: data breach and misuse, economic harms, privacy harms, psychological harms

Procedia PDF Downloads 174
24299 Technologies for Phosphorus Removal from Wastewater: Review

Authors: Thandie Veronicah Sima, Moatlhodi Wiseman Letshwenyo

Abstract:

Discharge of wastewater is one of the major sources of phosphorus entering streams, lakes and other water bodies causing undesired environmental problem such as eutrophication. This condition not only puts the ecosystem at risk but also causes severe economic damages. Stringent laws have been developed globally by different bodies to control the level of phosphorus concentrations into receiving environments. In order to satisfy the constraints, a high degree of tertiary treatment or at least a significant reduction of phosphorus concentration is obligatory. This comprehensive review summarizes phosphorus removal technologies, from the most commonly used conventional technologies such as chemical precipitation through metal addition, membrane filtration, reverse osmosis and enhanced biological phosphorus removal using activated sludge system to passive systems such as constructed wetlands and filtration systems. Trends, perspectives and scientific procedures conducted by different researchers have been presented. This review critically evaluates the advantages and limitations behind each of the technologies. Enhancement of passive systems using reactive media such as industrial wastes to provide additional uptake through adsorption or precipitation is also discussed in this article.

Keywords: adsorption, chemical precipitation, enhanced biological phosphorus removal, phosphorus removal

Procedia PDF Downloads 303
24298 Machine Learning Analysis of Student Success in Introductory Calculus Based Physics I Course

Authors: Chandra Prayaga, Aaron Wade, Lakshmi Prayaga, Gopi Shankar Mallu

Abstract:

This paper presents the use of machine learning algorithms to predict the success of students in an introductory physics course. Data having 140 rows pertaining to the performance of two batches of students was used. The lack of sufficient data to train robust machine learning models was compensated for by generating synthetic data similar to the real data. CTGAN and CTGAN with Gaussian Copula (Gaussian) were used to generate synthetic data, with the real data as input. To check the similarity between the real data and each synthetic dataset, pair plots were made. The synthetic data was used to train machine learning models using the PyCaret package. For the CTGAN data, the Ada Boost Classifier (ADA) was found to be the ML model with the best fit, whereas the CTGAN with Gaussian Copula yielded Logistic Regression (LR) as the best model. Both models were then tested for accuracy with the real data. ROC-AUC analysis was performed for all the ten classes of the target variable (Grades A, A-, B+, B, B-, C+, C, C-, D, F). The ADA model with CTGAN data showed a mean AUC score of 0.4377, but the LR model with the Gaussian data showed a mean AUC score of 0.6149. ROC-AUC plots were obtained for each Grade value separately. The LR model with Gaussian data showed consistently better AUC scores compared to the ADA model with CTGAN data, except in two cases of the Grade value, C- and A-.

Keywords: machine learning, student success, physics course, grades, synthetic data, CTGAN, gaussian copula CTGAN

Procedia PDF Downloads 28
24297 Data Access, AI Intensity, and Scale Advantages

Authors: Chuping Lo

Abstract:

This paper presents a simple model demonstrating that ceteris paribus countries with lower barriers to accessing global data tend to earn higher incomes than other countries. Therefore, large countries that inherently have greater data resources tend to have higher incomes than smaller countries, such that the former may be more hesitant than the latter to liberalize cross-border data flows to maintain this advantage. Furthermore, countries with higher artificial intelligence (AI) intensity in production technologies tend to benefit more from economies of scale in data aggregation, leading to higher income and more trade as they are better able to utilize global data.

Keywords: digital intensity, digital divide, international trade, scale of economics

Procedia PDF Downloads 48
24296 Secured Transmission and Reserving Space in Images Before Encryption to Embed Data

Authors: G. R. Navaneesh, E. Nagarajan, C. H. Rajam Raju

Abstract:

Nowadays the multimedia data are used to store some secure information. All previous methods allocate a space in image for data embedding purpose after encryption. In this paper, we propose a novel method by reserving space in image with a boundary surrounded before encryption with a traditional RDH algorithm, which makes it easy for the data hider to reversibly embed data in the encrypted images. The proposed method can achieve real time performance, that is, data extraction and image recovery are free of any error. A secure transmission process is also discussed in this paper, which improves the efficiency by ten times compared to other processes as discussed.

Keywords: secure communication, reserving room before encryption, least significant bits, image encryption, reversible data hiding

Procedia PDF Downloads 394
24295 Identity Verification Using k-NN Classifiers and Autistic Genetic Data

Authors: Fuad M. Alkoot

Abstract:

DNA data have been used in forensics for decades. However, current research looks at using the DNA as a biometric identity verification modality. The goal is to improve the speed of identification. We aim at using gene data that was initially used for autism detection to find if and how accurate is this data for identification applications. Mainly our goal is to find if our data preprocessing technique yields data useful as a biometric identification tool. We experiment with using the nearest neighbor classifier to identify subjects. Results show that optimal classification rate is achieved when the test set is corrupted by normally distributed noise with zero mean and standard deviation of 1. The classification rate is close to optimal at higher noise standard deviation reaching 3. This shows that the data can be used for identity verification with high accuracy using a simple classifier such as the k-nearest neighbor (k-NN). 

Keywords: biometrics, genetic data, identity verification, k nearest neighbor

Procedia PDF Downloads 234
24294 Heavy Metal Removal by Green Microalgae Biofilms from Industrial Wastewater

Authors: B. N. Makhanya, S. F. Ndulini, M. S. Mthembu

Abstract:

Heavy metals are hazardous pollutants present in both industrial and domestic wastewater. They are usually disposed directly into natural streams, and when left untreated, they are a major cause of natural degradation and diseases. This study aimed to determine the ability of microalgae to remove heavy metals from coal mine wastewater. The green algae were grown and used for heavy metal removal in a laboratory bench. The physicochemical parameters and heavy metal removal were determined at 24 hours intervals for 5 days. The highest removal efficiencies were found to be 85%, 95%, and 99%, for Fe, Zn, and Cd, respectively. Copper and aluminium both had 100%. The results also indicated that the correlation between physicochemical parameters and all heavy metals were ranging from (0.50 ≤ r ≤ 0.85) for temperature, which indicated moderate positive to a strong positive correlation, pH had a very weak negative to a very weak positive correlation (-0.27 ≤ r ≤ 0.11), and chemical oxygen demand had a fair positive to a very strong positive correlation (0.69 ≤ r ≤ 0.98). The paired t-test indicated the removal of heavy metals to be statistically significant (0.007 ≥ p ≥ 0.000). Therefore, results showed that the microalgae used in the study were capable of removing heavy metals from industrial wastewater using possible mechanisms such as binding and absorption. Compared to the currently used technology for wastewater treatment, the microalgae may be the alternative to industrial wastewater treatment.

Keywords: heavy metals, industrial wastewater, microalgae, physiochemical parameters

Procedia PDF Downloads 119
24293 Modelling of Recovery and Application of Low-Grade Thermal Resources in the Mining and Mineral Processing Industry

Authors: S. McLean, J. A. Scott

Abstract:

The research topic is focusing on improving sustainable operation through recovery and reuse of waste heat in process water streams, an area in the mining industry that is often overlooked. There are significant advantages to the application of this topic, including economic and environmental benefits. The smelting process in the mining industry presents an opportunity to recover waste heat and apply it to alternative uses, thereby enhancing the overall process. This applied research has been conducted at the Sudbury Integrated Nickel Operations smelter site, in particular on the water cooling towers. The aim was to determine and optimize methods for appropriate recovery and subsequent upgrading of thermally low-grade heat lost from the water cooling towers in a manner that makes it useful for repurposing in applications, such as within an acid plant. This would be valuable to mining companies as it would be an opportunity to reduce the cost of the process, as well as decrease environmental impact and primary fuel usage. The waste heat from the cooling towers needs to be upgraded before it can be beneficially applied, as lower temperatures result in a decrease of the number of potential applications. Temperature and flow rate data were collected from the water cooling towers at an acid plant over two years. The research includes process control strategies and the development of a model capable of determining if the proposed heat recovery technique is economically viable, as well as assessing any environmental impact with the reduction in net energy consumption by the process. Therefore, comprehensive cost and impact analyses are carried out to determine the best area of application for the recovered waste heat. This method will allow engineers to easily identify the value of thermal resources available to them and determine if a full feasibility study should be carried out. The rapid scoping model developed will be applicable to any site that generates large amounts of waste heat. Results show that heat pumps are an economically viable solution for this application, allowing for reduced cost and CO₂ emissions.

Keywords: environment, heat recovery, mining engineering, sustainability

Procedia PDF Downloads 96
24292 A Review on Intelligent Systems for Geoscience

Authors: R Palson Kennedy, P.Kiran Sai

Abstract:

This article introduces machine learning (ML) researchers to the hurdles that geoscience problems present, as well as the opportunities for improvement in both ML and geosciences. This article presents a review from the data life cycle perspective to meet that need. Numerous facets of geosciences present unique difficulties for the study of intelligent systems. Geosciences data is notoriously difficult to analyze since it is frequently unpredictable, intermittent, sparse, multi-resolution, and multi-scale. The first half addresses data science’s essential concepts and theoretical underpinnings, while the second section contains key themes and sharing experiences from current publications focused on each stage of the data life cycle. Finally, themes such as open science, smart data, and team science are considered.

Keywords: Data science, intelligent system, machine learning, big data, data life cycle, recent development, geo science

Procedia PDF Downloads 122
24291 Enhancing of Biogas Production from Slaughterhouse and Dairy Farm Waste with Pasteurization

Authors: Mahmoud Hassan Onsa, Saadelnour Abdueljabbar Adam

Abstract:

Wastes from slaughterhouses in most towns in Sudan are often poorly managed and sometimes discharged into adjoining streams due to poor implementation of standards, thus causing environmental and public health hazards and also there is a large amount of manure from dairy farms. This paper presents solution of organic waste from cow dairy farms and slaughterhouse the anaerobic digestion and biogas production. The paper presents the findings of experimental investigation of biogas production with and without pasteurization using cow manure, blood and rumen content were mixed at two proportions, 72.3% manure, 21.7%, rumen content and 6% blood for bio digester1with 62% dry matter at the beginning and without pasteurization and 72.3% manure, 21.7%, rumen content and 6% blood for bio-digester2 with 10% dry matter and pasteurization. The paper analyses the quantitative and qualitative composition of biogas: gas content, the concentration of methane. The highest biogas output 2.9 mL/g dry matter/day (from bio-digester2) together with a high quality biogas of 87.4% methane content which is useful for combustion and energy production and healthy bio-fertilizer but biodigester1 gave 1.68 mL/g dry matter/day with methane content 85% which is useful for combustion, energy production and can be considered as new technology of dryer bio-digesters.

Keywords: anaerobic digestion, bio-digester, blood, cow manure, rumen content

Procedia PDF Downloads 704
24290 Data Quality as a Pillar of Data-Driven Organizations: Exploring the Benefits of Data Mesh

Authors: Marc Bachelet, Abhijit Kumar Chatterjee, José Manuel Avila

Abstract:

Data quality is a key component of any data-driven organization. Without data quality, organizations cannot effectively make data-driven decisions, which often leads to poor business performance. Therefore, it is important for an organization to ensure that the data they use is of high quality. This is where the concept of data mesh comes in. Data mesh is an organizational and architectural decentralized approach to data management that can help organizations improve the quality of data. The concept of data mesh was first introduced in 2020. Its purpose is to decentralize data ownership, making it easier for domain experts to manage the data. This can help organizations improve data quality by reducing the reliance on centralized data teams and allowing domain experts to take charge of their data. This paper intends to discuss how a set of elements, including data mesh, are tools capable of increasing data quality. One of the key benefits of data mesh is improved metadata management. In a traditional data architecture, metadata management is typically centralized, which can lead to data silos and poor data quality. With data mesh, metadata is managed in a decentralized manner, ensuring accurate and up-to-date metadata, thereby improving data quality. Another benefit of data mesh is the clarification of roles and responsibilities. In a traditional data architecture, data teams are responsible for managing all aspects of data, which can lead to confusion and ambiguity in responsibilities. With data mesh, domain experts are responsible for managing their own data, which can help provide clarity in roles and responsibilities and improve data quality. Additionally, data mesh can also contribute to a new form of organization that is more agile and adaptable. By decentralizing data ownership, organizations can respond more quickly to changes in their business environment, which in turn can help improve overall performance by allowing better insights into business as an effect of better reports and visualization tools. Monitoring and analytics are also important aspects of data quality. With data mesh, monitoring, and analytics are decentralized, allowing domain experts to monitor and analyze their own data. This will help in identifying and addressing data quality problems in quick time, leading to improved data quality. Data culture is another major aspect of data quality. With data mesh, domain experts are encouraged to take ownership of their data, which can help create a data-driven culture within the organization. This can lead to improved data quality and better business outcomes. Finally, the paper explores the contribution of AI in the coming years. AI can help enhance data quality by automating many data-related tasks, like data cleaning and data validation. By integrating AI into data mesh, organizations can further enhance the quality of their data. The concepts mentioned above are illustrated by AEKIDEN experience feedback. AEKIDEN is an international data-driven consultancy that has successfully implemented a data mesh approach. By sharing their experience, AEKIDEN can help other organizations understand the benefits and challenges of implementing data mesh and improving data quality.

Keywords: data culture, data-driven organization, data mesh, data quality for business success

Procedia PDF Downloads 115
24289 Economic Growth through Quality in Higher Education

Authors: Mohammad Mushir Khan, C. Satyanarayana

Abstract:

Education is considered as one of the prime bottlenecks in the economic growth of India. The Ministry of Human Resource & Development, Government of India has, therefore, given special attention to this issue and the Gross Enrollment Ratio (GER) in Higher Education has increased marginally during last five years, with the efforts and various policy decisions like Right to Education (RTE) and other fee reimbursement schemes, initiated by the State Governments. But still this is one of the lowest, if assessed at the global level. It is true that the GER has improved but the survey reveals that the quality has been badly affected. This paper tries to assess the impact of lack of quality education in various sectors that affects Indian Economy and thereby signifies the need of immediate policy decision at the government level. It is to be noted that in higher education, science, management, engineering and technology plays vital role as far as shaping country’s economy is concerned and as such the quality needs to be addressed, particularly, in these streams. The paper, after carefully studying lots of survey reports and other government/ non-government documents recommends measures to be initiated by the Central Government, on priority, for improving quality of education. The quality up-gradation in higher education single handedly provides real fuel to the India’s growth Engine, as it has potential to touch each and every sector that strengthens country’s economy.

Keywords: higher education, economy, accreditation, industry, technology

Procedia PDF Downloads 404
24288 Big Data Analysis with RHadoop

Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim

Abstract:

It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop

Procedia PDF Downloads 416