Search results for: Atomic data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25015

Search results for: Atomic data

24145 Haemocompatibility of Surface Modified AISI 316L Austenitic Stainless Steel Tested in Artificial Plasma

Authors: W. Walke, J. Przondziono, K. Nowińska

Abstract:

The study comprises evaluation of suitability of passive layer created on the surface of AISI 316L stainless steel for products that are intended to have contact with blood. For that purpose, prior to and after chemical passivation, samples were subject to 7 day exposure in artificial plasma at the temperature of T=37°C. Next, tests of metallic ions infiltration from the surface to the solution were performed. The tests were performed with application of spectrometer JY 2000, by Yobin – Yvon, employing Inductively Coupled Plasma Atomic Emission Spectrometry (ICP-AES). In order to characterize physical and chemical features of electrochemical processes taking place during exposure of samples to artificial plasma, tests with application of electrochemical impedance spectroscopy were suggested. The tests were performed with application of measuring unit equipped with potentiostat PGSTAT 302n with an attachment for impedance tests FRA2. Measurements were made in the environment simulating human blood at the temperature of T=37°C. Performed tests proved that application of chemical passivation process for AISI 316L stainless steel used for production of goods intended to have contact with blood is well-grounded and useful in order to improve safety of their usage.

Keywords: AISI 316L stainless steel, chemical passivation, artificial plasma, ions infiltration, EIS

Procedia PDF Downloads 250
24144 Integration of Knowledge and Metadata for Complex Data Warehouses and Big Data

Authors: Jean Christian Ralaivao, Fabrice Razafindraibe, Hasina Rakotonirainy

Abstract:

This document constitutes a resumption of work carried out in the field of complex data warehouses (DW) relating to the management and formalization of knowledge and metadata. It offers a methodological approach for integrating two concepts, knowledge and metadata, within the framework of a complex DW architecture. The objective of the work considers the use of the technique of knowledge representation by description logics and the extension of Common Warehouse Metamodel (CWM) specifications. This will lead to a fallout in terms of the performance of a complex DW. Three essential aspects of this work are expected, including the representation of knowledge in description logics and the declination of this knowledge into consistent UML diagrams while respecting or extending the CWM specifications and using XML as pivot. The field of application is large but will be adapted to systems with heteroge-neous, complex and unstructured content and moreover requiring a great (re)use of knowledge such as medical data warehouses.

Keywords: data warehouse, description logics, integration, knowledge, metadata

Procedia PDF Downloads 122
24143 Data Analytics in Energy Management

Authors: Sanjivrao Katakam, Thanumoorthi I., Antony Gerald, Ratan Kulkarni, Shaju Nair

Abstract:

With increasing energy costs and its impact on the business, sustainability today has evolved from a social expectation to an economic imperative. Therefore, finding methods to reduce cost has become a critical directive for Industry leaders. Effective energy management is the only way to cut costs. However, Energy Management has been a challenge because it requires a change in old habits and legacy systems followed for decades. Today exorbitant levels of energy and operational data is being captured and stored by Industries, but they are unable to convert these structured and unstructured data sets into meaningful business intelligence. It must be noted that for quick decisions, organizations must learn to cope with large volumes of operational data in different formats. Energy analytics not only helps in extracting inferences from these data sets, but also is instrumental in transformation from old approaches of energy management to new. This in turn assists in effective decision making for implementation. It is the requirement of organizations to have an established corporate strategy for reducing operational costs through visibility and optimization of energy usage. Energy analytics play a key role in optimization of operations. The paper describes how today energy data analytics is extensively used in different scenarios like reducing operational costs, predicting energy demands, optimizing network efficiency, asset maintenance, improving customer insights and device data insights. The paper also highlights how analytics helps transform insights obtained from energy data into sustainable solutions. The paper utilizes data from an array of segments such as retail, transportation, and water sectors.

Keywords: energy analytics, energy management, operational data, business intelligence, optimization

Procedia PDF Downloads 349
24142 Efficient Frequent Itemset Mining Methods over Real-Time Spatial Big Data

Authors: Hamdi Sana, Emna Bouazizi, Sami Faiz

Abstract:

In recent years, there is a huge increase in the use of spatio-temporal applications where data and queries are continuously moving. As a result, the need to process real-time spatio-temporal data seems clear and real-time stream data management becomes a hot topic. Sliding window model and frequent itemset mining over dynamic data are the most important problems in the context of data mining. Thus, sliding window model for frequent itemset mining is a widely used model for data stream mining due to its emphasis on recent data and its bounded memory requirement. These methods use the traditional transaction-based sliding window model where the window size is based on a fixed number of transactions. Actually, this model supposes that all transactions have a constant rate which is not suited for real-time applications. And the use of this model in such applications endangers their performance. Based on these observations, this paper relaxes the notion of window size and proposes the use of a timestamp-based sliding window model. In our proposed frequent itemset mining algorithm, support conditions are used to differentiate frequents and infrequent patterns. Thereafter, a tree is developed to incrementally maintain the essential information. We evaluate our contribution. The preliminary results are quite promising.

Keywords: real-time spatial big data, frequent itemset, transaction-based sliding window model, timestamp-based sliding window model, weighted frequent patterns, tree, stream query

Procedia PDF Downloads 145
24141 The Extent of Big Data Analysis by the External Auditors

Authors: Iyad Ismail, Fathilatul Abdul Hamid

Abstract:

This research was mainly investigated to recognize the extent of big data analysis by external auditors. This paper adopts grounded theory as a framework for conducting a series of semi-structured interviews with eighteen external auditors. The research findings comprised the availability extent of big data and big data analysis usage by the external auditors in Palestine, Gaza Strip. Considering the study's outcomes leads to a series of auditing procedures in order to improve the external auditing techniques, which leads to high-quality audit process. Also, this research is crucial for auditing firms by giving an insight into the mechanisms of auditing firms to identify the most important strategies that help in achieving competitive audit quality. These results are aims to instruct the auditing academic and professional institutions in developing techniques for external auditors in order to the big data analysis. This paper provides appropriate information for the decision-making process and a source of future information which affects technological auditing.

Keywords: big data analysis, external auditors, audit reliance, internal audit function

Procedia PDF Downloads 50
24140 Preparation and Characterization of a Nickel-Based Catalyst Supported by Silica Promoted by Cerium for the Methane Steam Reforming Reaction

Authors: Ali Zazi, Ouiza Cherifi

Abstract:

Natural gas currently represents a raw material of choice for the manufacture of a wide range of chemical products via synthesis gas, among the routes of transformation of methane into synthesis gas The reaction of the oxidation of methane by gas vapor 'water. This work focuses on the study of the effect of cerieum on the nickel-based catalyst supported by silica for the methane vapor reforming reaction, with a variation of certain parameters of the reaction. The reaction temperature, the H₂O / CH₄ ratio and the flow rate of the reaction mixture (CH₄-H₂O). Two catalysts were prepared by impregnation of Degussa silica with a solution of nickel nitrates and a solution of cerium nitrates [Ni (NO₃) 2 6H₂O and Ce (NO₃) 3 6H₂O] so as to obtain the 1.5% nickel concentrations. For both catalysts and plus 1% cerium for the second catalyst. These Catalysts have been characterized by physical and chemical analysis techniques: BET technique, Atomic Absorption, IR Spectroscopy, X-ray diffraction. These characterizations indicated that the nitrates had impregnated the silica. And that the NiO and Ce₂O3 phases are present and Ni°(after reaction). The BET surface of the silica decreases without being affected. The catalytic tests carried out on the two catalysts for the steam reforming reactions show that the addition of cerium to the nickel improves the catalytic performances of the nickel. And that these performances also depend on the parameters of the reaction, namely the temperature, the rate of the reaction mixture, and the ratio (H₂O / CH₄).

Keywords: heterogeneous catalysis, steam reforming, Methane, Nickel, Cerium, synthesis gas, hydrogen

Procedia PDF Downloads 148
24139 A Model of Teacher Leadership in History Instruction

Authors: Poramatdha Chutimant

Abstract:

The objective of the research was to propose a model of teacher leadership in history instruction for utilization. Everett M. Rogers’ Diffusion of Innovations Theory is applied as theoretical framework. Qualitative method is to be used in the study, and the interview protocol used as an instrument to collect primary data from best practices who awarded by Office of National Education Commission (ONEC). Open-end questions will be used in interview protocol in order to gather the various data. Then, information according to international context of history instruction is the secondary data used to support in the summarizing process (Content Analysis). Dendrogram is a key to interpret and synthesize the primary data. Thus, secondary data comes as the supportive issue in explanation and elaboration. In-depth interview is to be used to collected information from seven experts in educational field. The focal point is to validate a draft model in term of future utilization finally.

Keywords: history study, nationalism, patriotism, responsible citizenship, teacher leadership

Procedia PDF Downloads 265
24138 The Effect of Institutions on Economic Growth: An Analysis Based on Bayesian Panel Data Estimation

Authors: Mohammad Anwar, Shah Waliullah

Abstract:

This study investigated panel data regression models. This paper used Bayesian and classical methods to study the impact of institutions on economic growth from data (1990-2014), especially in developing countries. Under the classical and Bayesian methodology, the two-panel data models were estimated, which are common effects and fixed effects. For the Bayesian approach, the prior information is used in this paper, and normal gamma prior is used for the panel data models. The analysis was done through WinBUGS14 software. The estimated results of the study showed that panel data models are valid models in Bayesian methodology. In the Bayesian approach, the effects of all independent variables were positively and significantly affected by the dependent variables. Based on the standard errors of all models, we must say that the fixed effect model is the best model in the Bayesian estimation of panel data models. Also, it was proved that the fixed effect model has the lowest value of standard error, as compared to other models.

Keywords: Bayesian approach, common effect, fixed effect, random effect, Dynamic Random Effect Model

Procedia PDF Downloads 58
24137 Investigation of the Mechanical Performance of Carbon Nanomembranes for Water Separation Technologies

Authors: Marinos Dimitropoulos, George Trakakis, Nikolaus Meyerbröker, Raphael Dalpke, Polina Angelova, Albert Schnieders, Christos Pavlou, Christos Kostaras, Costas Galiotis, Konstantinos Dassios

Abstract:

Intended for purifying water, water separation technologies are widely employed in a variety of contemporary household and industrial applications. Ultrathin Carbon Nanomembranes (CNMs) offer a highly selective, fast-flow, energy-efficient water separation technology intended for demanding water treatment applications as a technological replacement for biological filtration membranes. The membranes are two-dimensional (2D) materials with sub-nm functional pores and a thickness of roughly 1 nm; they may be generated in large quantities on porous supporting substrates and have customizable properties. The purpose of this work was to investigate and analyze the mechanical characteristics of CNMs and their substrates in order to ensure the structural stability of the membrane during operation. Contrary to macro-materials, it is difficult to measure the mechanical properties of membranes that are only a few nanometers thick. The membranes were supported on atomically flat substrates as well as suspended over patterned substrates, and their inherent mechanical properties were tested with atomic force microscopy. Quantitative experiments under nanomechanical loading, nanoindentation, and nano fatigue demonstrated the membranes' potential for usage in water separation applications.

Keywords: carbon nanomembranes, mechanical properties, AFM

Procedia PDF Downloads 71
24136 Diagnosis of the Heart Rhythm Disorders by Using Hybrid Classifiers

Authors: Sule Yucelbas, Gulay Tezel, Cuneyt Yucelbas, Seral Ozsen

Abstract:

In this study, it was tried to identify some heart rhythm disorders by electrocardiography (ECG) data that is taken from MIT-BIH arrhythmia database by subtracting the required features, presenting to artificial neural networks (ANN), artificial immune systems (AIS), artificial neural network based on artificial immune system (AIS-ANN) and particle swarm optimization based artificial neural network (PSO-NN) classifier systems. The main purpose of this study is to evaluate the performance of hybrid AIS-ANN and PSO-ANN classifiers with regard to the ANN and AIS. For this purpose, the normal sinus rhythm (NSR), atrial premature contraction (APC), sinus arrhythmia (SA), ventricular trigeminy (VTI), ventricular tachycardia (VTK) and atrial fibrillation (AF) data for each of the RR intervals were found. Then these data in the form of pairs (NSR-APC, NSR-SA, NSR-VTI, NSR-VTK and NSR-AF) is created by combining discrete wavelet transform which is applied to each of these two groups of data and two different data sets with 9 and 27 features were obtained from each of them after data reduction. Afterwards, the data randomly was firstly mixed within themselves, and then 4-fold cross validation method was applied to create the training and testing data. The training and testing accuracy rates and training time are compared with each other. As a result, performances of the hybrid classification systems, AIS-ANN and PSO-ANN were seen to be close to the performance of the ANN system. Also, the results of the hybrid systems were much better than AIS, too. However, ANN had much shorter period of training time than other systems. In terms of training times, ANN was followed by PSO-ANN, AIS-ANN and AIS systems respectively. Also, the features that extracted from the data affected the classification results significantly.

Keywords: AIS, ANN, ECG, hybrid classifiers, PSO

Procedia PDF Downloads 426
24135 Topic Modelling Using Latent Dirichlet Allocation and Latent Semantic Indexing on SA Telco Twitter Data

Authors: Phumelele Kubheka, Pius Owolawi, Gbolahan Aiyetoro

Abstract:

Twitter is one of the most popular social media platforms where users can share their opinions on different subjects. As of 2010, The Twitter platform generates more than 12 Terabytes of data daily, ~ 4.3 petabytes in a single year. For this reason, Twitter is a great source for big mining data. Many industries such as Telecommunication companies can leverage the availability of Twitter data to better understand their markets and make an appropriate business decision. This study performs topic modeling on Twitter data using Latent Dirichlet Allocation (LDA). The obtained results are benchmarked with another topic modeling technique, Latent Semantic Indexing (LSI). The study aims to retrieve topics on a Twitter dataset containing user tweets on South African Telcos. Results from this study show that LSI is much faster than LDA. However, LDA yields better results with higher topic coherence by 8% for the best-performing model represented in Table 1. A higher topic coherence score indicates better performance of the model.

Keywords: big data, latent Dirichlet allocation, latent semantic indexing, telco, topic modeling, twitter

Procedia PDF Downloads 137
24134 Enhance the Power of Sentiment Analysis

Authors: Yu Zhang, Pedro Desouza

Abstract:

Since big data has become substantially more accessible and manageable due to the development of powerful tools for dealing with unstructured data, people are eager to mine information from social media resources that could not be handled in the past. Sentiment analysis, as a novel branch of text mining, has in the last decade become increasingly important in marketing analysis, customer risk prediction and other fields. Scientists and researchers have undertaken significant work in creating and improving their sentiment models. In this paper, we present a concept of selecting appropriate classifiers based on the features and qualities of data sources by comparing the performances of five classifiers with three popular social media data sources: Twitter, Amazon Customer Reviews, and Movie Reviews. We introduced a couple of innovative models that outperform traditional sentiment classifiers for these data sources, and provide insights on how to further improve the predictive power of sentiment analysis. The modelling and testing work was done in R and Greenplum in-database analytic tools.

Keywords: sentiment analysis, social media, Twitter, Amazon, data mining, machine learning, text mining

Procedia PDF Downloads 336
24133 Real-Time Big-Data Warehouse a Next-Generation Enterprise Data Warehouse and Analysis Framework

Authors: Abbas Raza Ali

Abstract:

Big Data technology is gradually becoming a dire need of large enterprises. These enterprises are generating massively large amount of off-line and streaming data in both structured and unstructured formats on daily basis. It is a challenging task to effectively extract useful insights from the large scale datasets, even though sometimes it becomes a technology constraint to manage transactional data history of more than a few months. This paper presents a framework to efficiently manage massively large and complex datasets. The framework has been tested on a communication service provider producing massively large complex streaming data in binary format. The communication industry is bound by the regulators to manage history of their subscribers’ call records where every action of a subscriber generates a record. Also, managing and analyzing transactional data allows service providers to better understand their customers’ behavior, for example, deep packet inspection requires transactional internet usage data to explain internet usage behaviour of the subscribers. However, current relational database systems limit service providers to only maintain history at semantic level which is aggregated at subscriber level. The framework addresses these challenges by leveraging Big Data technology which optimally manages and allows deep analysis of complex datasets. The framework has been applied to offload existing Intelligent Network Mediation and relational Data Warehouse of the service provider on Big Data. The service provider has 50+ million subscriber-base with yearly growth of 7-10%. The end-to-end process takes not more than 10 minutes which involves binary to ASCII decoding of call detail records, stitching of all the interrogations against a call (transformations) and aggregations of all the call records of a subscriber.

Keywords: big data, communication service providers, enterprise data warehouse, stream computing, Telco IN Mediation

Procedia PDF Downloads 158
24132 Preparation of Silicon-Based Oxide Hollow Nanofibers Using Single-Nozzle Electrospinning

Authors: Juiwen Liang, Choliang Chung

Abstract:

In this study, the silicon-base oxide nanofibers with hollow structure were prepared using single-nozzle electrospinning and heat treatment. Firstly, precursor solution was prepared: the Polyvinylpyrrolidone (PVP) and Tetraethyl orthosilicate (TEOS) dissolved in ethanol and to make sure the concentration of solution in appropriate using single-nozzle electrospinning to produce the nanofibers. Secondly, control morphology of the electrostatic spinning nanofibers was conducted, and design the temperature profile to created hollow nanofibers, exploring the morphology and properties of nanofibers. The characterized of nanofibers, following instruments were used: Atomic force microscopy (AFM), Field Emission Scanning Electron Microscope (FE-SEM), Transmission electron microscopy (TEM), Photoluminescence (PL), X-ray Diffraction (XRD). The AFM was used to scan the nanofibers, and 3D Graphics were applied to explore the surface morphology of fibers. FE-SEM and TEM were used to explore the morphology and diameter of nanofibers and hollow nanofiber. The excitation and emission spectra explored by PL. Finally, XRD was used for identified crystallization of ceramic nanofibers. Using electrospinning technique followed by subsequent heat treatment, we have successfully prepared silicon-base oxide nanofibers with hollow structure. Thus, the microstructure and morphology of electrostatic spinning silicon-base oxide hollow nanofibers were explored. Major characteristics of the nanofiber in terms of crystalline, optical properties and crystal structure were identified.

Keywords: electrospinning, single-nozzle, hollow, nanofibers

Procedia PDF Downloads 337
24131 Programming with Grammars

Authors: Peter M. Maurer Maurer

Abstract:

DGL is a context free grammar-based tool for generating random data. Many types of simulator input data require some computation to be placed in the proper format. For example, it might be necessary to generate ordered triples in which the third element is the sum of the first two elements, or it might be necessary to generate random numbers in some sorted order. Although DGL is universal in computational power, generating these types of data is extremely difficult. To overcome this problem, we have enhanced DGL to include features that permit direct computation within the structure of a context free grammar. The features have been implemented as special types of productions, preserving the context free flavor of DGL specifications.

Keywords: DGL, Enhanced Context Free Grammars, Programming Constructs, Random Data Generation

Procedia PDF Downloads 132
24130 A Model Architecture Transformation with Approach by Modeling: From UML to Multidimensional Schemas of Data Warehouses

Authors: Ouzayr Rabhi, Ibtissam Arrassen

Abstract:

To provide a complete analysis of the organization and to help decision-making, leaders need to have relevant data; Data Warehouses (DW) are designed to meet such needs. However, designing DW is not trivial and there is no formal method to derive a multidimensional schema from heterogeneous databases. In this article, we present a Model-Driven based approach concerning the design of data warehouses. We describe a multidimensional meta-model and also specify a set of transformations starting from a Unified Modeling Language (UML) metamodel. In this approach, the UML metamodel and the multidimensional one are both considered as a platform-independent model (PIM). The first meta-model is mapped into the second one through transformation rules carried out by the Query View Transformation (QVT) language. This proposal is validated through the application of our approach to generating a multidimensional schema of a Balanced Scorecard (BSC) DW. We are interested in the BSC perspectives, which are highly linked to the vision and the strategies of an organization.

Keywords: data warehouse, meta-model, model-driven architecture, transformation, UML

Procedia PDF Downloads 140
24129 Secured Embedding of Patient’s Confidential Data in Electrocardiogram Using Chaotic Maps

Authors: Butta Singh

Abstract:

This paper presents a chaotic map based approach for secured embedding of patient’s confidential data in electrocardiogram (ECG) signal. The chaotic map generates predefined locations through the use of selective control parameters. The sample value difference method effectually hides the confidential data in ECG sample pairs at these predefined locations. Evaluation of proposed method on all 48 records of MIT-BIH arrhythmia ECG database demonstrates that the embedding does not alter the diagnostic features of cover ECG. The secret data imperceptibility in stego-ECG is evident through various statistical and clinical performance measures. Statistical metrics comprise of Percentage Root Mean Square Difference (PRD) and Peak Signal to Noise Ratio (PSNR). Further, a comparative analysis between proposed method and existing approaches was also performed. The results clearly demonstrated the superiority of proposed method.

Keywords: chaotic maps, ECG steganography, data embedding, electrocardiogram

Procedia PDF Downloads 171
24128 The Imagined Scientific Drawing as a Representative of the Content Provided by Emotions to Scientific Rationality

Authors: Dení Stincer Gómez, Zuraya Monroy Nasr

Abstract:

From the epistemology of emotions, one of the topics of current reflection is the function that emotions fulfill in the rational processes involved in scientific activity. So far, three functions have been assigned to them: selective, heuristic, and carriers of content. In this last function, it is argued that emotions, like our perceptual organs, contribute relevant content to reasoning, which is then converted into linguistic statements or graphic representations. In this paper, of a qualitative and philosophical nature, arguments are provided for two hypotheses 1) if emotions provide content to the mind, which then translates it into language or representations, then it is important to take up the idea of the Saussurean linguistic sign to understand this process. This sign has two elements: the signified and the signifier. Emotions would provide meanings, and reasoning creates the signifier, and 2) the meanings provided by emotions are properties and qualities of phenomena generally not accessible to the sense organs. These meanings must be imagined, and the imagination is nurtured by the feeling that "maybe this is the way." One way to access the content provided by emotions can be through imagined scientific drawings. The atomic models created since Thomson, the structure of crystals by René Just, the representations of lunar eclipses by Johannes, fractal geometry, and the structure of DNA, among others, have resulted fundamentally from the imagination. These representations, not provided by the sense organs, seem to come from the emotional involvement of scientists in their desire to understand, explain and discover.

Keywords: emotions, epistemic functions of emotions, scientific drawing, linguistic sign

Procedia PDF Downloads 56
24127 Phytochemical Screening, Antioxidant Potential, and Mineral Composition of Dried Abelmoschus esculentus L. Fruits Consume in Gada Area of Sokoto State, Nigeria

Authors: I. Sani, F. Bello, I. M. Fakai, A. Abdulhamid

Abstract:

Abelmoschus esculentus L. fruit is very common especially in northern part of Nigeria, but people are ignorant of its medicinal and pharmacological benefits. Preliminary phytochemical screening, antioxidant potential and mineral composition of the dried form of this fruit were determined. The Phytochemical screening was conducted using standard methods. Antioxidant potential screening was carried out using Ferric Reducing Antioxidant Power Assay (FRAP) method, while, the mineral compositions were analyzed using an atomic absorption spectrophotometer by wet digest method. The result of the qualitative phytochemical screening revealed that the fruits contain saponins, flavonoids, tannins, steroids, and terpenoids, while, anthraquinone, alkaloids, phenols, glycosides, and phlobatannins were not detected. The quantitative analysis revealed that the fruits contain saponnins (380 ± 0.020 mg/g), flavonoids (240±0.01 mg/g), and tannins (21.71 ± 0.66 mg/ml). The antioxidant potential was determined to be 54.1 ± 0.19%. The mineral composition revealed that 100 g of the fruits contains 97.52 ± 1.04 mg of magnesium (Mg), 94.53 ± 3.21 mg of calcium (Ca), 77.10 ± 0.79 mg of iron (Fe), 47.14 ± 0.41 mg of zinc (Zn), 43.96 ± 1.49 mg of potassium (K), 42.02 ± 1.09 mg of sodium (Na), 0.47 ± 0.08 mg of copper (Cu) and 0.10 ± 0.02 mg of lead (Pb). These results showed that the Abelmoschus esculentus L. fruit is a good source of antioxidants, and contains an appreciable amount of phytochemicals, therefore, it has some pharmacological attributes. On the other side, the fruit can serve as a nutritional supplement for Mg, Ca, Fe, Zn, K, and Na, but a poor source of Cu, and contains no significant amount of Pb.

Keywords: Abelmoschus esculentus Fruits, antioxidant potential, mineral composition, phytochemical screening

Procedia PDF Downloads 357
24126 Detection Efficient Enterprises via Data Envelopment Analysis

Authors: S. Turkan

Abstract:

In this paper, the Turkey’s Top 500 Industrial Enterprises data in 2014 were analyzed by data envelopment analysis. Data envelopment analysis is used to detect efficient decision-making units such as universities, hospitals, schools etc. by using inputs and outputs. The decision-making units in this study are enterprises. To detect efficient enterprises, some financial ratios are determined as inputs and outputs. For this reason, financial indicators related to productivity of enterprises are considered. The efficient foreign weighted owned capital enterprises are detected via super efficiency model. According to the results, it is said that Mercedes-Benz is the most efficient foreign weighted owned capital enterprise in Turkey.

Keywords: data envelopment analysis, super efficiency, logistic regression, financial ratios

Procedia PDF Downloads 312
24125 Use of Carica papaya as a Bio-Sorbent for Removal of Heavy Metals in Wastewater

Authors: W. E. Igwegbe, B. C. Okoro, J. C. Osuagwu

Abstract:

The study was aimed at assessing the effectiveness of reducing the concentrations of heavy metals in waste water using Pawpaw (Carica papaya) wood as a bio-sorbent. The heavy metals considered include; zinc, cadmium, lead, copper, iron, selenium, nickel, and manganese. The physiochemical properties of carica papaya stem were studied. The experimental sample was obtained from a felled trunk of matured pawpaw tree. Waste water for experimental use was prepared by dissolving soil samples collected from a dump site at Owerri, Imo state in water. The concentration of each metal remaining in solution as residual metal after bio-sorption was determined using Atomic absorption Spectrometer. The effects of ph, contact time and initial heavy metal concentration were studied in a batch reactor. The results of Spectrometer test showed that there were different functional groups detected in the carica papaya stem biomass. Optimum bio-sorption occurred at pH 5.9 with 5g/100ml solution of bio-sorbent. The results of the study showed that the treated wastewater is fit for irrigation purpose based on Canada wastewater quality guideline for the protection of Agricultural standard. This approach thus provides a cost effective and environmentally friendly option for treating waste water.

Keywords: biomass, bio-sorption, Carica papaya, heavy metal, wastewater

Procedia PDF Downloads 354
24124 Intelligent Process Data Mining for Monitoring for Fault-Free Operation of Industrial Processes

Authors: Hyun-Woo Cho

Abstract:

The real-time fault monitoring and diagnosis of large scale production processes is helpful and necessary in order to operate industrial process safely and efficiently producing good final product quality. Unusual and abnormal events of the process may have a serious impact on the process such as malfunctions or breakdowns. This work try to utilize process measurement data obtained in an on-line basis for the safe and some fault-free operation of industrial processes. To this end, this work evaluated the proposed intelligent process data monitoring framework based on a simulation process. The monitoring scheme extracts the fault pattern in the reduced space for the reliable data representation. Moreover, this work shows the results of using linear and nonlinear techniques for the monitoring purpose. It has shown that the nonlinear technique produced more reliable monitoring results and outperforms linear methods. The adoption of the qualitative monitoring model helps to reduce the sensitivity of the fault pattern to noise.

Keywords: process data, data mining, process operation, real-time monitoring

Procedia PDF Downloads 620
24123 A Machine Learning Approach for the Leakage Classification in the Hydraulic Final Test

Authors: Christian Neunzig, Simon Fahle, Jürgen Schulz, Matthias Möller, Bernd Kuhlenkötter

Abstract:

The widespread use of machine learning applications in production is significantly accelerated by improved computing power and increasing data availability. Predictive quality enables the assurance of product quality by using machine learning models as a basis for decisions on test results. The use of real Bosch production data based on geometric gauge blocks from machining, mating data from assembly and hydraulic measurement data from final testing of directional valves is a promising approach to classifying the quality characteristics of workpieces.

Keywords: machine learning, classification, predictive quality, hydraulics, supervised learning

Procedia PDF Downloads 189
24122 Analysis of Cyber Activities of Potential Business Customers Using Neo4j Graph Databases

Authors: Suglo Tohari Luri

Abstract:

Data analysis is an important aspect of business performance. With the application of artificial intelligence within databases, selecting a suitable database engine for an application design is also very crucial for business data analysis. The application of business intelligence (BI) software into some relational databases such as Neo4j has proved highly effective in terms of customer data analysis. Yet what remains of great concern is the fact that not all business organizations have the neo4j business intelligence software applications to implement for customer data analysis. Further, those with the BI software lack personnel with the requisite expertise to use it effectively with the neo4j database. The purpose of this research is to demonstrate how the Neo4j program code alone can be applied for the analysis of e-commerce website customer visits. As the neo4j database engine is optimized for handling and managing data relationships with the capability of building high performance and scalable systems to handle connected data nodes, it will ensure that business owners who advertise their products at websites using neo4j as a database are able to determine the number of visitors so as to know which products are visited at routine intervals for the necessary decision making. It will also help in knowing the best customer segments in relation to specific goods so as to place more emphasis on their advertisement on the said websites.

Keywords: data, engine, intelligence, customer, neo4j, database

Procedia PDF Downloads 182
24121 Decision Making System for Clinical Datasets

Authors: P. Bharathiraja

Abstract:

Computer Aided decision making system is used to enhance diagnosis and prognosis of diseases and also to assist clinicians and junior doctors in clinical decision making. Medical Data used for decision making should be definite and consistent. Data Mining and soft computing techniques are used for cleaning the data and for incorporating human reasoning in decision making systems. Fuzzy rule based inference technique can be used for classification in order to incorporate human reasoning in the decision making process. In this work, missing values are imputed using the mean or mode of the attribute. The data are normalized using min-ma normalization to improve the design and efficiency of the fuzzy inference system. The fuzzy inference system is used to handle the uncertainties that exist in the medical data. Equal-width-partitioning is used to partition the attribute values into appropriate fuzzy intervals. Fuzzy rules are generated using Class Based Associative rule mining algorithm. The system is trained and tested using heart disease data set from the University of California at Irvine (UCI) Machine Learning Repository. The data was split using a hold out approach into training and testing data. From the experimental results it can be inferred that classification using fuzzy inference system performs better than trivial IF-THEN rule based classification approaches. Furthermore it is observed that the use of fuzzy logic and fuzzy inference mechanism handles uncertainty and also resembles human decision making. The system can be used in the absence of a clinical expert to assist junior doctors and clinicians in clinical decision making.

Keywords: decision making, data mining, normalization, fuzzy rule, classification

Procedia PDF Downloads 499
24120 Estimating Bridge Deterioration for Small Data Sets Using Regression and Markov Models

Authors: Yina F. Muñoz, Alexander Paz, Hanns De La Fuente-Mella, Joaquin V. Fariña, Guilherme M. Sales

Abstract:

The primary approach for estimating bridge deterioration uses Markov-chain models and regression analysis. Traditional Markov models have problems in estimating the required transition probabilities when a small sample size is used. Often, reliable bridge data have not been taken over large periods, thus large data sets may not be available. This study presents an important change to the traditional approach by using the Small Data Method to estimate transition probabilities. The results illustrate that the Small Data Method and traditional approach both provide similar estimates; however, the former method provides results that are more conservative. That is, Small Data Method provided slightly lower than expected bridge condition ratings compared with the traditional approach. Considering that bridges are critical infrastructures, the Small Data Method, which uses more information and provides more conservative estimates, may be more appropriate when the available sample size is small. In addition, regression analysis was used to calculate bridge deterioration. Condition ratings were determined for bridge groups, and the best regression model was selected for each group. The results obtained were very similar to those obtained when using Markov chains; however, it is desirable to use more data for better results.

Keywords: concrete bridges, deterioration, Markov chains, probability matrix

Procedia PDF Downloads 326
24119 Validation of Visibility Data from Road Weather Information Systems by Comparing Three Data Resources: Case Study in Ohio

Authors: Fan Ye

Abstract:

Adverse weather conditions, particularly those with low visibility, are critical to the driving tasks. However, the direct relationship between visibility distances and traffic flow/roadway safety is uncertain due to the limitation of visibility data availability. The recent growth of deployment of Road Weather Information Systems (RWIS) makes segment-specific visibility information available which can be integrated with other Intelligent Transportation System, such as automated warning system and variable speed limit, to improve mobility and safety. Before applying the RWIS visibility measurements in traffic study and operations, it is critical to validate the data. Therefore, an attempt was made in the paper to examine the validity and viability of RWIS visibility data by comparing visibility measurements among RWIS, airport weather stations, and weather information recorded by police in crash reports, based on Ohio data. The results indicated that RWIS visibility measurements were significantly different from airport visibility data in Ohio, but no conclusion regarding the reliability of RWIS visibility could be drawn in the consideration of no verified ground truth in the comparisons. It was suggested that more objective methods are needed to validate the RWIS visibility measurements, such as continuous in-field measurements associated with various weather events using calibrated visibility sensors.

Keywords: RWIS, visibility distance, low visibility, adverse weather

Procedia PDF Downloads 235
24118 Design and Simulation of All Optical Fiber to the Home Network

Authors: Rahul Malhotra

Abstract:

Fiber based access networks can deliver performance that can support the increasing demands for high speed connections. One of the new technologies that have emerged in recent years is Passive Optical Networks. This paper is targeted to show the simultaneous delivery of triple play service (data, voice and video). The comparative investigation and suitability of various data rates is presented. It is demonstrated that as we increase the data rate, number of users to be accommodated decreases due to increase in bit error rate.

Keywords: BER, PON, TDMPON, GPON, CWDM, OLT, ONT

Procedia PDF Downloads 537
24117 Troubleshooting Petroleum Equipment Based on Wireless Sensors Based on Bayesian Algorithm

Authors: Vahid Bayrami Rad

Abstract:

In this research, common methods and techniques have been investigated with a focus on intelligent fault finding and monitoring systems in the oil industry. In fact, remote and intelligent control methods are considered a necessity for implementing various operations in the oil industry, but benefiting from the knowledge extracted from countless data generated with the help of data mining algorithms. It is a avoid way to speed up the operational process for monitoring and troubleshooting in today's big oil companies. Therefore, by comparing data mining algorithms and checking the efficiency and structure and how these algorithms respond in different conditions, The proposed (Bayesian) algorithm using data clustering and their analysis and data evaluation using a colored Petri net has provided an applicable and dynamic model from the point of view of reliability and response time. Therefore, by using this method, it is possible to achieve a dynamic and consistent model of the remote control system and prevent the occurrence of leakage in oil pipelines and refineries and reduce costs and human and financial errors. Statistical data The data obtained from the evaluation process shows an increase in reliability, availability and high speed compared to other previous methods in this proposed method.

Keywords: wireless sensors, petroleum equipment troubleshooting, Bayesian algorithm, colored Petri net, rapid miner, data mining-reliability

Procedia PDF Downloads 47
24116 Quantitative Determination of Heavy Metals in Some Commonly Consumed Herbal Medicines in Kano State, Nigeria

Authors: Aliyu Umar, Mohammed Yau, Faruruwa M. Dahiru, Saed Garba

Abstract:

Evaluation of heavy metals in twelve commonly consumed herbal medicines/preparations in Kano State, Nigeria, was carried out. The samples comprised of five unregistered powdered medicines, namely, Zuwo, (ZW); Rai Dorai, (RD); Miyar Tsanya, (MTS); Bagaruwar Makka, (BM); and Madobiya, (M); five unregistered liquid herbal medicinal concussions for pile (MB), yellow fever (MS), typhoid (MT), stomach pain (MC), sexually transmitted diseases (STDs); and two registered herbal medicines; Alif Powder (AP) and Champion Leaf (CL). The heavy metals evaluation was carried out using Atomic Absorption Spectroscopy (AAS) and the result revealed the concentrations (ppm) ranges of the heavy metals as follows: Cadmium (0.0045 – 0.1601), Chromium (0.0418 – 0.2092), Cobalt (0.0038 – 0.0760), Copper (0.0547 – 0.2465), Iron (0.1197 – 0.3592), Manganese (0.0123 – 1.4462), Nickel (0.0073 – 0.0960), Lead (0.185 – 0.0927) and Zinc (0.0244 – 0.2444). Comparing the results obtained in this work with the standards of the World Health Organization (WHO), the Food and Agricultural Organization (FAO) and permissible limits of other countries, the concentrations of heavy metals in the herbal medicine/preparations are within the allowed permissible limits range in herbal medicines and their use could be safe.

Keywords: Kano state, herbal medicines, registered, unregistered

Procedia PDF Downloads 223