Search results for: tool wear prediction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2714

Search results for: tool wear prediction

2174 Use of Radial Basis Function Neural Network for Bearing Pressure Prediction of Strip Footing on Reinforced Granular Bed Overlying Weak Soil

Authors: Srinath Shetty K., Shivashankar R., Rashmi P. Shetty

Abstract:

Earth reinforcing techniques have become useful and economical to solve problems related to difficult grounds and provide satisfactory foundation performance. In this context, this paper uses radial basis function neural network (RBFNN) for predicting the bearing pressure of strip footing on reinforced granular bed overlying weak soil. The inputs for the neural network models included plate width, thickness of granular bed and number of layers of reinforcements, settlement ratio, water content, dry density, cohesion and angle of friction. The results indicated that RBFNN model exhibited more than 84 % prediction accuracy, thereby demonstrating its application in a geotechnical problem.

Keywords: Bearing pressure, granular bed, radial basis function neural network, strip footing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1933
2173 Modelling Indoor Air Carbon Dioxide (CO2)Concentration using Neural Network

Authors: J-P. Skön, M. Johansson, M. Raatikainen, K. Leiviskä, M. Kolehmainen

Abstract:

The use of neural networks is popular in various building applications such as prediction of heating load, ventilation rate and indoor temperature. Significant is, that only few papers deal with indoor carbon dioxide (CO2) prediction which is a very good indicator of indoor air quality (IAQ). In this study, a data-driven modelling method based on multilayer perceptron network for indoor air carbon dioxide in an apartment building is developed. Temperature and humidity measurements are used as input variables to the network. Motivation for this study derives from the following issues. First, measuring carbon dioxide is expensive and sensors power consumptions is high and secondly, this leads to short operating times of battery-powered sensors. The results show that predicting CO2 concentration based on relative humidity and temperature measurements, is difficult. Therefore, more additional information is needed.

Keywords: Indoor air quality, Modelling, Neural networks

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1880
2172 A Web and Cloud-Based Measurement System Analysis Tool for the Automotive Industry

Authors: C. A. Barros, Ana P. Barroso

Abstract:

Any industrial company needs to determine the amount of variation that exists within its measurement process and guarantee the reliability of their data, studying the performance of their measurement system, in terms of linearity, bias, repeatability and reproducibility and stability. This issue is critical for automotive industry suppliers, who are required to be certified by the 16949:2016 standard (replaces the ISO/TS 16949) of International Automotive Task Force, defining the requirements of a quality management system for companies in the automotive industry. Measurement System Analysis (MSA) is one of the mandatory tools. Frequently, the measurement system in companies is not connected to the equipment and do not incorporate the methods proposed by the Automotive Industry Action Group (AIAG). To address these constraints, an R&D project is in progress, whose objective is to develop a web and cloud-based MSA tool. This MSA tool incorporates Industry 4.0 concepts, such as, Internet of Things (IoT) protocols to assure the connection with the measuring equipment, cloud computing, artificial intelligence, statistical tools, and advanced mathematical algorithms. This paper presents the preliminary findings of the project. The web and cloud-based MSA tool is innovative because it implements all statistical tests proposed in the MSA-4 reference manual from AIAG as well as other emerging methods and techniques. As it is integrated with the measuring devices, it reduces the manual input of data and therefore the errors. The tool ensures traceability of all performed tests and can be used in quality laboratories and in the production lines. Besides, it monitors MSAs over time, allowing both the analysis of deviations from the variation of the measurements performed and the management of measurement equipment and calibrations. To develop the MSA tool a ten-step approach was implemented. Firstly, it was performed a benchmarking analysis of the current competitors and commercial solutions linked to MSA, concerning Industry 4.0 paradigm. Next, an analysis of the size of the target market for the MSA tool was done. Afterwards, data flow and traceability requirements were analysed in order to implement an IoT data network that interconnects with the equipment, preferably via wireless. The MSA web solution was designed under UI/UX principles and an API in python language was developed to perform the algorithms and the statistical analysis. Continuous validation of the tool by companies is being performed to assure real time management of the ‘big data’. The main results of this R&D project are: MSA Tool, web and cloud-based; Python API; New Algorithms to the market; and Style Guide of UI/UX of the tool. The MSA tool proposed adds value to the state of the art as it ensures an effective response to the new challenges of measurement systems, which are increasingly critical in production processes. Although the automotive industry has triggered the development of this innovative MSA tool, other industries would also benefit from it. Currently, companies from molds and plastics, chemical and food industry are already validating it.

Keywords: Automotive industry, Industry 4.0, internet of things, IATF 16949:2016, measurement system analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 979
2171 The Use of Voltage Stability Indices and Proposed Instability Prediction to Coordinate with Protection Systems

Authors: R. Leelaruji, V. Knazkins

Abstract:

This paper proposes a methodology for mitigating the occurrence of cascading failure in stressed power systems. The methodology is essentially based on predicting voltage instability in the power system using a voltage stability index and then devising a corrective action in order to increase the voltage stability margin. The paper starts with a brief description of the cascading failure mechanism which is probable root cause of severe blackouts. Then, the voltage instability indices are introduced in order to evaluate stability limit. The aim of the analysis is to assure that the coordination of protection, by adopting load shedding scheme, capable of enhancing performance of the system after the major location of instability is determined. Finally, the proposed method to generate instability prediction is introduced.

Keywords: Blackouts, cascading failure, voltage stability indices, singular value decomposition, load shedding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1538
2170 A Fuzzy Model and Tool to Analyze SIVD Diseases Using TMS

Authors: A. Faro, D. Giordano, M. Pennisi, G. Scarciofalo, C. Spampinato, F. Tramontana

Abstract:

The paper proposes a methodology to process the signals coming from the Transcranial Magnetic Stimulation (TMS) in order to identify the pathology and evaluate the therapy to treat the patients affected by demency diseases. In particular, a fuzzy model is developed to identify the demency of the patients affected by Subcortical Ischemic Vascular Dementia and to measure the positive effect, if any, of a repetitive TMS on their motor performances. A tool is also presented to support the mentioned analysis.

Keywords: TMS, SIVD, Electromiography , Fuzzy Logic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1562
2169 An Alternative Approach for Assessing the Impact of Cutting Conditions on Surface Roughness Using Single Decision Tree

Authors: S. Ghorbani, N. I. Polushin

Abstract:

In this study, an approach to identify factors affecting on surface roughness in a machining process is presented. This study is based on 81 data about surface roughness over a wide range of cutting tools (conventional, cutting tool with holes, cutting tool with composite material), workpiece materials (AISI 1045 Steel, AA2024 aluminum alloy, A48-class30 gray cast iron), spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev), depth of cut (0.05-0.15 mm) and tool overhang (41-65 mm). A single decision tree (SDT) analysis was done to identify factors for predicting a model of surface roughness, and the CART algorithm was employed for building and evaluating regression tree. Results show that a single decision tree is better than traditional regression models with higher rate and forecast accuracy and strong value.

Keywords: Cutting condition, surface roughness, decision tree, CART algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 857
2168 WPRiMA Tool: Managing Risks in Web Projects

Authors: Thamer Al-Rousan, Shahida Sulaiman, Rosalina Abdul Salam

Abstract:

Risk management is an essential fraction of project management, which plays a significant role in project success. Many failures associated with Web projects are the consequences of poor awareness of the risks involved and lack of process models that can serve as a guideline for the development of Web based applications. To circumvent this problem, contemporary process models have been devised for the development of conventional software. This paper introduces the WPRiMA (Web Project Risk Management Assessment) as the tool, which is used to implement RIAP, the risk identification architecture pattern model, which focuses upon the data from the proprietor-s and vendor-s perspectives. The paper also illustrates how WPRiMA tool works and how it can be used to calculate the risk level for a given Web project, to generate recommendations in order to facilitate risk avoidance in a project, and to improve the prospects of early risk management.

Keywords: Architecture pattern model, risk factors, risk identification, web project, web project risk management assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581
2167 A Meta-Model for Tubercle Design of Wing Planforms Inspired by Humpback Whale Flippers

Authors: A. Taheri

Abstract:

Inspired by topology of humpback whale flippers, a meta-model is designed for wing planform design. The net is trained based on experimental data using cascade-forward artificial neural network (ANN) to investigate effects of the amplitude and wavelength of sinusoidal leading edge configurations on the wing performance. Afterwards, the trained ANN is coupled with a genetic algorithm method towards an optimum design strategy. Finally, flow physics of the problem for an optimized rectangular planform and also a real flipper geometry planform is simulated using Lam-Bremhorst low Reynolds number turbulence model with damping wall-functions resolving to the wall. Lift and drag coefficients and also details of flow are presented along with comparisons to available experimental data. Results show that the proposed strategy can be adopted with success as a fast-estimation tool for performance prediction of wing planforms with wavy leading edge at preliminary design phase.  

Keywords: Humpback whale flipper, cascade-forward ANN, GA, CFD, Bionics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3573
2166 An Optimization of Machine Parameters for Modified Horizontal Boring Tool Using Taguchi Method

Authors: Thirasak Panyaphirawat, Pairoj Sapsmarnwong, Teeratas Pornyungyuen

Abstract:

This paper presents the findings of an experimental investigation of important machining parameters for the horizontal boring tool modified to mouth with a horizontal lathe machine to bore an overlength workpiece. In order to verify a usability of a modified tool, design of experiment based on Taguchi method is performed. The parameters investigated are spindle speed, feed rate, depth of cut and length of workpiece. Taguchi L9 orthogonal array is selected for four factors three level parameters in order to minimize surface roughness (Ra and Rz) of S45C steel tubes. Signal to noise ratio analysis and analysis of variance (ANOVA) is performed to study an effect of said parameters and to optimize the machine setting for best surface finish. The controlled factors with most effect are depth of cut, spindle speed, length of workpiece, and feed rate in order. The confirmation test is performed to test the optimal setting obtained from Taguchi method and the result is satisfactory.

Keywords: Design of Experiment, Taguchi Design, Optimization, Analysis of Variance, Machining Parameters, Horizontal Boring Tool.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2701
2165 Novel Use of a Quality Assurance Tool for Integrating Technology to HSE

Authors: Ragi Poyyara, Vivek V., Ashish Khaparde

Abstract:

The product development process (PDP) in the Technology group plays a very important role in the launch of any product. While a manufacturing process encourages the use of certain measures to reduce health, safety and environmental (HSE) risks on the shop floor, the PDP concentrates on the use of Geometric Dimensioning and Tolerancing (GD&T) to develop a flawless design. Furthermore, PDP distributes and coordinates activities between different departments such as marketing, purchasing, and manufacturing. However, it is seldom realized that PDP makes a significant contribution to developing a product that reduces HSE risks by encouraging the Technology group to use effective GD&T. The GD&T is a precise communication tool that uses a set of symbols, rules, and definitions to mathematically define parts to be manufactured. It is a quality assurance method widely used in the oil and gas sector. Traditionally it is used to ensure the interchangeability of a part without affecting its form, fit, and function. Parts that do not meet these requirements are rejected during quality audits. This paper discusses how the Technology group integrates this quality assurance tool into the PDP and how the tool plays a major role in helping the HSE department in its goal towards eliminating HSE incidents. The PDP involves a thorough risk assessment and establishes a method to address those risks during the design stage. An illustration shows how GD&T helped reduce safety risks by ergonomically improving assembling operations. A brief discussion explains how tolerances provided on a part help prevent finger injury. This tool has equipped Technology to produce fixtures, which are used daily in operations as well as manufacturing. By applying GD&T to create good fits, HSE risks are mitigated for operating personnel. Both customers and service providers benefit from reduced safety risks.

Keywords: HSE, PDP, GD&T, risks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1822
2164 Ray Tracing Technique based 60 GHz Band Propagation Modelling and Influence of People Shadowing

Authors: A. Khafaji, R. Saadane, J. El Abbadi, M. Belkasmi

Abstract:

The main objectif of this paper is to present a tool that we have developed subject to characterize and modelling indoor radio channel propagation at millimetric wave. The tool is based on the ray tracing technique (RTT). As, in realistic environment we cannot neglect the significant impact of Human Body Shadowing and other objects in motion on indoor 60 GHz propagation channel. Hence, our proposed model allows a simulation of propagation in a dynamic indoor environment. First, we describe a model of human body. Second, RTT with this model is used to simulate the propagation of millimeter waves in the presence of persons in motion. Results of the simulation show that this tool gives results in agreement with those reported in the literature. Specially, the effects of people motion on temporal channel properties.

Keywords: Simulation. 60 GHz band, Ray Tracing Technique, Indoor channel, Propagation, Human Body Model, Level crossing rate,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2502
2163 Finite Element Simulation of Multi-Stage Deep Drawing Processes and Comparison with Experimental Results

Authors: A. Pourkamali Anaraki, M. Shahabizadeh, B. Babaee

Abstract:

The plastic forming process of sheet plate takes an important place in forming metals. The traditional techniques of tool design for sheet forming operations used in industry are experimental and expensive methods. Prediction of the forming results, determination of the punching force, blank holder forces and the thickness distribution of the sheet metal will decrease the production cost and time of the material to be formed. In this paper, multi-stage deep drawing simulation of an Industrial Part has been presented with finite element method. The entire production steps with additional operations such as intermediate annealing and springback has been simulated by ABAQUS software under axisymmetric conditions. The simulation results such as sheet thickness distribution, Punch force and residual stresses have been extracted in any stages and sheet thickness distribution was compared with experimental results. It was found through comparison of results, the FE model have proven to be in close agreement with those of experiment.

Keywords: Deep drawing, Finite element method, Simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5060
2162 Mining of Interesting Prediction Rules with Uniform Two-Level Genetic Algorithm

Authors: Bilal Alatas, Ahmet Arslan

Abstract:

The main goal of data mining is to extract accurate, comprehensible and interesting knowledge from databases that may be considered as large search spaces. In this paper, a new, efficient type of Genetic Algorithm (GA) called uniform two-level GA is proposed as a search strategy to discover truly interesting, high-level prediction rules, a difficult problem and relatively little researched, rather than discovering classification knowledge as usual in the literatures. The proposed method uses the advantage of uniform population method and addresses the task of generalized rule induction that can be regarded as a generalization of the task of classification. Although the task of generalized rule induction requires a lot of computations, which is usually not satisfied with the normal algorithms, it was demonstrated that this method increased the performance of GAs and rapidly found interesting rules.

Keywords: Classification rule mining, data mining, genetic algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1587
2161 A Statistical Prediction of Likely Distress in Nigeria Banking Sector Using a Neural Network Approach

Authors: D. A. Farinde

Abstract:

One of the most significant threats to the economy of a nation is the bankruptcy of its banks. This study evaluates the susceptibility of Nigerian banks to failure with a view to identifying ratios and financial data that are sensitive to solvency of the bank. Further, a predictive model is generated to guide all stakeholders in the industry. Thirty quoted banks that had published Annual Reports for the year preceding the consolidation i.e. year 2004 were selected. They were examined for distress using the Multilayer Perceptron Neural Network Analysis. The model was used to analyze further reforms by the Central Bank of Nigeria using published Annual Reports of twenty quoted banks for the year 2008 and 2011. The model can thus be used for future prediction of failure in the Nigerian banking system.

Keywords: Bank, Bankruptcy, Financial Ratios, Neural Network, Multilayer Perceptron, Predictive Model

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2690
2160 A New Technique for Solar Activity Forecasting Using Recurrent Elman Networks

Authors: Salvatore Marra, Francesco C. Morabito

Abstract:

In this paper we present an efficient approach for the prediction of two sunspot-related time series, namely the Yearly Sunspot Number and the IR5 Index, that are commonly used for monitoring solar activity. The method is based on exploiting partially recurrent Elman networks and it can be divided into three main steps: the first one consists in a “de-rectification" of the time series under study in order to obtain a new time series whose appearance, similar to a sum of sinusoids, can be modelled by our neural networks much better than the original dataset. After that, we normalize the derectified data so that they have zero mean and unity standard deviation and, finally, train an Elman network with only one input, a recurrent hidden layer and one output using a back-propagation algorithm with variable learning rate and momentum. The achieved results have shown the efficiency of this approach that, although very simple, can perform better than most of the existing solar activity forecasting methods.

Keywords: Elman neural networks, sunspot, solar activity, time series prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1846
2159 A Comparison between Hybrid and Experimental Extended Polars for the Numerical Prediction of Vertical-Axis Wind Turbine Performance using Blade Element-Momentum Algorithm

Authors: Gabriele Bedon, Marco Raciti Castelli, Ernesto Benini

Abstract:

A dynamic stall-corrected Blade Element-Momentum algorithm based on a hybrid polar is validated through the comparison with Sandia experimental measurements on a 5-m diameter wind turbine of Troposkien shape. Different dynamic stall models are evaluated. The numerical predictions obtained using the extended aerodynamic coefficients provided by both Sheldal and Klimas and Raciti Castelli et al. are compared to experimental data, determining the potential of the hybrid database for the numerical prediction of vertical-axis wind turbine performances.

Keywords: Darrieus wind turbine, Blade Element-Momentum Theory, extended airfoil database, hybrid database, Sandia 5-m wind turbine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2547
2158 System Reliability by Prediction of Generator Output and Losses in a Competitive Energy Market

Authors: Perumal Nallagownden, Ravindra N. Mukerjee, Syafrudin Masri

Abstract:

In a competitive energy market, system reliability should be maintained at all times. Power system operation being of online in nature, the energy balance requirements must be satisfied to ensure reliable operation the system. To achieve this, information regarding the expected status of the system, the scheduled transactions and the relevant inputs necessary to make either a transaction contract or a transmission contract operational, have to be made available in real time. The real time procedure proposed, facilitates this. This paper proposes a quadratic curve learning procedure, which enables a generator-s contribution to the retailer demand, power loss of transaction in a line at the retail end and its associated losses for an oncoming operating scenario to be predicted. Matlab program was used to test in on a 24-bus IEE Reliability Test System, and the results are found to be acceptable.

Keywords: Deregulation, learning coefficients, reliability, prediction, competitive energy market.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1470
2157 Intelligent Heart Disease Prediction System Using CANFIS and Genetic Algorithm

Authors: Latha Parthiban, R. Subramanian

Abstract:

Heart disease (HD) is a major cause of morbidity and mortality in the modern society. Medical diagnosis is an important but complicated task that should be performed accurately and efficiently and its automation would be very useful. All doctors are unfortunately not equally skilled in every sub specialty and they are in many places a scarce resource. A system for automated medical diagnosis would enhance medical care and reduce costs. In this paper, a new approach based on coactive neuro-fuzzy inference system (CANFIS) was presented for prediction of heart disease. The proposed CANFIS model combined the neural network adaptive capabilities and the fuzzy logic qualitative approach which is then integrated with genetic algorithm to diagnose the presence of the disease. The performances of the CANFIS model were evaluated in terms of training performances and classification accuracies and the results showed that the proposed CANFIS model has great potential in predicting the heart disease.

Keywords: CANFIS, genetic algorithms, heart disease, membership function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3970
2156 Statistical Assessment of Models for Determination of Soil – Water Characteristic Curves of Sand Soils

Authors: S. J. Matlan, M. Mukhlisin, M. R. Taha

Abstract:

Characterization of the engineering behavior of unsaturated soil is dependent on the soil-water characteristic curve (SWCC), a graphical representation of the relationship between water content or degree of saturation and soil suction. A reasonable description of the SWCC is thus important for the accurate prediction of unsaturated soil parameters. The measurement procedures for determining the SWCC, however, are difficult, expensive, and timeconsuming. During the past few decades, researchers have laid a major focus on developing empirical equations for predicting the SWCC, with a large number of empirical models suggested. One of the most crucial questions is how precisely existing equations can represent the SWCC. As different models have different ranges of capability, it is essential to evaluate the precision of the SWCC models used for each particular soil type for better SWCC estimation. It is expected that better estimation of SWCC would be achieved via a thorough statistical analysis of its distribution within a particular soil class. With this in view, a statistical analysis was conducted in order to evaluate the reliability of the SWCC prediction models against laboratory measurement. Optimization techniques were used to obtain the best-fit of the model parameters in four forms of SWCC equation, using laboratory data for relatively coarse-textured (i.e., sandy) soil. The four most prominent SWCCs were evaluated and computed for each sample. The result shows that the Brooks and Corey model is the most consistent in describing the SWCC for sand soil type. The Brooks and Corey model prediction also exhibit compatibility with samples ranging from low to high soil water content in which subjected to the samples that evaluated in this study.

Keywords: Soil-water characteristic curve (SWCC), statistical analysis, unsaturated soil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2655
2155 A Wall Law for Two-Phase Turbulent Boundary Layers

Authors: Dhahri Maher, Aouinet Hana

Abstract:

The presence of bubbles in the boundary layer introduces corrections into the log law, which must be taken into account. In this work, a logarithmic wall law was presented for bubbly two phase flows. The wall law presented in this work was based on the postulation of additional turbulent viscosity associated with bubble wakes in the boundary layer. The presented wall law contained empirical constant accounting both for shear induced turbulence interaction and for non-linearity of bubble. This constant was deduced from experimental data. The wall friction prediction achieved with the wall law was compared to the experimental data, in the case of a turbulent boundary layer developing on a vertical flat plate in the presence of millimetric bubbles. A very good agreement between experimental and numerical wall friction prediction was verified. The agreement was especially noticeable for the low void fraction when bubble induced turbulence plays a significant role.

Keywords: Bubbly flows, log law, boundary layer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1115
2154 An Artificial Neural Network Model for Earthquake Prediction and Relations between Environmental Parameters and Earthquakes

Authors: S. Niksarlioglu, F. Kulahci

Abstract:

Earthquakes are natural phenomena that occur with influence of a lot of parameters such as seismic activity, changing in the ground waters' motion, changing in the water-s temperature, etc. On the other hand, the radon gas concentrations in soil vary as nonlinear generally with earthquakes. Continuous measurement of the soil radon gas is very important for determination of characteristic of the seismic activity. The radon gas changes as continuous with strain occurring within the Earth-s surface during an earthquake and effects from the physical and the chemical processes such as soil structure, soil permeability, soil temperature, the barometric pressure, etc. Therefore, at the modeling researches are notsufficient to knowthe concentration ofradon gas. In this research, we determined relationships between radon emissions based on the environmental parameters and earthquakes occurring along the East Anatolian Fault Zone (EAFZ), Turkiye and predicted magnitudes of some earthquakes with the artificial neural network (ANN) model.

Keywords: Earthquake, Modeling, Prediction, Radon.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3001
2153 ANN Based Model Development for Material Removal Rate in Dry Turning in Indian Context

Authors: Mangesh R. Phate, V. H. Tatwawadi

Abstract:

This paper is intended to develop an artificial neural network (ANN) based model of material removal rate (MRR) in the turning of ferrous and nonferrous material in a Indian small-scale industry. MRR of the formulated model was proved with the testing data and artificial neural network (ANN) model was developed for the analysis and prediction of the relationship between inputs and output parameters during the turning of ferrous and nonferrous materials. The input parameters of this model are operator, work-piece, cutting process, cutting tool, machine and the environment.

The ANN model consists of a three layered feedforward back propagation neural network. The network is trained with pairs of independent/dependent datasets generated when machining ferrous and nonferrous material. A very good performance of the neural network, in terms of contract with experimental data, was achieved. The model may be used for the testing and forecast of the complex relationship between dependent and the independent parameters in turning operations.

Keywords: Field data based model, Artificial neural network, Simulation, Convectional Turning, Material removal rate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959
2152 Forecasting Stock Indexes Using Bayesian Additive Regression Tree

Authors: Darren Zou

Abstract:

Forecasting the stock market is a very challenging task. Various economic indicators such as GDP, exchange rates, interest rates, and unemployment have a substantial impact on the stock market. Time series models are the traditional methods used to predict stock market changes. In this paper, a machine learning method, Bayesian Additive Regression Tree (BART) is used in predicting stock market indexes based on multiple economic indicators. BART can be used to model heterogeneous treatment effects, and thereby works well when models are misspecified. It also has the capability to handle non-linear main effects and multi-way interactions without much input from financial analysts. In this research, BART is proposed to provide a reliable prediction on day-to-day stock market activities. By comparing the analysis results from BART and with time series method, BART can perform well and has better prediction capability than the traditional methods.

Keywords: Bayesian, Forecast, Stock, BART.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 720
2151 Eukaryotic Gene Prediction by an Investigation of Nonlinear Dynamical Modeling Techniques on EIIP Coded Sequences

Authors: Mai S. Mabrouk, Nahed H. Solouma, Abou-Bakr M. Youssef, Yasser M. Kadah

Abstract:

Many digital signal processing, techniques have been used to automatically distinguish protein coding regions (exons) from non-coding regions (introns) in DNA sequences. In this work, we have characterized these sequences according to their nonlinear dynamical features such as moment invariants, correlation dimension, and largest Lyapunov exponent estimates. We have applied our model to a number of real sequences encoded into a time series using EIIP sequence indicators. In order to discriminate between coding and non coding DNA regions, the phase space trajectory was first reconstructed for coding and non-coding regions. Nonlinear dynamical features are extracted from those regions and used to investigate a difference between them. Our results indicate that the nonlinear dynamical characteristics have yielded significant differences between coding (CR) and non-coding regions (NCR) in DNA sequences. Finally, the classifier is tested on real genes where coding and non-coding regions are well known.

Keywords: Gene prediction, nonlinear dynamics, correlation dimension, Lyapunov exponent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1808
2150 Lifelong Education for Teachers: A Tool for Achieving Effective Teaching and Learning in Secondary Schools in Benue State, Nigeria

Authors: P. I. Adzongo, O. A. Aloga

Abstract:

The purpose of the study was to examine lifelong education for teachers as a tool for achieving effective teaching and learning. Lifelong education enhances social inclusion, personal development, citizenship, employability, teaching and learning, community and the nation. It is imperative that the teacher needs to update his knowledge regularly to be able to perform optimally, since he has a major position in the inculcation of desirable elements in students, and the challenges of lifelong education were also discussed. Descriptive survey design was adopted for the study. A simple random sampling technique was used to select 80 teachers as sample from a population of 105 senior secondary school teachers in Makurdi Local Government Area of Benue State. A 20-item self designed questionnaire subjected to expert validation and reliability was used to collect data. The reliability Alpha coefficient of 0.87 was established using Cronbach’s Alpha technique, mean scores and standard deviation were used to answer the 2 research questions while chi-square was used to analyse data for the 2 null hypotheses, which states that lifelong education for teachers is not a significant tool for achieving effective teaching and lifelong education for teachers does not significantly impact on effective learning. The findings of the study revealed that, lifelong education for teachers can be used as a tool for achieving effective teaching and learning, and the study recommended among others that government, organizations and individuals should in collaboration put lifelong education programmes for teachers on the priority list. The paper concluded that the strategic position of lifelong education for teachers towards enhanced teaching, learning and the production of quality manpower in the society makes it imperative for all hands to be on “deck” to support the programme financially and otherwise.

Keywords: Lifelong Education, Tool, Effective Teaching and Learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1461
2149 An Enhanced Tool for Implementing Dialogue Forms in Conversational Applications

Authors: Ilias Spais, George Bafas

Abstract:

Natural Language Understanding Systems (NLU) will not be widely deployed unless they are technically mature and cost effective to develop. Cost effective development hinges on the availability of tools and techniques enabling the rapid production of NLU applications through minimal human resources. Further, these tools and techniques should allow quick development of applications in a user friendly way and should be easy to upgrade in order to continuously follow the evolving technologies and standards. This paper presents a visual tool for the structuring and editing of dialog forms, the key element of driving conversation in NLU applications based on IBM technology. The main focus is given on the basic component used to describe Human – Machine interactions of that kind, the Dialogue Manager. In essence, the description of a tool that enables the visual representation of the Dialogue Manager mainly during the implementation phase is illustrated.

Keywords: Conversational Applications, Forms Dialogue Manager (FDM), Natural Language Processing, Natural Language Understanding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1444
2148 Reliability Analysis of Underground Pipelines Using Subset Simulation

Authors: Kong Fah Tee, Lutfor Rahman Khan, Hongshuang Li

Abstract:

An advanced Monte Carlo simulation method, called Subset Simulation (SS) for the time-dependent reliability prediction for underground pipelines has been presented in this paper. The SS can provide better resolution for low failure probability level with efficient investigating of rare failure events which are commonly encountered in pipeline engineering applications. In SS method, random samples leading to progressive failure are generated efficiently and used for computing probabilistic performance by statistical variables. SS gains its efficiency as small probability event as a product of a sequence of intermediate events with larger conditional probabilities. The efficiency of SS has been demonstrated by numerical studies and attention in this work is devoted to scrutinise the robustness of the SS application in pipe reliability assessment. It is hoped that the development work can promote the use of SS tools for uncertainty propagation in the decision-making process of underground pipelines network reliability prediction.

Keywords: Underground pipelines, Probability of failure, Reliability and Subset Simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3539
2147 Discovering Complex Regularities: from Tree to Semi-Lattice Classifications

Authors: A. Faro, D. Giordano, F. Maiorana

Abstract:

Data mining uses a variety of techniques each of which is useful for some particular task. It is important to have a deep understanding of each technique and be able to perform sophisticated analysis. In this article we describe a tool built to simulate a variation of the Kohonen network to perform unsupervised clustering and support the entire data mining process up to results visualization. A graphical representation helps the user to find out a strategy to optimize classification by adding, moving or delete a neuron in order to change the number of classes. The tool is able to automatically suggest a strategy to optimize the number of classes optimization, but also support both tree classifications and semi-lattice organizations of the classes to give to the users the possibility of passing from one class to the ones with which it has some aspects in common. Examples of using tree and semi-lattice classifications are given to illustrate advantages and problems. The tool is applied to classify macroeconomic data that report the most developed countries- import and export. It is possible to classify the countries based on their economic behaviour and use the tool to characterize the commercial behaviour of a country in a selected class from the analysis of positive and negative features that contribute to classes formation. Possible interrelationships between the classes and their meaning are also discussed.

Keywords: Unsupervised classification, Kohonen networks, macroeconomics, Visual data mining, Cluster interpretation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1532
2146 Prediction the Deformation in Upsetting Process by Neural Network and Finite Element

Authors: H.Mohammadi Majd, M.Jalali Azizpour , Foad Saadi

Abstract:

In this paper back-propagation artificial neural network (BPANN) is employed to predict the deformation of the upsetting process. To prepare a training set for BPANN, some finite element simulations were carried out. The input data for the artificial neural network are a set of parameters generated randomly (aspect ratio d/h, material properties, temperature and coefficient of friction). The output data are the coefficient of polynomial that fitted on barreling curves. Neural network was trained using barreling curves generated by finite element simulations of the upsetting and the corresponding material parameters. This technique was tested for three different specimens and can be successfully employed to predict the deformation of the upsetting process

Keywords: Back-propagation artificial neural network(BPANN), prediction, upsetting

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1540
2145 A Reinforcement Learning Approach for Evaluation of Real-Time Disaster Relief Demand and Network Condition

Authors: Ali Nadi, Ali Edrissi

Abstract:

Relief demand and transportation links availability is the essential information that is needed for every natural disaster operation. This information is not in hand once a disaster strikes. Relief demand and network condition has been evaluated based on prediction method in related works. Nevertheless, prediction seems to be over or under estimated due to uncertainties and may lead to a failure operation. Therefore, in this paper a stochastic programming model is proposed to evaluate real-time relief demand and network condition at the onset of a natural disaster. To address the time sensitivity of the emergency response, the proposed model uses reinforcement learning for optimization of the total relief assessment time. The proposed model is tested on a real size network problem. The simulation results indicate that the proposed model performs well in the case of collecting real-time information.

Keywords: Disaster management, real-time demand, reinforcement learning, relief demand.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1921