Search results for: error minimization
1002 Regression Model Evaluation on Depth Camera Data for Gaze Estimation
Authors: James Purnama, Riri Fitri Sari
Abstract:
We investigate the machine learning algorithm selection problem in the term of a depth image based eye gaze estimation, with respect to its essential difficulty in reducing the number of required training samples and duration time of training. Statistics based prediction accuracy are increasingly used to assess and evaluate prediction or estimation in gaze estimation. This article evaluates Root Mean Squared Error (RMSE) and R-Squared statistical analysis to assess machine learning methods on depth camera data for gaze estimation. There are 4 machines learning methods have been evaluated: Random Forest Regression, Regression Tree, Support Vector Machine (SVM), and Linear Regression. The experiment results show that the Random Forest Regression has the lowest RMSE and the highest R-Squared, which means that it is the best among other methods.Keywords: gaze estimation, gaze tracking, eye tracking, kinect, regression model, orange python
Procedia PDF Downloads 5401001 Thin-Layer Drying Characteristics and Modelling of Instant Coffee Solution
Authors: Apolinar Picado, Ronald Solís, Rafael Gamero
Abstract:
The thin-layer drying characteristics of instant coffee solution were investigated in a laboratory tunnel dryer. Drying experiments were carried out at three temperatures (80, 100 and 120 °C) and an air velocity of 1.2 m/s. Drying experimental data obtained are fitted to six (6) thin-layer drying models using the non-linear least squares regression analysis. The acceptability of the thin-layer drying model has been based on a value of the correlation coefficient that should be close to one, and low values for root mean square error (RMSE) and chi-square (x²). According to this evaluation, the most suitable model for describing drying process of thin-layer instant coffee solution is the Page model. Further, the effective moisture diffusivity and the activation energy were computed employing the drying experimental data. The effective moisture diffusivity values varied from 1.6133 × 10⁻⁹ to 1.6224 × 10⁻⁹ m²/s over the temperature range studied and the activation energy was estimated to be 162.62 J/mol.Keywords: activation energy, diffusivity, instant coffee, thin-layer models
Procedia PDF Downloads 2641000 An Approach on Robust Multi Inversion of a Nonlinear Model for an Omni-Directional Mobile
Authors: Fernando P. Silva, Valter J. S. Leite, Erivelton G. Nepomuceno
Abstract:
In this paper, a nonlinear controller design for an omnidirectional mobile is presented. The robot controller consists of an inner-loop controller and an outer-loop controller, the first is designed using state feedback (robust allocation) and the second controller is designed based on Robust Multi Inversion (RMI) approach. The objective of RMI controller is rendering the robust inversion of the dynamic, when the model is affected by uncertainties. A model nonlinear MIMO of an omni-directional robot (small-league of Robocup) is used to simulate the RMI approach. The parameters of linear and nonlinear model are varied to cause modelling uncertainties among the model and the real model (real system) generating an error in inner-loop controller signal that must be compensated by RMI controller. The simulation test results show that the RMI is capable of compensating the uncertainties and keep the system stable and controlled under uncertainties.Keywords: robust multi inversion, omni-directional robot, robocup, nonlinear control
Procedia PDF Downloads 591999 Prediction of the Transmittance of Various Bended Angles Lightpipe by Using Neural Network under Different Sky Clearness Condition
Authors: Li Zhang, Yuehong Su
Abstract:
Lightpipe as a mature solar light tube technique has been employed worldwide. Accurately assessing the performance of lightpipe and evaluate daylighting available has been a challenging topic. Previous research had used regression model and computational simulation methods to estimate the performance of lightpipe. However, due to the nonlinear nature of solar light transferring in lightpipe, the methods mentioned above express inaccurate and time-costing issues. In the present study, a neural network model as an alternative method is investigated to predict the transmittance of lightpipe. Four types of commercial lightpipe with bended angle 0°, 30°, 45° and 60° are discussed under clear, intermediate and overcast sky conditions respectively. The neural network is generated in MATLAB by using the outcomes of an optical software Photopia simulations as targets for networks training and testing. The coefficient of determination (R²) for each model is higher than 0.98, and the mean square error (MSE) is less than 0.0019, which indicate the neural network strong predictive ability and the use of the neural network method could be an efficient technique for determining the performance of lightpipe.Keywords: neural network, bended lightpipe, transmittance, Photopia
Procedia PDF Downloads 155998 Mobile Robot Manipulator Kinematics Motion Control Analysis with MATLAB/Simulink
Authors: Wayan Widhiada, Cok Indra Partha, Gusti Ngurah Nitya Santhiarsa
Abstract:
The purpose of this paper is to investigate the sophistication of the use of Proportional Integral and Derivative Control to control the kinematic motion of the mobile robot manipulator. Simulation and experimental methods will be used to investigate the sophistication of PID control to control the mobile robot arm in the collection and placement of several kinds of objects quickly, accurately and correctly. Mathematical modeling will be done by utilizing the integration of Solidworks and MATLAB / Simmechanics software. This method works by converting the physical model file into the xml file. This method is easy, fast and accurate done in modeling and design robotics. The automatic control design of this robot manipulator will be validated in simulations and experimental in control labs as evidence that the mobile robot manipulator gripper control design can achieve the best performance such as the error signal is lower than 5%, small overshoot and get steady signal response as quickly.Keywords: control analysis, kinematics motion, mobile robot manipulator, performance
Procedia PDF Downloads 411997 A Convolution Neural Network PM-10 Prediction System Based on a Dense Measurement Sensor Network in Poland
Authors: Piotr A. Kowalski, Kasper Sapala, Wiktor Warchalowski
Abstract:
PM10 is a suspended dust that primarily has a negative effect on the respiratory system. PM10 is responsible for attacks of coughing and wheezing, asthma or acute, violent bronchitis. Indirectly, PM10 also negatively affects the rest of the body, including increasing the risk of heart attack and stroke. Unfortunately, Poland is a country that cannot boast of good air quality, in particular, due to large PM concentration levels. Therefore, based on the dense network of Airly sensors, it was decided to deal with the problem of prediction of suspended particulate matter concentration. Due to the very complicated nature of this issue, the Machine Learning approach was used. For this purpose, Convolution Neural Network (CNN) neural networks have been adopted, these currently being the leading information processing methods in the field of computational intelligence. The aim of this research is to show the influence of particular CNN network parameters on the quality of the obtained forecast. The forecast itself is made on the basis of parameters measured by Airly sensors and is carried out for the subsequent day, hour after hour. The evaluation of learning process for the investigated models was mostly based upon the mean square error criterion; however, during the model validation, a number of other methods of quantitative evaluation were taken into account. The presented model of pollution prediction has been verified by way of real weather and air pollution data taken from the Airly sensor network. The dense and distributed network of Airly measurement devices enables access to current and archival data on air pollution, temperature, suspended particulate matter PM1.0, PM2.5, and PM10, CAQI levels, as well as atmospheric pressure and air humidity. In this investigation, PM2.5, and PM10, temperature and wind information, as well as external forecasts of temperature and wind for next 24h served as inputted data. Due to the specificity of the CNN type network, this data is transformed into tensors and then processed. This network consists of an input layer, an output layer, and many hidden layers. In the hidden layers, convolutional and pooling operations are performed. The output of this system is a vector containing 24 elements that contain prediction of PM10 concentration for the upcoming 24 hour period. Over 1000 models based on CNN methodology were tested during the study. During the research, several were selected out that give the best results, and then a comparison was made with the other models based on linear regression. The numerical tests carried out fully confirmed the positive properties of the presented method. These were carried out using real ‘big’ data. Models based on the CNN technique allow prediction of PM10 dust concentration with a much smaller mean square error than currently used methods based on linear regression. What's more, the use of neural networks increased Pearson's correlation coefficient (R²) by about 5 percent compared to the linear model. During the simulation, the R² coefficient was 0.92, 0.76, 0.75, 0.73, and 0.73 for 1st, 6th, 12th, 18th, and 24th hour of prediction respectively.Keywords: air pollution prediction (forecasting), machine learning, regression task, convolution neural networks
Procedia PDF Downloads 150996 Advanced Deployable/Retractable Solar Panel System for Satellite Applications
Authors: Zane Brough, Claudio Paoloni
Abstract:
Modern low earth orbit (LEO) satellites that require multi-mission flexibility are highly likely to be repositioned between different operational orbits. While executing this process the satellite may experience high levels of vibration and environmental hazards, exposing the deployed solar panel to dangerous stress levels, fatigue and space debris, hence it is desirable to retract the solar array before satellite repositioning to avoid damage or failure. Furthermore, to accommodate for today's technological world, the power demand of a modern LEO satellite is rapidly increasing, which consequently provides pressure upon the design of the satellites solar array system to conform to the strict volume and mass limitations. A novel concept of deployable/retractable hybrid solar array system, aimed to provide a greater power to volume ratio while dramatically reducing the disadvantages of system mass and cost is proposed. Taking advantage of the new lightweight technology in solar panels, a mechanical system composed of both rigid and flexible solar panels arranged within a petal formation is proposed to yield a stowed to deployment area ratio up to at least 1:7, which improves the power density dramatically. The system consists of five subsystems, the outer ones based on a novel eight-petal configuration that provides a large surface and supports the flexible solar panels. A single cable and spool based hinge mechanism were designed to synchronously deploy/retract the panels in a safe, simple and efficient manner while the mass compared to the previous systems is considerably reduced. The relevant challenge to assure a smooth movement is resolved by a proper minimization of the gearing system and the use of a micro-controller system. A prototype was designed by 3D simulators and successfully constructed and tested. Further design works are in progress to implement an epicyclical gear hinge mechanism, which will further reduce the volume, mass and complexity of the system significantly. The proposed system due to an effective and reliable mechanism provides a large active surface, whilst being very compact. It could be extremely advantageous for use as ground portable solar panel system.Keywords: mechatronic engineering, satellite, solar panel, deployable/retractable mechanism
Procedia PDF Downloads 381995 An Efficient Collocation Method for Solving the Variable-Order Time-Fractional Partial Differential Equations Arising from the Physical Phenomenon
Authors: Haniye Dehestani, Yadollah Ordokhani
Abstract:
In this work, we present an efficient approach for solving variable-order time-fractional partial differential equations, which are based on Legendre and Laguerre polynomials. First, we introduced the pseudo-operational matrices of integer and variable fractional order of integration by use of some properties of Riemann-Liouville fractional integral. Then, applied together with collocation method and Legendre-Laguerre functions for solving variable-order time-fractional partial differential equations. Also, an estimation of the error is presented. At last, we investigate numerical examples which arise in physics to demonstrate the accuracy of the present method. In comparison results obtained by the present method with the exact solution and the other methods reveals that the method is very effective.Keywords: collocation method, fractional partial differential equations, legendre-laguerre functions, pseudo-operational matrix of integration
Procedia PDF Downloads 169994 Reducing Later Life Loneliness: A Systematic Literature Review of Loneliness Interventions
Authors: Dhruv Sharma, Lynne Blair, Stephen Clune
Abstract:
Later life loneliness is a social issue that is increasing alongside an upward global population trend. As a society, one way that we have responded to this social challenge is through developing non-pharmacological interventions such as befriending services, activity clubs, meet-ups, etc. Through a systematic literature review, this paper suggests that currently there is an underrepresentation of radical innovation, and underutilization of digital technologies in developing loneliness interventions for older adults. This paper examines intervention studies that were published in English language, within peer reviewed journals between January 2005 and December 2014 across 4 electronic databases. In addition to academic databases, interventions found in grey literature in the form of websites, blogs, and Twitter were also included in the overall review. This approach yielded 129 interventions that were included in the study. A systematic approach allowed the minimization of any bias dictating the selection of interventions to study. A coding strategy based on a pattern analysis approach was devised to be able to compare and contrast the loneliness interventions. Firstly, interventions were categorized on the basis of their objective to identify whether they were preventative, supportive, or remedial in nature. Secondly, depending on their scope, they were categorized as one-to-one, community-based, or group based. It was also ascertained whether interventions represented an improvement, an incremental innovation, a major advance or a radical departure, in comparison to the most basic form of a loneliness intervention. Finally, interventions were also assessed on the basis of the extent to which they utilized digital technologies. Individual visualizations representing the four levels of coding were created for each intervention, followed by an aggregated visual to facilitate analysis. To keep the inquiry within scope and to present a coherent view of the findings, the analysis was primarily concerned the level of innovation, and the use of digital technologies. This analysis highlights a weak but positive correlation between the level of innovation and the use of digital technologies in designing and deploying loneliness interventions, and also emphasizes how certain existing interventions could be tweaked to enable their migration from representing incremental innovation to radical innovation for example. This analysis also points out the value of including grey literature, especially from Twitter, in systematic literature reviews to get a contemporary view of latest work in the area under investigation.Keywords: ageing, loneliness, innovation, digital
Procedia PDF Downloads 125993 Metrology-Inspired Methods to Assess the Biases of Artificial Intelligence Systems
Authors: Belkacem Laimouche
Abstract:
With the field of artificial intelligence (AI) experiencing exponential growth, fueled by technological advancements that pave the way for increasingly innovative and promising applications, there is an escalating need to develop rigorous methods for assessing their performance in pursuit of transparency and equity. This article proposes a metrology-inspired statistical framework for evaluating bias and explainability in AI systems. Drawing from the principles of metrology, we propose a pioneering approach, using a concrete example, to evaluate the accuracy and precision of AI models, as well as to quantify the sources of measurement uncertainty that can lead to bias in their predictions. Furthermore, we explore a statistical approach for evaluating the explainability of AI systems based on their ability to provide interpretable and transparent explanations of their predictions.Keywords: artificial intelligence, metrology, measurement uncertainty, prediction error, bias, machine learning algorithms, probabilistic models, interlaboratory comparison, data analysis, data reliability, measurement of bias impact on predictions, improvement of model accuracy and reliability
Procedia PDF Downloads 107992 Off Design Modelling of 650MW Combined Cycle Gas Turbine Power Plant Integrated with a Retrofitted Inlet Fogging System
Authors: Osarobo Omorogieva Ighodaro, Josephus Otejere
Abstract:
This paper contains the modelling and simulation of GT13E2 combined cycle gas turbine with the aid of the software EBSILON PROFESSIONAL. The design mode was modeled using guaranteed performance data from the power plant, in the off design, temperature variation of ambient air and fogging (spray water at inlet to compressor) was simulated. The fogging was simulated under two different modes; constant fuel consumption and constant turbine exhaust temperature .The model results were validated using actual operating data by applying error percentage analysis. The validation results obtained ranged from -0.0038% to 0% in design condition while the results varied from -0.9202% to 10.24% The model shows that fogging decreases compressor inlet temperature which in turn decreases the power required to drive the compressor hence improving the simple cycle efficiency and hence increasing power generated.Keywords: inlet fogging, off design, combined cycle, modelling
Procedia PDF Downloads 44991 Investigation on Mesh Sensitivity of a Transient Model for Nozzle Clogging
Authors: H. Barati, M. Wu, A. Kharicha, A. Ludwig
Abstract:
A transient model for nozzle clogging has been developed and successfully validated against a laboratory experiment. Key steps of clogging are considered: transport of particles by turbulent flow towards the nozzle wall; interactions between fluid flow and nozzle wall, and the adhesion of the particle on the wall; the growth of the clog layer and its interaction with the flow. The current paper is to investigate the mesh (size and type) sensitivity of the model in both two and three dimensions. It is found that the algorithm for clog growth alone excluding the flow effect is insensitive to the mesh type and size, but the calculation including flow becomes sensitive to the mesh quality. The use of 2D meshes leads to overestimation of the clog growth because the 3D nature of flow in the boundary layer cannot be properly solved by 2D calculation. 3D simulation with tetrahedron mesh can also lead to an error estimation of the clog growth. A mesh-independent result can be achieved with hexahedral mesh, or at least with triangular prism (inflation layer) for near-wall regions.Keywords: clogging, continuous casting, inclusion, simulation, submerged entry nozzle
Procedia PDF Downloads 286990 Water Leakage Detection System of Pipe Line using Radial Basis Function Neural Network
Authors: A. Ejah Umraeni Salam, M. Tola, M. Selintung, F. Maricar
Abstract:
Clean water is an essential and fundamental human need. Therefore, its supply must be assured by maintaining the quality, quantity and water pressure. However the fact is, on its distribution system, leakage happens and becomes a common world issue. One of the technical causes of the leakage is a leaking pipe. The purpose of the research is how to use the Radial Basis Function Neural (RBFNN) model to detect the location and the magnitude of the pipeline leakage rapidly and efficiently. In this study the RBFNN are trained and tested on data from EPANET hydraulic modeling system. Method of Radial Basis Function Neural Network is proved capable to detect location and magnitude of pipeline leakage with of the accuracy of the prediction results based on the value of RMSE (Root Meant Square Error), comparison prediction and actual measurement approaches 0.000049 for the whole pipeline system.Keywords: radial basis function neural network, leakage pipeline, EPANET, RMSE
Procedia PDF Downloads 361989 Numerical Solution of Porous Media Equation Using Jacobi Operational Matrix
Authors: Shubham Jaiswal
Abstract:
During modeling of transport phenomena in porous media, many nonlinear partial differential equations (NPDEs) encountered which greatly described the convection, diffusion and reaction process. To solve such types of nonlinear problems, a reliable and efficient technique is needed. In this article, the numerical solution of NPDEs encountered in porous media is derived. Here Jacobi collocation method is used to solve the considered problems which convert the NPDEs in systems of nonlinear algebraic equations that can be solved using Newton-Raphson method. The numerical results of some illustrative examples are reported to show the efficiency and high accuracy of the proposed approach. The comparison of the numerical results with the existing analytical results already reported in the literature and the error analysis for each example exhibited through graphs and tables confirms the exponential convergence rate of the proposed method.Keywords: nonlinear porous media equation, shifted Jacobi polynomials, operational matrix, spectral collocation method
Procedia PDF Downloads 443988 Dynamic Thermal Modelling of a PEMFC-Type Fuel Cell
Authors: Marco Avila Lopez, Hasnae Ait-Douchi, Silvia De Los Santos, Badr Eddine Lebrouhi, Pamela Ramírez Vidal
Abstract:
In the context of the energy transition, fuel cell technology has emerged as a solution for harnessing hydrogen energy and mitigating greenhouse gas emissions. An in-depth study was conducted on a PEMFC-type fuel cell, with an initiation of an analysis of its operational principles and constituent components. Subsequently, the modelling of the fuel cell was undertaken using the Python programming language, encompassing both steady-state and transient regimes. In the case of the steady-state regime, the physical and electrochemical phenomena occurring within the fuel cell were modelled, with the assumption of uniform temperature throughout all cell compartments. Parametric identification was carried out, resulting in a remarkable mean error of only 1.62% when the model results were compared to experimental data documented in the literature. The dynamic model that was developed enabled the scrutiny of the fuel cell's response in terms of temperature and voltage under varying current conditions.Keywords: fuel cell, modelling, dynamic, thermal model, PEMFC
Procedia PDF Downloads 84987 Suitable Die Shaping for a Rectangular Shape Bottle by Application of FEM and AI Technique
Authors: N. Ploysook, R. Rugsaj, C. Suvanjumrat
Abstract:
The characteristic requirement for producing rectangular shape bottles was a uniform thickness of the plastic bottle wall. Die shaping was a good technique which controlled the wall thickness of bottles. An advance technology which was the finite element method (FEM) for blowing parison to be a rectangular shape bottle was conducted to reduce waste plastic from a trial and error method of a die shaping and parison control method. The artificial intelligent (AI) comprised of artificial neural network and genetic algorithm was selected to optimize the die gap shape from the FEM results. The application of AI technique could optimize the suitable die gap shape for the parison blow molding which did not depend on the parison control method to produce rectangular bottles with the uniform wall. Particularly, this application can be used with cheap blow molding machines without a parison controller therefore it will reduce cost of production in the bottle blow molding process.Keywords: AI, bottle, die shaping, FEM
Procedia PDF Downloads 240986 Intrusion Detection Using Dual Artificial Techniques
Authors: Rana I. Abdulghani, Amera I. Melhum
Abstract:
With the abnormal growth of the usage of computers over networks and under the consideration or agreement of most of the computer security experts who said that the goal of building a secure system is never achieved effectively, all these points led to the design of the intrusion detection systems(IDS). This research adopts a comparison between two techniques for network intrusion detection, The first one used the (Particles Swarm Optimization) that fall within the field (Swarm Intelligence). In this Act, the algorithm Enhanced for the purpose of obtaining the minimum error rate by amending the cluster centers when better fitness function is found through the training stages. Results show that this modification gives more efficient exploration of the original algorithm. The second algorithm used a (Back propagation NN) algorithm. Finally a comparison between the results of two methods used were based on (NSL_KDD) data sets for the construction and evaluation of intrusion detection systems. This research is only interested in clustering the two categories (Normal and Abnormal) for the given connection records. Practices experiments result in intrude detection rate (99.183818%) for EPSO and intrude detection rate (69.446416%) for BP neural network.Keywords: IDS, SI, BP, NSL_KDD, PSO
Procedia PDF Downloads 384985 A Syntactic Errors Analysis in the Malaysian ESL Learners' Written Composition
Authors: Annie Gedion, Johan Severinus Tati, Jacinta Caroline Peter
Abstract:
Syntax error analysis studies have a significant role in English language teaching especially in the second language. This study investigates the syntax errors in written composition by 50 multilingual ESL learners in Politeknik Kota Kinabalu Sabah, Malaysia. The subjects speak their own dialect, Malay as their second language and English as their third or foreign language. Data were collected from the written discourse in the form of descriptive essays. The subjects were asked to write in the classroom within 45 minutes. 15 categories of errors were classified into a set of syntactic categories and were analysed based on the five steps of the syntactic analysis procedure. The findings of the study showed that the mother tongue interference, as well as lack of vocabulary and grammar knowledge, were the major sources of syntax errors in the learners’ written composition. Learners should be exposed to the differentiation of Malay and English grammar to avoid interference and effective learning of second language writing.Keywords: errors analysis, syntactic analysis, English as a second language, ESL writing
Procedia PDF Downloads 287984 Design of a Pneumonia Ontology for Diagnosis Decision Support System
Authors: Sabrina Azzi, Michal Iglewski, Véronique Nabelsi
Abstract:
Diagnosis error problem is frequent and one of the most important safety problems today. One of the main objectives of our work is to propose an ontological representation that takes into account the diagnostic criteria in order to improve the diagnostic. We choose pneumonia disease since it is one of the frequent diseases affected by diagnosis errors and have harmful effects on patients. To achieve our aim, we use a semi-automated method to integrate diverse knowledge sources that include publically available pneumonia disease guidelines from international repositories, biomedical ontologies and electronic health records. We follow the principles of the Open Biomedical Ontologies (OBO) Foundry. The resulting ontology covers symptoms and signs, all the types of pneumonia, antecedents, pathogens, and diagnostic testing. The first evaluation results show that most of the terms are covered by the ontology. This work is still in progress and represents a first and major step toward a development of a diagnosis decision support system for pneumonia.Keywords: Clinical decision support system, Diagnostic errors, Ontology, Pneumonia
Procedia PDF Downloads 190983 Neural Network Based Compressor Flow Estimator in an Aircraft Vapor Cycle System
Authors: Justin Reverdi, Sixin Zhang, Serge Gratton, Said Aoues, Thomas Pellegrini
Abstract:
In Vapor Cycle Systems, the flow sensor plays a key role in different monitoring and control purposes. However, physical sensors can be expensive, inaccurate, heavy, cumbersome, or highly sensitive to vibrations, which is especially problematic when embedded into an aircraft. The conception of a virtual sensor based on other standard sensors is a good alternative. In this paper, a data-driven model using a Convolutional Neural Network is proposed to estimate the flow of the compressor. To fit the model to our dataset, we tested different loss functions. We show in our application that a Dynamic Time Warping based loss function called DILATE leads to better dynamical performance than the vanilla mean squared error (MSE) loss function. DILATE allows choosing a trade-off between static and dynamic performance.Keywords: deep learning, dynamic time warping, vapor cycle system, virtual sensor
Procedia PDF Downloads 148982 Modeling Sediment Yield Using the SWAT Model: A Case Study of Upper Ankara River Basin, Turkey
Authors: Umit Duru
Abstract:
The Soil and Water Assessment Tool (SWAT) was tested for prediction of water balance and sediment yield in the Ankara gauged basin, Turkey. The overall objective of this study was to evaluate the performance and applicability of the SWAT in this region of Turkey. Thirteen years of monthly stream flow, and suspended sediment, data were used for calibration and validation. This research assessed model performance based on differences between observed and predicted suspended sediment yield during calibration (1987-1996) and validation (1982-1984) periods. Statistical comparisons of suspended sediment produced values for NSE (Nash Sutcliffe efficiency), RE (relative error), and R² (coefficient of determination), of 0.81, -1.55, and 0.93, respectively, during the calibration period, and NSE, RE (%), and R² of 0.77, -2.61, and 0.87, respectively, during the validation period. Based on the analyses, SWAT satisfactorily simulated observed hydrology and sediment yields and can be used as a tool in decision making for water resources planning and management in the basin.Keywords: calibration, GIS, sediment yield, SWAT, validation
Procedia PDF Downloads 284981 Housing Price Prediction Using Machine Learning Algorithms: The Case of Melbourne City, Australia
Authors: The Danh Phan
Abstract:
House price forecasting is a main topic in the real estate market research. Effective house price prediction models could not only allow home buyers and real estate agents to make better data-driven decisions but may also be beneficial for the property policymaking process. This study investigates the housing market by using machine learning techniques to analyze real historical house sale transactions in Australia. It seeks useful models which could be deployed as an application for house buyers and sellers. Data analytics show a high discrepancy between the house price in the most expensive suburbs and the most affordable suburbs in the city of Melbourne. In addition, experiments demonstrate that the combination of Stepwise and Support Vector Machine (SVM), based on the Mean Squared Error (MSE) measurement, consistently outperforms other models in terms of prediction accuracy.Keywords: house price prediction, regression trees, neural network, support vector machine, stepwise
Procedia PDF Downloads 234980 Modeling the Demand for the Healthcare Services Using Data Analysis Techniques
Authors: Elizaveta S. Prokofyeva, Svetlana V. Maltseva, Roman D. Zaitsev
Abstract:
Rapidly evolving modern data analysis technologies in healthcare play a large role in understanding the operation of the system and its characteristics. Nowadays, one of the key tasks in urban healthcare is to optimize the resource allocation. Thus, the application of data analysis in medical institutions to solve optimization problems determines the significance of this study. The purpose of this research was to establish the dependence between the indicators of the effectiveness of the medical institution and its resources. Hospital discharges by diagnosis; hospital days of in-patients and in-patient average length of stay were selected as the performance indicators and the demand of the medical facility. The hospital beds by type of care, medical technology (magnetic resonance tomography, gamma cameras, angiographic complexes and lithotripters) and physicians characterized the resource provision of medical institutions for the developed models. The data source for the research was an open database of the statistical service Eurostat. The choice of the source is due to the fact that the databases contain complete and open information necessary for research tasks in the field of public health. In addition, the statistical database has a user-friendly interface that allows you to quickly build analytical reports. The study provides information on 28 European for the period from 2007 to 2016. For all countries included in the study, with the most accurate and complete data for the period under review, predictive models were developed based on historical panel data. An attempt to improve the quality and the interpretation of the models was made by cluster analysis of the investigated set of countries. The main idea was to assess the similarity of the joint behavior of the variables throughout the time period under consideration to identify groups of similar countries and to construct the separate regression models for them. Therefore, the original time series were used as the objects of clustering. The hierarchical agglomerate algorithm k-medoids was used. The sampled objects were used as the centers of the clusters obtained, since determining the centroid when working with time series involves additional difficulties. The number of clusters used the silhouette coefficient. After the cluster analysis it was possible to significantly improve the predictive power of the models: for example, in the one of the clusters, MAPE error was only 0,82%, which makes it possible to conclude that this forecast is highly reliable in the short term. The obtained predicted values of the developed models have a relatively low level of error and can be used to make decisions on the resource provision of the hospital by medical personnel. The research displays the strong dependencies between the demand for the medical services and the modern medical equipment variable, which highlights the importance of the technological component for the successful development of the medical facility. Currently, data analysis has a huge potential, which allows to significantly improving health services. Medical institutions that are the first to introduce these technologies will certainly have a competitive advantage.Keywords: data analysis, demand modeling, healthcare, medical facilities
Procedia PDF Downloads 146979 Hydraulic Analysis of Irrigation Approach Channel Using HEC-RAS Model
Authors: Muluegziabher Semagne Mekonnen
Abstract:
This study was intended to show the irrigation water requirements and evaluation of canal hydraulics steady state conditions to improve on scheme performance of the Meki-Ziway irrigation project. The methodology used was the CROPWAT 8.0 model to estimate the irrigation water requirements of five major crops irrigated in the study area. The results showed that for the whole existing and potential irrigation development area of 2000 ha and 2599 ha, crop water requirements were 3,339,200 and 4,339,090.4 m³, respectively. Hydraulic simulation models are fundamental tools for understanding the hydraulic flow characteristics of irrigation systems. Hydraulic simulation models are fundamental tools for understanding the hydraulic flow characteristics of irrigation systems. In this study Hydraulic Analysis of Irrigation Canals Using HEC-RAS Model was conducted in Meki-Ziway Irrigation Scheme. The HEC-RAS model was tested in terms of error estimation and used to determine canal capacity potential.Keywords: HEC-RAS, irrigation, hydraulic. canal reach, capacity
Procedia PDF Downloads 63978 A Study of Islamic Stock Indices and Macroeconomic Variables
Authors: Mohammad Irfan
Abstract:
The purpose of this paper is to investigate the relationship among the key macroeconomic variables and Islamic stock market in India. This study is based on the time series data of financial years 2009-2015 to explore the consistency of relationship between macroeconomic variables and Shariah Indices. The ADF (Augmented Dickey–Fuller Test Statistic) and PP (Phillips–Perron Test Statistic) tests are employed to check stationarity of the data. The study depicts the long run relationship between Shariah indices and macroeconomic variables by using the Johansen Co-integration test. BSE Shariah and Nifty Shariah have uni-direct Granger causality. The outcome of VECM is significantly confirming the applicability of best fitted model. Thus, Islamic stock indices are proficiently working for the development of Indian economy. It suggests that by keeping eyes on Islamic stock market which will be more interactive in the future with other macroeconomic variables.Keywords: Indian Shariah Indices, macroeconomic variables, co-integration, Granger causality, vector error correction model (VECM)
Procedia PDF Downloads 281977 Investigation of Roll-Off Factor in Pulse Shaping Filter on Maximal Ratio Combining for CDMA 2000 System
Authors: G. S. Walia, H. P. Singh, D. Padma
Abstract:
The integration of wide variety of communication services is made possible with invention of 3G technology. Code Division Multiple Access 2000 operates on various RF channel bandwidths 1.2288 or 3.6864 Mcps (1x or 3x systems). It is a 3G system which offers high bandwidth and wireless broadband services but its efficiency is lowered due to various factors like fading, interference, scattering, absorption etc. This paper investigates the effect of diversity (MRC), roll off factor in Root Raised Cosine (RRC) filter for the BPSK and QPSK modulation schemes. It is possible to transmit data with minimum Inter symbol Interference and within limited bandwidth with proper pulse shaping technique. Bit error rate (BER) performance is analyzed by applying diversity technique by varying the roll off factor for BPSK and QPSK. Roll off factor reduces the ISI and diversity reduces the Fading.Keywords: CDMA2000, root raised cosine, roll-off factor, ISI, diversity, interference, fading
Procedia PDF Downloads 409976 Leverage Effect for Volatility with Generalized Laplace Error
Authors: Farrukh Javed, Krzysztof Podgórski
Abstract:
We propose a new model that accounts for the asymmetric response of volatility to positive ('good news') and negative ('bad news') shocks in economic time series the so-called leverage effect. In the past, asymmetric powers of errors in the conditionally heteroskedastic models have been used to capture this effect. Our model is using the gamma difference representation of the generalized Laplace distributions that efficiently models the asymmetry. It has one additional natural parameter, the shape, that is used instead of power in the asymmetric power models to capture the strength of a long-lasting effect of shocks. Some fundamental properties of the model are provided including the formula for covariances and an explicit form for the conditional distribution of 'bad' and 'good' news processes given the past the property that is important for the statistical fitting of the model. Relevant features of volatility models are illustrated using S&P 500 historical data.Keywords: heavy tails, volatility clustering, generalized asymmetric laplace distribution, leverage effect, conditional heteroskedasticity, asymmetric power volatility, GARCH models
Procedia PDF Downloads 387975 Granger Causal Nexus between Financial Development and Energy Consumption: Evidence from Cross Country Panel Data
Authors: Rudra P. Pradhan
Abstract:
This paper examines the Granger causal nexus between financial development and energy consumption in the group of 35 Financial Action Task Force (FATF) Countries over the period 1988-2012. The study uses two financial development indicators such as private sector credit and stock market capitalization and seven energy consumption indicators such as coal, oil, gas, electricity, hydro-electrical, nuclear and biomass. Using panel cointegration tests, the study finds that financial development and energy consumption are cointegrated, indicating the presence of a long-run relationship between the two. Using a panel vector error correction model (VECM), the study detects both bidirectional and unidirectional causality between financial development and energy consumption. The variation of this causality is due to the use of different proxies for both financial development and energy consumption. The policy implication of this study is that economic policies should recognize the differences in the financial development-energy consumption nexus in order to maintain sustainable development in the selected 35 FATF countries.Keywords: energy consumption, financial development, FATF countries, Panel VECM
Procedia PDF Downloads 269974 Software Verification of Systematic Resampling for Optimization of Particle Filters
Authors: Osiris Terry, Kenneth Hopkinson, Laura Humphrey
Abstract:
Systematic resampling is the most popularly used resampling method in particle filters. This paper seeks to further the understanding of systematic resampling by defining a formula made up of variables from the sampling equation and the particle weights. The formula is then verified via SPARK, a software verification language. The verified systematic resampling formula states that the minimum/maximum number of possible samples taken of a particle is equal to the floor/ceiling value of particle weight divided by the sampling interval, respectively. This allows for the creation of a randomness spectrum that each resampling method can fall within. Methods on the lower end, e.g., systematic resampling, have less randomness and, thus, are quicker to reach an estimate. Although lower randomness allows for error by having a larger bias towards the size of the weight, having this bias creates vulnerabilities to the noise in the environment, e.g., jamming. Conclusively, this is the first step in characterizing each resampling method. This will allow target-tracking engineers to pick the best resampling method for their environment instead of choosing the most popularly used one.Keywords: SPARK, software verification, resampling, systematic resampling, particle filter, tracking
Procedia PDF Downloads 86973 Ultra-High Precision Diamond Turning of Infrared Lenses
Authors: Khaled Abou-El-Hossein
Abstract:
The presentation will address the features of two IR convex lenses that have been manufactured using an ultra-high precision machining centre based on single-point diamond turning. The lenses are made from silicon and germanium with a radius of curvature of 500 mm. Because of the brittle nature of silicon and germanium, machining parameters were selected in such a way that ductile regime was achieved. The cutting speed was 800 rpm while the feed rate and depth cut were 20 mm/min and 20 um, respectively. Although both materials comprise a mono-crystalline microstructure and are quite similar in terms of optical properties, machining of silicon was accompanied with more difficulties in terms of form accuracy compared to germanium machining. The P-V error of the silicon profile was 0.222 um while it was only 0.055 um for the germanium lens. This could be attributed to the accelerated wear that takes place on the tool edge when turning mono-crystalline silicon. Currently, we are using other ranges of the machining parameters in order to determine their optimal range that could yield satisfactory performance in terms of form accuracy when fabricating silicon lenses.Keywords: diamond turning, optical surfaces, precision machining, surface roughness
Procedia PDF Downloads 319