Search results for: bayesian estimation
1211 On Estimating the Headcount Index by Using the Logistic Regression Estimator
Authors: Encarnación Álvarez, Rosa M. García-Fernández, Juan F. Muñoz, Francisco J. Blanco-Encomienda
Abstract:
The problem of estimating a proportion has important applications in the field of economics, and in general, in many areas such as social sciences. A common application in economics is the estimation of the headcount index. In this paper, we define the general headcount index as a proportion. Furthermore, we introduce a new quantitative method for estimating the headcount index. In particular, we suggest to use the logistic regression estimator for the problem of estimating the headcount index. Assuming a real data set, results derived from Monte Carlo simulation studies indicate that the logistic regression estimator can be more accurate than the traditional estimator of the headcount index.Keywords: poverty line, poor, risk of poverty, Monte Carlo simulations, sample
Procedia PDF Downloads 4231210 The Role of Logistics Services in Influencing Customer Satisfaction and Reviews in an Online Marketplace
Authors: nafees mahbub, blake tindol, utkarsh shrivastava, kuanchin chen
Abstract:
Online shopping has become an integral part of businesses today. Big players such as Amazon are setting the bar for delivery services, and many businesses are working towards meeting them. However, what happens if a seller underestimates or overestimates the delivery time? Does it translate to consumer comments, ratings, or lost sales? Although several prior studies have investigated the impact of poor logistics on customer satisfaction, that impact of under estimation of delivery times has been rarely considered. The study uses real-time customer online purchase data to study the impact of missed delivery times on satisfaction.Keywords: LOST SALES, DELIVERY TIME, CUSTOMER SATISFACTION, CUSTOMER REVIEWS
Procedia PDF Downloads 2141209 Application of Adaptive Particle Filter for Localizing a Mobile Robot Using 3D Camera Data
Authors: Maysam Shahsavari, Seyed Jamalaldin Haddadi
Abstract:
There are several methods to localize a mobile robot such as relative, absolute and probabilistic. In this paper, particle filter due to its simple implementation and the fact that it does not need to know to the starting position will be used. This method estimates the position of the mobile robot using a probabilistic distribution, relying on a known map of the environment instead of predicting it. Afterwards, it updates this estimation by reading input sensors and control commands. To receive information from the surrounding world, distance to obstacles, for example, a Kinect is used which is much cheaper than a laser range finder. Finally, after explaining the Adaptive Particle Filter method and its implementation in detail, we will compare this method with the dead reckoning method and show that this method is much more suitable for situations in which we have a map of the environment.Keywords: particle filter, localization, methods, odometry, kinect
Procedia PDF Downloads 2691208 Matrix Completion with Heterogeneous Cost
Authors: Ilqar Ramazanli
Abstract:
The matrix completion problem has been studied broadly under many underlying conditions. The problem has been explored under adaptive or non-adaptive, exact or estimation, single-phase or multi-phase, and many other categories. In most of these cases, the observation cost of each entry is uniform and has the same cost across the columns. However, in many real-life scenarios, we could expect elements from distinct columns or distinct positions to have a different cost. In this paper, we explore this generalization under adaptive conditions. We approach the problem under two different cost models. The first one is that entries from different columns have different observation costs, but within the same column, each entry has a uniform cost. The second one is any two entry has different observation cost, despite being the same or different columns. We provide complexity analysis of our algorithms and provide tightness guarantees.Keywords: matroid optimization, matrix completion, linear algebra, algorithms
Procedia PDF Downloads 1091207 Application of Principal Component Analysis and Ordered Logit Model in Diabetic Kidney Disease Progression in People with Type 2 Diabetes
Authors: Mequanent Wale Mekonen, Edoardo Otranto, Angela Alibrandi
Abstract:
Diabetic kidney disease is one of the main microvascular complications caused by diabetes. Several clinical and biochemical variables are reported to be associated with diabetic kidney disease in people with type 2 diabetes. However, their interrelations could distort the effect estimation of these variables for the disease's progression. The objective of the study is to determine how the biochemical and clinical variables in people with type 2 diabetes are interrelated with each other and their effects on kidney disease progression through advanced statistical methods. First, principal component analysis was used to explore how the biochemical and clinical variables intercorrelate with each other, which helped us reduce a set of correlated biochemical variables to a smaller number of uncorrelated variables. Then, ordered logit regression models (cumulative, stage, and adjacent) were employed to assess the effect of biochemical and clinical variables on the order-level response variable (progression of kidney function) by considering the proportionality assumption for more robust effect estimation. This retrospective cross-sectional study retrieved data from a type 2 diabetic cohort in a polyclinic hospital at the University of Messina, Italy. The principal component analysis yielded three uncorrelated components. These are principal component 1, with negative loading of glycosylated haemoglobin, glycemia, and creatinine; principal component 2, with negative loading of total cholesterol and low-density lipoprotein; and principal component 3, with negative loading of high-density lipoprotein and a positive load of triglycerides. The ordered logit models (cumulative, stage, and adjacent) showed that the first component (glycosylated haemoglobin, glycemia, and creatinine) had a significant effect on the progression of kidney disease. For instance, the cumulative odds model indicated that the first principal component (linear combination of glycosylated haemoglobin, glycemia, and creatinine) had a strong and significant effect on the progression of kidney disease, with an effect or odds ratio of 0.423 (P value = 0.000). However, this effect was inconsistent across levels of kidney disease because the first principal component did not meet the proportionality assumption. To address the proportionality problem and provide robust effect estimates, alternative ordered logit models, such as the partial cumulative odds model, the partial adjacent category model, and the partial continuation ratio model, were used. These models suggested that clinical variables such as age, sex, body mass index, medication (metformin), and biochemical variables such as glycosylated haemoglobin, glycemia, and creatinine have a significant effect on the progression of kidney disease.Keywords: diabetic kidney disease, ordered logit model, principal component analysis, type 2 diabetes
Procedia PDF Downloads 391206 Improving 99mTc-tetrofosmin Myocardial Perfusion Images by Time Subtraction Technique
Authors: Yasuyuki Takahashi, Hayato Ishimura, Masao Miyagawa, Teruhito Mochizuki
Abstract:
Quantitative measurement of myocardium perfusion is possible with single photon emission computed tomography (SPECT) using a semiconductor detector. However, accumulation of 99mTc-tetrofosmin in the liver may make it difficult to assess that accurately in the inferior myocardium. Our idea is to reduce the high accumulation in the liver by using dynamic SPECT imaging and a technique called time subtraction. We evaluated the performance of a new SPECT system with a cadmium-zinc-telluride solid-state semi- conductor detector (Discovery NM 530c; GE Healthcare). Our system acquired list-mode raw data over 10 minutes for a typical patient. From the data, ten SPECT images were reconstructed, one for every minute of acquired data. Reconstruction with the semiconductor detector was based on an implementation of a 3-D iterative Bayesian reconstruction algorithm. We studied 20 patients with coronary artery disease (mean age 75.4 ± 12.1 years; range 42-86; 16 males and 4 females). In each subject, 259 MBq of 99mTc-tetrofosmin was injected intravenously. We performed both a phantom and a clinical study using dynamic SPECT. An approximation to a liver-only image is obtained by reconstructing an image from the early projections during which time the liver accumulation dominates (0.5~2.5 minutes SPECT image-5~10 minutes SPECT image). The extracted liver-only image is then subtracted from a later SPECT image that shows both the liver and the myocardial uptake (5~10 minutes SPECT image-liver-only image). The time subtraction of liver was possible in both a phantom and the clinical study. The visualization of the inferior myocardium was improved. In past reports, higher accumulation in the myocardium due to the overlap of the liver is un-diagnosable. Using our time subtraction method, the image quality of the 99mTc-tetorofosmin myocardial SPECT image is considerably improved.Keywords: 99mTc-tetrofosmin, dynamic SPECT, time subtraction, semiconductor detector
Procedia PDF Downloads 3351205 Application of Multilinear Regression Analysis for Prediction of Synthetic Shear Wave Velocity Logs in Upper Assam Basin
Authors: Triveni Gogoi, Rima Chatterjee
Abstract:
Shear wave velocity (Vs) estimation is an important approach in the seismic exploration and characterization of a hydrocarbon reservoir. There are varying methods for prediction of S-wave velocity, if recorded S-wave log is not available. But all the available methods for Vs prediction are empirical mathematical models. Shear wave velocity can be estimated using P-wave velocity by applying Castagna’s equation, which is the most common approach. The constants used in Castagna’s equation vary for different lithologies and geological set-ups. In this study, multiple regression analysis has been used for estimation of S-wave velocity. The EMERGE module from Hampson-Russel software has been used here for generation of S-wave log. Both single attribute and multi attributes analysis have been carried out for generation of synthetic S-wave log in Upper Assam basin. Upper Assam basin situated in North Eastern India is one of the most important petroleum provinces of India. The present study was carried out using four wells of the study area. Out of these wells, S-wave velocity was available for three wells. The main objective of the present study is a prediction of shear wave velocities for wells where S-wave velocity information is not available. The three wells having S-wave velocity were first used to test the reliability of the method and the generated S-wave log was compared with actual S-wave log. Single attribute analysis has been carried out for these three wells within the depth range 1700-2100m, which corresponds to Barail group of Oligocene age. The Barail Group is the main target zone in this study, which is the primary producing reservoir of the basin. A system generated list of attributes with varying degrees of correlation appeared and the attribute with the highest correlation was concerned for the single attribute analysis. Crossplot between the attributes shows the variation of points from line of best fit. The final result of the analysis was compared with the available S-wave log, which shows a good visual fit with a correlation of 72%. Next multi-attribute analysis has been carried out for the same data using all the wells within the same analysis window. A high correlation of 85% has been observed between the output log from the analysis and the recorded S-wave. The almost perfect fit between the synthetic S-wave and the recorded S-wave log validates the reliability of the method. For further authentication, the generated S-wave data from the wells have been tied to the seismic and correlated them. Synthetic share wave log has been generated for the well M2 where S-wave is not available and it shows a good correlation with the seismic. Neutron porosity, density, AI and P-wave velocity are proved to be the most significant variables in this statistical method for S-wave generation. Multilinear regression method thus can be considered as a reliable technique for generation of shear wave velocity log in this study.Keywords: Castagna's equation, multi linear regression, multi attribute analysis, shear wave logs
Procedia PDF Downloads 2291204 QCARNet: Networks for Quality-Adaptive Compression Artifact
Authors: Seung Ho Park, Young Su Moon, Nam Ik Cho
Abstract:
We propose a convolution neural network (CNN) for quality adaptive compression artifact reduction named QCARNet. The proposed method is different from the existing discriminative models that learn a specific model at a certain quality level. The method is composed of a quality estimation CNN (QECNN) and a compression artifact reduction CNN (CARCNN), which are two functionally separate CNNs. By connecting the QECNN and CARCNN, each CARCNN layer is able to adaptively reduce compression artifacts and preserve details depending on the estimated quality level map generated by the QECNN. We experimentally demonstrate that the proposed method achieves better performance compared to other state-of-the-art blind compression artifact reduction methods.Keywords: compression artifact reduction, deblocking, image denoising, image restoration
Procedia PDF Downloads 1391203 The Impact of Trade Liberalization on Current Account Deficit: The Turkish Case
Authors: E. Selçuk, Z. Karaçor, P. Yardımcı
Abstract:
Trade liberalization and its effects on the economies of developing countries have been investigated by many different studies, and some of them have focused on its impact on the current account balance. Turkey, as being one of the countries, which has liberalized its foreign trade in the 1980s, also needs to be studied in terms of the impact of liberalization on current account deficits. Therefore, the aim of this study is to find out whether trade liberalization has affected Turkey’s trade and current account balances. In order to determine this, yearly data of Turkey from 1980 to 2013 is used. As liberalization dummy, the year 1989, which was set for Turkey, is selected. Structural break test and model estimation results show that trade liberalization has a negative impact on trade balance but do not have a significant impact on the current account balance.Keywords: budget deficit, liberalization, Turkish economy, current account
Procedia PDF Downloads 3801202 Comparison of MODIS-Based Rice Extent Map and Landsat-Based Rice Classification Map in Determining Biomass Energy Potential of Rice Hull in Nueva Ecija, Philippines
Authors: Klathea Sevilla, Marjorie Remolador, Bryan Baltazar, Imee Saladaga, Loureal Camille Inocencio, Ma. Rosario Concepcion Ang
Abstract:
The underutilization of biomass resources in the Philippines, combined with its growing population and the rise in fossil fuel prices confirms demand for alternative energy sources. The goal of this paper is to provide a comparison of MODIS-based and Landsat-based agricultural land cover maps when used in the estimation of rice hull’s available energy potential. Biomass resource assessment was done using mathematical models and remote sensing techniques employed in a GIS platform.Keywords: biomass, geographic information system (GIS), remote sensing, renewable energy
Procedia PDF Downloads 4811201 Application of the Discrete Rationalized Haar Transform to Distributed Parameter System
Authors: Joon-Hoon Park
Abstract:
In this paper the rationalized Haar transform is applied for distributed parameter system identification and estimation. A distributed parameter system is a dynamical and mathematical model described by a partial differential equation. And system identification concerns the problem of determining mathematical models from observed data. The Haar function has some disadvantages of calculation because it contains irrational numbers, for these reasons the rationalized Haar function that has only rational numbers. The algorithm adopted in this paper is based on the transform and operational matrix of the rationalized Haar function. This approach provides more convenient and efficient computational results.Keywords: distributed parameter system, rationalized Haar transform, operational matrix, system identification
Procedia PDF Downloads 5091200 Comparison of Different Artificial Intelligence-Based Protein Secondary Structure Prediction Methods
Authors: Jamerson Felipe Pereira Lima, Jeane Cecília Bezerra de Melo
Abstract:
The difficulty and cost related to obtaining of protein tertiary structure information through experimental methods, such as X-ray crystallography or NMR spectroscopy, helped raising the development of computational methods to do so. An approach used in these last is prediction of tridimensional structure based in the residue chain, however, this has been proved an NP-hard problem, due to the complexity of this process, explained by the Levinthal paradox. An alternative solution is the prediction of intermediary structures, such as the secondary structure of the protein. Artificial Intelligence methods, such as Bayesian statistics, artificial neural networks (ANN), support vector machines (SVM), among others, were used to predict protein secondary structure. Due to its good results, artificial neural networks have been used as a standard method to predict protein secondary structure. Recent published methods that use this technique, in general, achieved a Q3 accuracy between 75% and 83%, whereas the theoretical accuracy limit for protein prediction is 88%. Alternatively, to achieve better results, support vector machines prediction methods have been developed. The statistical evaluation of methods that use different AI techniques, such as ANNs and SVMs, for example, is not a trivial problem, since different training sets, validation techniques, as well as other variables can influence the behavior of a prediction method. In this study, we propose a prediction method based on artificial neural networks, which is then compared with a selected SVM method. The chosen SVM protein secondary structure prediction method is the one proposed by Huang in his work Extracting Physico chemical Features to Predict Protein Secondary Structure (2013). The developed ANN method has the same training and testing process that was used by Huang to validate his method, which comprises the use of the CB513 protein data set and three-fold cross-validation, so that the comparative analysis of the results can be made comparing directly the statistical results of each method.Keywords: artificial neural networks, protein secondary structure, protein structure prediction, support vector machines
Procedia PDF Downloads 6211199 Developing a DNN Model for the Production of Biogas From a Hybrid BO-TPE System in an Anaerobic Wastewater Treatment Plant
Authors: Hadjer Sadoune, Liza Lamini, Scherazade Krim, Amel Djouadi, Rachida Rihani
Abstract:
Deep neural networks are highly regarded for their accuracy in predicting intricate fermentation processes. Their ability to learn from a large amount of datasets through artificial intelligence makes them particularly effective models. The primary obstacle in improving the performance of these models is to carefully choose the suitable hyperparameters, including the neural network architecture (number of hidden layers and hidden units), activation function, optimizer, learning rate, and other relevant factors. This study predicts biogas production from real wastewater treatment plant data using a sophisticated approach: hybrid Bayesian optimization with a tree-structured Parzen estimator (BO-TPE) for an optimised deep neural network (DNN) model. The plant utilizes an Upflow Anaerobic Sludge Blanket (UASB) digester that treats industrial wastewater from soft drinks and breweries. The digester has a working volume of 1574 m3 and a total volume of 1914 m3. Its internal diameter and height were 19 and 7.14 m, respectively. The data preprocessing was conducted with meticulous attention to preserving data quality while avoiding data reduction. Three normalization techniques were applied to the pre-processed data (MinMaxScaler, RobustScaler and StandardScaler) and compared with the Non-Normalized data. The RobustScaler approach has strong predictive ability for estimating the volume of biogas produced. The highest predicted biogas volume was 2236.105 Nm³/d, with coefficient of determination (R2), mean absolute error (MAE), and root mean square error (RMSE) values of 0.712, 164.610, and 223.429, respectively.Keywords: anaerobic digestion, biogas production, deep neural network, hybrid bo-tpe, hyperparameters tuning
Procedia PDF Downloads 381198 Cooperative AF Scheme for Multi Source and Terminal in Edge of Cell Coverage
Authors: Myoung-Jin Kim, Chang-Bin Ha, Yeong-Seop Ahn, Hyoung-Kyu Song
Abstract:
This paper proposes a cooperative communication scheme for improve wireless communication performance. When the receiver is located in the edge of coverage, the signal from the transmitter is distorted for various reasons such as inter-cell interference (ICI), power reduction, incorrect channel estimation. In order to improve communication performance, the proposed scheme adds the relay. By the relay, the receiver has diversity gain. In this paper, two base stations, one relay and one destination are considered. The two base stations transmit same time to relay and destination. The relay forwarding to destination and the destination detects signals.Keywords: cooperative communication, diversity gain, OFDM, MMSE
Procedia PDF Downloads 3891197 Fault Detection and Isolation in Attitude Control Subsystem of Spacecraft Formation Flying Using Extended Kalman Filters
Authors: S. Ghasemi, K. Khorasani
Abstract:
In this paper, the problem of fault detection and isolation in the attitude control subsystem of spacecraft formation flying is considered. In order to design the fault detection method, an extended Kalman filter is utilized which is a nonlinear stochastic state estimation method. Three fault detection architectures, namely, centralized, decentralized, and semi-decentralized are designed based on the extended Kalman filters. Moreover, the residual generation and threshold selection techniques are proposed for these architectures.Keywords: component, formation flight of satellites, extended Kalman filter, fault detection and isolation, actuator fault
Procedia PDF Downloads 4341196 A New Method for Estimating the Mass Recession Rate for Ablator Systems
Authors: Bianca A. Szasz, Keiichi Okuyama
Abstract:
As the human race will continue to explore the space by creating new space transportation means and sending them to other planets, the enhance of atmospheric reentry study is crucial. In this context, an analysis of mass recession rate of ablative materials for thermal shields of reentry spacecrafts is important to be carried out. The paper describes a new estimation method for calculating the mass recession of an ablator system, this method combining an old method with a new one, which was recently elaborated by Okuyama et al. The space mission of USERS spacecraft is taken as a case study and the possibility of implementing lighter ablative materials in future space missions is taking into consideration.Keywords: ablator system, mass recession, reentry spacecraft, ablative materials
Procedia PDF Downloads 2721195 Model-Driven and Data-Driven Approaches for Crop Yield Prediction: Analysis and Comparison
Authors: Xiangtuo Chen, Paul-Henry Cournéde
Abstract:
Crop yield prediction is a paramount issue in agriculture. The main idea of this paper is to find out efficient way to predict the yield of corn based meteorological records. The prediction models used in this paper can be classified into model-driven approaches and data-driven approaches, according to the different modeling methodologies. The model-driven approaches are based on crop mechanistic modeling. They describe crop growth in interaction with their environment as dynamical systems. But the calibration process of the dynamic system comes up with much difficulty, because it turns out to be a multidimensional non-convex optimization problem. An original contribution of this paper is to propose a statistical methodology, Multi-Scenarios Parameters Estimation (MSPE), for the parametrization of potentially complex mechanistic models from a new type of datasets (climatic data, final yield in many situations). It is tested with CORNFLO, a crop model for maize growth. On the other hand, the data-driven approach for yield prediction is free of the complex biophysical process. But it has some strict requirements about the dataset. A second contribution of the paper is the comparison of these model-driven methods with classical data-driven methods. For this purpose, we consider two classes of regression methods, methods derived from linear regression (Ridge and Lasso Regression, Principal Components Regression or Partial Least Squares Regression) and machine learning methods (Random Forest, k-Nearest Neighbor, Artificial Neural Network and SVM regression). The dataset consists of 720 records of corn yield at county scale provided by the United States Department of Agriculture (USDA) and the associated climatic data. A 5-folds cross-validation process and two accuracy metrics: root mean square error of prediction(RMSEP), mean absolute error of prediction(MAEP) were used to evaluate the crop prediction capacity. The results show that among the data-driven approaches, Random Forest is the most robust and generally achieves the best prediction error (MAEP 4.27%). It also outperforms our model-driven approach (MAEP 6.11%). However, the method to calibrate the mechanistic model from dataset easy to access offers several side-perspectives. The mechanistic model can potentially help to underline the stresses suffered by the crop or to identify the biological parameters of interest for breeding purposes. For this reason, an interesting perspective is to combine these two types of approaches.Keywords: crop yield prediction, crop model, sensitivity analysis, paramater estimation, particle swarm optimization, random forest
Procedia PDF Downloads 2311194 A Multi Function Myocontroller for Upper Limb Prostheses
Authors: Ayad Asaad Ibrahim
Abstract:
Myoelectrically controlled prostheses are becoming more and more popular, for below-elbow amputation, the wrist flexor and extensor muscle group, while for above-elbow biceps and triceps brachii muscles are used for control of the prosthesis. A two site multi-function controller is presented. Two stainless steel bipolar electrode pairs are used to monitor the activities in both muscles. The detected signals are processed by new pre-whitening technique to identify the accurate tension estimation in these muscles. These estimates will activate the relevant prosthesis control signal, with a time constant of 200 msec. It is ensured that the tension states in the control muscle to activate a particular prosthesis function are similar to those used to activate normal functions in the natural hand. This facilitates easier training.Keywords: prosthesis, biosignal processing, pre-whitening, myoelectric controller
Procedia PDF Downloads 3631193 Estimation of the Pore Electrical Conductivity Using Dielectric Sensors
Authors: Fethi Bouksila, Magnus Persson, Ronny Berndtsson, Akissa Bahri
Abstract:
Under salinity conditions, we evaluate the performance of Hilhost (2000) model to predict pore electrical conductivity ECp from dielectric permittivity and bulk electrical conductivity (ECa) using Time and Frequency Domain Reflectometry sensors (TDR, FDR). Using FDR_WET sensor, RMSE of ECp was 4.15 dS m-1. By replacing the standard soil parameter (K0) in Hilhost model by K0-ECa relationship, the RMSE of ECp decreased to 0.68 dS m-1. WET sensor could give similar accuracy to estimate ECp than TDR if calibrated values of K0 were used instead of standard values in Hilhost model.Keywords: hilhost model, soil salinity, time domain reflectometry, frequency domain reflectometry, dielectric methods
Procedia PDF Downloads 1351192 How Do Crisis Affect Economic Policy?
Authors: Eva Kotlánová
Abstract:
After recession that began in 2007 in the United States and subsequently spilled over the Europe we could expect recovery of economic growth. According to the last estimation of economic progress of European countries, this recovery is not strong enough. Among others, it will depend on economic policy, where and in which way, the economic indicators will proceed. Economic theories postulate that the economic subjects prefer stably, continual economic policy without repeated and strong fluctuations. This policy is perceived as support of economic growth. Mostly in crises period, when the government must cope with consequences of recession, the economic policy becomes unpredictable for many subjects and economic policy uncertainty grows, which have negative influence on economic growth. The aim of this paper is to use panel regression to prove or disprove this hypothesis on the example of five largest European economies in the period 2008–2012.Keywords: economic crises in Europe, economic policy, uncertainty, panel analysis regression
Procedia PDF Downloads 3861191 Financial Literacy in Greek High-School Students
Authors: Vasiliki A. Tzora, Nikolaos D. Philippas
Abstract:
The paper measures the financial literacy of youth in Greece derived from the examined aspects of financial knowledge, behaviours, and attitudes that high school students performed. The findings reveal that less than half of participant high school students have an acceptable level of financial literacy. Also, students who are in the top of their class cohort exhibit higher levels of financial literacy. We also find that the father’s education level has a significant effect on financial literacy. Students who keep records of their income and expenses are likely to show better levels of financial literacy than students who do not. Students’ perception/estimation of their parents’ income changes is also related to their levels of financial literacy. We conclude that financial education initiatives should be embedded in schools in order to embrace the young generation.Keywords: financial literacy, financial knowledge, financial behaviour, financial attitude, financial wellbeing, 15-year-old students
Procedia PDF Downloads 1411190 Urban Design via Estimation Model for Traffic Index of Cities Based on an Artificial Intelligence
Authors: Seyed Sobhan Alvani, Mohammad Gohari
Abstract:
By developing cities and increasing the population, traffic congestion has become a vital problem. Due to this crisis, urban designers try to present solutions to decrease this difficulty. On the other hand, predicting the model with perfect accuracy is essential for solution-providing. The current study presents a model based on artificial intelligence which can predict traffic index based on city population, growth rate, and area. The accuracy of the model was evaluated, which is acceptable and it is around 90%. Thus, urban designers and planners can employ it for predicting traffic index in the future to provide strategies.Keywords: traffic index, population growth rate, cities wideness, artificial neural network
Procedia PDF Downloads 401189 Predatory Pricing at Services Markets: Incentives, Mechanisms, Standards of Proving, and Remedies
Authors: Mykola G. Boichuk
Abstract:
The paper concerns predatory pricing incentives and mechanisms in the markets of services, as well as its anti-competitive effects. As cost estimation at services markets is more complex in comparison to markets of goods, predatory pricing is more difficult to detect in the provision of services. For instance, this is often the case for professional services, which is analyzed in the paper. The special attention is given to employment markets as de-facto main supply markets for professional services markets. Also, the paper concerns such instances as travel agents' services, where predatory pricing may have implications not only on competition but on a wider range of public interest as well. Thus, the paper develops on effective ways to apply competition law rules on predatory pricing to the provision of services.Keywords: employment markets, predatory pricing, services markets, unfair competition
Procedia PDF Downloads 3251188 Service Life Prediction of Tunnel Structures Subjected to Water Seepage
Authors: Hassan Baji, Chun-Qing Li, Wei Yang
Abstract:
Water seepage is one of the most common causes of damage in tunnel structures, which can cause direct and indirect e.g. reinforcement corrosion and calcium leaching damages. Estimation of water seepage or inflow is one of the main challenges in probabilistic assessment of tunnels. The methodology proposed in this study is an attempt for mathematically modeling the water seepage in tunnel structures and further predicting its service life. Using the time-dependent reliability, water seepage is formulated as a failure mode, which can be used for prediction of service life. Application of the formulated seepage failure mode to a case study tunnel is presented.Keywords: water seepage, tunnels, time-dependent reliability, service life
Procedia PDF Downloads 4821187 An Application-Driven Procedure for Optimal Signal Digitization of Automotive-Grade Ultrasonic Sensors
Authors: Mohamed Shawki Elamir, Heinrich Gotzig, Raoul Zoellner, Patrick Maeder
Abstract:
In this work, a methodology is presented for identifying the optimal digitization parameters for the analog signal of ultrasonic sensors. These digitization parameters are the resolution of the analog to digital conversion and the sampling rate. This is accomplished through the derivation of characteristic curves based on Fano inequality and the calculation of the mutual information content over a given dataset. The mutual information is calculated between the examples in the dataset and the corresponding variation in the feature that needs to be estimated. The optimal parameters are identified in a manner that ensures optimal estimation performance while preventing inefficiency in using unnecessarily powerful analog to digital converters.Keywords: analog to digital conversion, digitization, sampling rate, ultrasonic
Procedia PDF Downloads 2071186 A Fuzzy Nonlinear Regression Model for Interval Type-2 Fuzzy Sets
Authors: O. Poleshchuk, E. Komarov
Abstract:
This paper presents a regression model for interval type-2 fuzzy sets based on the least squares estimation technique. Unknown coefficients are assumed to be triangular fuzzy numbers. The basic idea is to determine aggregation intervals for type-1 fuzzy sets, membership functions of whose are low membership function and upper membership function of interval type-2 fuzzy set. These aggregation intervals were called weighted intervals. Low and upper membership functions of input and output interval type-2 fuzzy sets for developed regression models are considered as piecewise linear functions.Keywords: interval type-2 fuzzy sets, fuzzy regression, weighted interval
Procedia PDF Downloads 3731185 Block Matching Based Stereo Correspondence for Depth Calculation
Authors: G. Balakrishnan
Abstract:
Stereo Correspondence plays a major role in estimation of distance of an object from the stereo camera pair for various applications. In this paper, a stereo correspondence algorithm based on block-matching technique is presented. Initially, an energy matrix is calculated for every disparity obtained using modified Sum of Absolute Difference (SAD). Higher energy matrix errors are removed by using threshold value in order to reduce the mismatch errors. A smoothening filter is applied to eliminate unreliable disparity estimate across the object boundaries. The purpose is to improve the reliability of calculation of disparity map. The experimental results obtained shows that the final depth map produce better results and can be used to all the applications using stereo cameras.Keywords: stereo matching, filters, energy matrix, disparity
Procedia PDF Downloads 2151184 Estimation of Sediment Transport into a Reservoir Dam
Authors: Kiyoumars Roushangar, Saeid Sadaghian
Abstract:
Although accurate sediment load prediction is very important in planning, designing, operating and maintenance of water resources structures, the transport mechanism is complex, and the deterministic transport models are based on simplifying assumptions often lead to large prediction errors. In this research, firstly, two intelligent ANN methods, Radial Basis and General Regression Neural Networks, are adopted to model of total sediment load transport into Madani Dam reservoir (north of Iran) using the measured data and then applicability of the sediment transport methods developed by Engelund and Hansen, Ackers and White, Yang, and Toffaleti for predicting of sediment load discharge are evaluated. Based on comparison of the results, it is found that the GRNN model gives better estimates than the sediment rating curve and mentioned classic methods.Keywords: sediment transport, dam reservoir, RBF, GRNN, prediction
Procedia PDF Downloads 4961183 Identifying Protein-Coding and Non-Coding Regions in Transcriptomes
Authors: Angela U. Makolo
Abstract:
Protein-coding and Non-coding regions determine the biology of a sequenced transcriptome. Research advances have shown that Non-coding regions are important in disease progression and clinical diagnosis. Existing bioinformatics tools have been targeted towards Protein-coding regions alone. Therefore, there are challenges associated with gaining biological insights from transcriptome sequence data. These tools are also limited to computationally intensive sequence alignment, which is inadequate and less accurate to identify both Protein-coding and Non-coding regions. Alignment-free techniques can overcome the limitation of identifying both regions. Therefore, this study was designed to develop an efficient sequence alignment-free model for identifying both Protein-coding and Non-coding regions in sequenced transcriptomes. Feature grouping and randomization procedures were applied to the input transcriptomes (37,503 data points). Successive iterations were carried out to compute the gradient vector that converged the developed Protein-coding and Non-coding Region Identifier (PNRI) model to the approximate coefficient vector. The logistic regression algorithm was used with a sigmoid activation function. A parameter vector was estimated for every sample in 37,503 data points in a bid to reduce the generalization error and cost. Maximum Likelihood Estimation (MLE) was used for parameter estimation by taking the log-likelihood of six features and combining them into a summation function. Dynamic thresholding was used to classify the Protein-coding and Non-coding regions, and the Receiver Operating Characteristic (ROC) curve was determined. The generalization performance of PNRI was determined in terms of F1 score, accuracy, sensitivity, and specificity. The average generalization performance of PNRI was determined using a benchmark of multi-species organisms. The generalization error for identifying Protein-coding and Non-coding regions decreased from 0.514 to 0.508 and to 0.378, respectively, after three iterations. The cost (difference between the predicted and the actual outcome) also decreased from 1.446 to 0.842 and to 0.718, respectively, for the first, second and third iterations. The iterations terminated at the 390th epoch, having an error of 0.036 and a cost of 0.316. The computed elements of the parameter vector that maximized the objective function were 0.043, 0.519, 0.715, 0.878, 1.157, and 2.575. The PNRI gave an ROC of 0.97, indicating an improved predictive ability. The PNRI identified both Protein-coding and Non-coding regions with an F1 score of 0.970, accuracy (0.969), sensitivity (0.966), and specificity of 0.973. Using 13 non-human multi-species model organisms, the average generalization performance of the traditional method was 74.4%, while that of the developed model was 85.2%, thereby making the developed model better in the identification of Protein-coding and Non-coding regions in transcriptomes. The developed Protein-coding and Non-coding region identifier model efficiently identified the Protein-coding and Non-coding transcriptomic regions. It could be used in genome annotation and in the analysis of transcriptomes.Keywords: sequence alignment-free model, dynamic thresholding classification, input randomization, genome annotation
Procedia PDF Downloads 681182 Statistical Estimation of Ionospheric Energy Dissipation Using ØStgaard's Empirical Relation
Authors: M. A. Ahmadu, S. S. Rabia
Abstract:
During the past few decades, energy dissipation in the ionosphere resulting from the geomagnetic activity has caused an increasing number of major disruptions of important power and communication services, malfunctions and loss of expensive facilities. Here, the electron precipitation energy, w(ep) and joule heating energy, w(jh) was used in the computation of this dissipation using Østgaard’s empirical relation from hourly geomagnetic indices of 2012, under the assumption that the magnetosphere does not store any energy, so that at the beginning of the activity t1=0 and end at t2=t, the statistical results obtained show that ionospheric dissipation varies month to month, day to day and hour to hour and estimated with a value ~3.6 w(ep), which is in agreement with experimental result.Keywords: Ostgaard's, ionospheric dissipation, joule heating, electron precipitation, geomagnetic indices, empirical relation
Procedia PDF Downloads 293