Search results for: Factor analysis model
13994 Performance Analysis of Adaptive LMS Filter through Regression Analysis using SystemC
Authors: Hyeong-Geon Lee, Jae-Young Park, Suk-ki Lee, Jong-Tae Kim
Abstract:
The LMS adaptive filter has several parameters which can affect their performance. From among these parameters, most papers handle the step size parameter for controlling the performance. In this paper, we approach three parameters: step-size, filter tap-size and filter form. The regression analysis is used for defining the relation between parameters and performance of LMS adaptive filter with using the system level simulation results. The results present that all parameters have performance trends in each own particular form, which can be estimated from equations drawn by regression analysis.
Keywords: System level model, adaptive LMS FIR filter, regression analysis, systemC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 280013993 Effect of Unbound Granular Materials Nonlinear Resilient Behavior on Pavement Response and Performance of Low Volume Roads
Authors: K. Sandjak, B. Tiliouine
Abstract:
Structural analysis of flexible pavements has been and still is currently performed using multi-layer elastic theory. However, for thinly surfaced pavements subjected to low to medium volumes of traffics, the importance of non-linear stress-strain behavior of unbound granular materials (UGM) requires the use of more sophisticated numerical models for structural design and performance of such pavements. In the present work, nonlinear unbound aggregates constitutive model is implemented within an axisymmetric finite element code developed to simulate the nonlinear behavior of pavement structures including two local aggregates of different mineralogical nature, typically used in Algerian pavements. The performance of the mechanical model is examined about its capability of representing adequately, under various conditions, the granular material non-linearity in pavement analysis. In addition, deflection data collected by Falling Weight Deflectometer (FWD) are incorporated into the analysis in order to assess the sensitivity of critical pavement design criteria and pavement design life to the constitutive model. Finally, conclusions of engineering significance are formulated.
Keywords: Nonlinear resilient behavior, unbound granular materials, RLT test results, FWD backcalculations, finite element simulations, pavement response and performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 229313992 Net-Banking System as a Game
Authors: N. Ghoualmi-Zine, A. Araar
Abstract:
In this article we propose to model Net-banking system by game theory. We adopt extensive game to model our web application. We present the model in term of players and strategy. We present UML diagram related the protocol game.Keywords: Game theory, model, state, web application.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155113991 Hemodynamic Characteristics in the Human Carotid Artery Model Induced by Blood-Arterial Wall Interactions
Authors: Taewon Seo
Abstract:
The characteristics of physiological blood flow in human carotid arterial bifurcation model have been numerically studied using a fully coupled fluid-structure interaction (FSI) analysis. This computational model with the fluid-structure interaction is constructed to investigate the flow characteristics and wall shear stress in the carotid artery. As the flow begins to decelerate after the peak flow, a large recirculation zone develops at the non-divider wall of both internal carotid artery (ICA) and external carotid artery (ECA) in FSI model due to the elastic energy stored in the expanding compliant wall. The calculated difference in wall shear stress (WSS) in both Non-FSI and FSI models is a range of between 5 and 11% at the mean WSS. The low WSS corresponds to regions of carotid artery that are more susceptible to atherosclerosis.
Keywords: Carotid artery, Fluid-structure interaction, Hemodynamics, Wall shear stress.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 286713990 Optimization of the Input Layer Structure for Feed-Forward Narx Neural Networks
Authors: Zongyan Li, Matt Best
Abstract:
This paper presents an optimization method for reducing the number of input channels and the complexity of the feed-forward NARX neural network (NN) without compromising the accuracy of the NN model. By utilizing the correlation analysis method, the most significant regressors are selected to form the input layer of the NN structure. An application of vehicle dynamic model identification is also presented in this paper to demonstrate the optimization technique and the optimal input layer structure and the optimal number of neurons for the neural network is investigated.Keywords: Correlation analysis, F-ratio, Levenberg-Marquardt, MSE, NARX, neural network, optimisation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 218813989 Metabolic Analysis of Fibroblast Conditioned Media and Comparison with Theoretical Modeling
Authors: Priyanka Gupta, Paul Verma, Kerry Hourigan, Jayesh Bellare, Sameer Jadhav
Abstract:
Understanding the consumption and production of various metabolites of fibroblast conditioned media is needed for its proper and optimized use in expansion of pluripotent stem cells. For this purpose, we have used the HPLC method to analyse the consumption of glucose and the production of lactate over time by mouse embryonic fibroblasts. The experimental data have also been compared with mathematical model fits. 0.025 moles of lactate was produced after 72 hrs while the glucose concentration decreased from 0.017 moles to 0.011 moles. The mathematical model was able to predict the trends of glucose consumption and lactate production.Keywords: Conditioned media, HPLC, metabolite analysis, mouse embryonic fibroblast.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 261513988 The Research of Taiwan Green Building Materials (GBM) system and GBM Eco-Efficiency Model on Climate Change
Authors: Ting-Ting Hsieh, Che-Ming Chiang, Ming-Chin Ho, Kwang-Pang Lai
Abstract:
The globe Sustainability has become the subject of international attention, the key reason is that global climate change. Climate and disasters around the abnormal frequency multiplier, the global temperature of the catastrophe and disaster continue to occur throughout the world, as well as countries around the world. Currently there are many important international conferences and policy, it is a "global environmental sustainability " and "living human health " as the goal of development, including the APEC 2007 meeting to "climate Clean Energy" as the theme Sydney Declaration, 2008 World Economic Forum's "Carbon - promote Cool Earth energy efficiency improvement project", the EU proposed "Green Idea" program, the Japanese annual policy, "low-carbon society, sustainable eco-city environment (Eco City) "And from 2009 to 2010 to promote the "Eco-Point" to promote green energy and carbon reduction products .And the 2010 World Climate Change Conference (COP16 United Nations Climate Change Conference Copenhagen), the world has been the subject of Negative conservative "Environmental Protection ", "save energy consumption, " into a positive response to the "Sustainable " and" LOHAS", while Taiwan has actively put forward eco-cities, green building, green building materials and other related environmental response Measures, especially green building construction environment that is the basis of factors, the most widely used application level, and direct contact with human health and the key to sustainable planet. "Sustainable development "is a necessary condition for continuation of the Earth, "healthy and comfortable" is a necessary condition for the continuation of life, and improve the "quality" is a necessary condition for economic development, balance between the three is "to enhance the efficiency of ", According to the World Business Council for Sustainable Development (WBCSD) for the "environmental efficiency "(Eco-Efficiency) proposed: " the achievement of environmental efficiency, the price to be competitive in the provision of goods or services to meet people's needs, improve living Quality at the same time, the goods or services throughout the life cycle. Its impact on the environment and natural resource utilization and gradually reduced to the extent the Earth can load. "whichever is the economy "Economic" and " Ecologic". The research into the methodology to obtain the Taiwan Green Building Material Labeling product as the scope of the study, by investigating and weight analysis to explore green building environmental load (Ln) factor and the Green Building Quality (Qn) factor to Establish green building environmental efficiency assessment model (GBM Eco-Efficiency). And building materials for healthy green label products for priority assessment object, the object is set in the material evidence for the direct response to the environmental load from the floor class-based, explicit feedback correction to the Green Building environmental efficiency assessment model, "efficiency " as a starting point to achieve balance between human "health "and Earth "sustainable development of win-win strategy. The study is expected to reach 1.To establish green building materials and the quality of environmental impact assessment system, 2. To establish value of GBM Eco-Efficiency model, 3. To establish the GBM Eco-Efficiency model for application of green building material feedback mechanisms.
Keywords: Climate Change, Green Building Material (GBM), Eco-Efficiency, Life Cycle Assessment, Performance Evaluation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 233113987 Competitors’ Influence Analysis of a Retailer by Using Customer Value and Huff’s Gravity Model
Authors: Yepeng Cheng, Yasuhiko Morimoto
Abstract:
Customer relationship analysis is vital for retail stores, especially for supermarkets. The point of sale (POS) systems make it possible to record the daily purchasing behaviors of customers as an identification point of sale (ID-POS) database, which can be used to analyze customer behaviors of a supermarket. The customer value is an indicator based on ID-POS database for detecting the customer loyalty of a store. In general, there are many supermarkets in a city, and other nearby competitor supermarkets significantly affect the customer value of customers of a supermarket. However, it is impossible to get detailed ID-POS databases of competitor supermarkets. This study firstly focused on the customer value and distance between a customer's home and supermarkets in a city, and then constructed the models based on logistic regression analysis to analyze correlations between distance and purchasing behaviors only from a POS database of a supermarket chain. During the modeling process, there are three primary problems existed, including the incomparable problem of customer values, the multicollinearity problem among customer value and distance data, and the number of valid partial regression coefficients. The improved customer value, Huff’s gravity model, and inverse attractiveness frequency are considered to solve these problems. This paper presents three types of models based on these three methods for loyal customer classification and competitors’ influence analysis. In numerical experiments, all types of models are useful for loyal customer classification. The type of model, including all three methods, is the most superior one for evaluating the influence of the other nearby supermarkets on customers' purchasing of a supermarket chain from the viewpoint of valid partial regression coefficients and accuracy.Keywords: Customer value, Huff's Gravity Model, POS, retailer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 61213986 BIDENS: Iterative Density Based Biclustering Algorithm With Application to Gene Expression Analysis
Authors: Mohamed A. Mahfouz, M. A. Ismail
Abstract:
Biclustering is a very useful data mining technique for identifying patterns where different genes are co-related based on a subset of conditions in gene expression analysis. Association rules mining is an efficient approach to achieve biclustering as in BIMODULE algorithm but it is sensitive to the value given to its input parameters and the discretization procedure used in the preprocessing step, also when noise is present, classical association rules miners discover multiple small fragments of the true bicluster, but miss the true bicluster itself. This paper formally presents a generalized noise tolerant bicluster model, termed as μBicluster. An iterative algorithm termed as BIDENS based on the proposed model is introduced that can discover a set of k possibly overlapping biclusters simultaneously. Our model uses a more flexible method to partition the dimensions to preserve meaningful and significant biclusters. The proposed algorithm allows discovering biclusters that hard to be discovered by BIMODULE. Experimental study on yeast, human gene expression data and several artificial datasets shows that our algorithm offers substantial improvements over several previously proposed biclustering algorithms.Keywords: Machine learning, biclustering, bi-dimensional clustering, gene expression analysis, data mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 196313985 Cirrhosis Mortality Prediction as Classification Using Frequent Subgraph Mining
Authors: Abdolghani Ebrahimi, Diego Klabjan, Chenxi Ge, Daniela Ladner, Parker Stride
Abstract:
In this work, we use machine learning and data analysis techniques to predict the one-year mortality of cirrhotic patients. Data from 2,322 patients with liver cirrhosis are collected at a single medical center. Different machine learning models are applied to predict one-year mortality. A comprehensive feature space including demographic information, comorbidity, clinical procedure and laboratory tests is being analyzed. A temporal pattern mining technic called Frequent Subgraph Mining (FSM) is being used. Model for End-stage liver disease (MELD) prediction of mortality is used as a comparator. All of our models statistically significantly outperform the MELD-score model and show an average 10% improvement of the area under the curve (AUC). The FSM technic itself does not improve the model significantly, but FSM, together with a machine learning technique called an ensemble, further improves the model performance. With the abundance of data available in healthcare through electronic health records (EHR), existing predictive models can be refined to identify and treat patients at risk for higher mortality. However, due to the sparsity of the temporal information needed by FSM, the FSM model does not yield significant improvements. Our work applies modern machine learning algorithms and data analysis methods on predicting one-year mortality of cirrhotic patients and builds a model that predicts one-year mortality significantly more accurate than the MELD score. We have also tested the potential of FSM and provided a new perspective of the importance of clinical features.
Keywords: machine learning, liver cirrhosis, subgraph mining, supervised learning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 44913984 Stiffness Modeling of 3-PRS Mechanism
Authors: Xiaohui Han, Yuhan Wang, Jing Shi
Abstract:
This paper proposed a stiffness analysis method for a 3-PRS mechanism for welding thick aluminum plate using FSW technology. In the molding process, elastic deformation of lead-screws and links are taken into account. This method is based on the virtual work principle. Through a survey of the commonly used stiffness performance indices, the minimum and maximum eigenvalues of the stiffness matrix are used to evaluate the stiffness of the 3-PRS mechanism. Furthermore, A FEA model has been constructed to verify the method. Finally, we redefined the workspace using the stiffness analysis method.Keywords: 3-PRS, parallel mechanism, stiffness analysis, workspace.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 226613983 Reliability Verification of the Performance Evaluation of Multiphase Pump
Authors: Joon-Hyung Kim, Him-Chan Lee, Jin-Hyuk Kim, Yong-Kab Lee, Young-Seok Choi
Abstract:
The crude oil in an oil well exists in various phases such as gas, seawater, and sand, as well as oil. Therefore, a phase separator is needed at the front of a single-phase pump for pressurization and transfer. On the other hand, the application of a multiphase pump can provide such advantages as simplification of the equipment structure and cost savings, because there is no need for a phase separation process. Therefore, the crude oil transfer method using a multiphase pump is being applied to recently developed oil wells. Due to this increase in demand, technical demands for the development of multiphase pumps are sharply increasing, but the progress of research into related technologies is insufficient, due to the nature of multiphase pumps that require high levels of skills. This study was conducted to verify the reliability of pump performance evaluation using numerical analysis, which is the basis of the development of a multiphase pump. For this study, a model was designed by selecting the specifications of this study. The performance of the designed model was evaluated through numerical analysis and experiment. The results of the performance evaluation were compared to verify the reliability of the result using numerical analysis.
Keywords: Multiphase pump, Numerical analysis, Experiment, Performance evaluation, Reliability verification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 313413982 An Investigation on Vegetable Oils as Potential Insulating Liquid
Authors: C. Kocatepe, E. Taslak, C. F. Kumru, O. Arıkan
Abstract:
While choosing insulating oil, characteristic features such as thermal cooling, endurance, efficiency and being environment-friendly should be considered. Mineral oils are referred as petroleum-based oil. In this study, vegetable oils investigated as an alternative insulating liquid to mineral oil. Dissipation factor, breakdown voltage, relative dielectric constant and resistivity changes with the frequency and voltage of mineral, rapeseed and nut oils were measured. Experimental studies were performed according to ASTM D924 and IEC 60156 standards.Keywords: Breakdown voltage, dielectric dissipation factor, mineral oil, vegetable oils.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 259613981 An Incomplete Factorization Preconditioner for LMS Adaptive Filter
Authors: Shazia Javed, Noor Atinah Ahmad
Abstract:
In this paper an efficient incomplete factorization preconditioner is proposed for the Least Mean Squares (LMS) adaptive filter. The proposed preconditioner is approximated from a priori knowledge of the factors of input correlation matrix with an incomplete strategy, motivated by the sparsity patter of the upper triangular factor in the QRD-RLS algorithm. The convergence properties of IPLMS algorithm are comparable with those of transform domain LMS(TDLMS) algorithm. Simulation results show efficiency and robustness of the proposed algorithm with reduced computational complexity.
Keywords: Autocorrelation matrix, Cholesky's factor, eigenvalue spread, Markov input.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 179313980 Analytics Model in a Telehealth Center Based on Cloud Computing and Local Storage
Authors: L. Ramirez, E. Guillén, J. Sánchez
Abstract:
Some of the main goals about telecare such as monitoring, treatment, telediagnostic are deployed with the integration of applications with specific appliances. In order to achieve a coherent model to integrate software, hardware, and healthcare systems, different telehealth models with Internet of Things (IoT), cloud computing, artificial intelligence, etc. have been implemented, and their advantages are still under analysis. In this paper, we propose an integrated model based on IoT architecture and cloud computing telehealth center. Analytics module is presented as a solution to control an ideal diagnostic about some diseases. Specific features are then compared with the recently deployed conventional models in telemedicine. The main advantage of this model is the availability of controlling the security and privacy about patient information and the optimization on processing and acquiring clinical parameters according to technical characteristics.Keywords: Analytics, telemedicine, internet of things, cloud computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156513979 A Stochastic Diffusion Process Based on the Two-Parameters Weibull Density Function
Authors: Meriem Bahij, Ahmed Nafidi, Boujemâa Achchab, Sílvio M. A. Gama, José A. O. Matos
Abstract:
Stochastic modeling concerns the use of probability to model real-world situations in which uncertainty is present. Therefore, the purpose of stochastic modeling is to estimate the probability of outcomes within a forecast, i.e. to be able to predict what conditions or decisions might happen under different situations. In the present study, we present a model of a stochastic diffusion process based on the bi-Weibull distribution function (its trend is proportional to the bi-Weibull probability density function). In general, the Weibull distribution has the ability to assume the characteristics of many different types of distributions. This has made it very popular among engineers and quality practitioners, who have considered it the most commonly used distribution for studying problems such as modeling reliability data, accelerated life testing, and maintainability modeling and analysis. In this work, we start by obtaining the probabilistic characteristics of this model, as the explicit expression of the process, its trends, and its distribution by transforming the diffusion process in a Wiener process as shown in the Ricciaardi theorem. Then, we develop the statistical inference of this model using the maximum likelihood methodology. Finally, we analyse with simulated data the computational problems associated with the parameters, an issue of great importance in its application to real data with the use of the convergence analysis methods. Overall, the use of a stochastic model reflects only a pragmatic decision on the part of the modeler. According to the data that is available and the universe of models known to the modeler, this model represents the best currently available description of the phenomenon under consideration.Keywords: Diffusion process, discrete sampling, likelihood estimation method, simulation, stochastic diffusion equation, trends functions, bi-parameters Weibull density function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 196713978 Analysis and Prediction of the Behavior of the Landslide at Ain El Hammam, Algeria Based on the Second Order Work Criterion
Authors: Zerarka Hizia, Akchiche Mustapha, Prunier Florent
Abstract:
The landslide of Ain El Hammam (AEH) is characterized by a complex geology and a high hydrogeology hazard. AEH's perpetual reactivation compels us to look closely at its triggers and to better understand the mechanisms of its evolution in mass and in depth. This study builds a numerical model to simulate the influencing factors such as precipitation, non-saturation, and pore pressure fluctuations, using Plaxis software. For a finer analysis of instabilities, we use Hill's criterion, based on the sign of the second order work, which is the most appropriate material stability criterion for non-associated elastoplastic materials. The results of this type of calculation allow us, in theory, to predict the shape and position of the slip surface(s) which are liable to ground movements of the slope, before reaching the rupture given by the plastic limit of Mohr Coulomb. To validate the numerical model, an analysis of inclinometer measures is performed to confirm the direction of movement and kinematic of the sliding mechanism of AEH’s slope.Keywords: Landslide, second order work, precipitation, inclinometers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 111313977 Lineup Optimization Model of Basketball Players Based on the Prediction of Recursive Neural Networks
Authors: Wang Yichen, Haruka Yamashita
Abstract:
In recent years, in the field of sports, decision making such as member in the game and strategy of the game based on then analysis of the accumulated sports data are widely attempted. In fact, in the NBA basketball league where the world's highest level players gather, to win the games, teams analyze the data using various statistical techniques. However, it is difficult to analyze the game data for each play such as the ball tracking or motion of the players in the game, because the situation of the game changes rapidly, and the structure of the data should be complicated. Therefore, it is considered that the analysis method for real time game play data is proposed. In this research, we propose an analytical model for "determining the optimal lineup composition" using the real time play data, which is considered to be difficult for all coaches. In this study, because replacing the entire lineup is too complicated, and the actual question for the replacement of players is "whether or not the lineup should be changed", and “whether or not Small Ball lineup is adopted”. Therefore, we propose an analytical model for the optimal player selection problem based on Small Ball lineups. In basketball, we can accumulate scoring data for each play, which indicates a player's contribution to the game, and the scoring data can be considered as a time series data. In order to compare the importance of players in different situations and lineups, we combine RNN (Recurrent Neural Network) model, which can analyze time series data, and NN (Neural Network) model, which can analyze the situation on the field, to build the prediction model of score. This model is capable to identify the current optimal lineup for different situations. In this research, we collected all the data of accumulated data of NBA from 2019-2020. Then we apply the method to the actual basketball play data to verify the reliability of the proposed model.Keywords: Recurrent Neural Network, players lineup, basketball data, decision making model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82813976 Impact of Financial System’s Development on Economic Development: An Empirical Investigation
Authors: Vilma Deltuvaitė
Abstract:
Comparisons of financial development across countries are central to answering many of the questions on factors leading to economic development. For this reason this study analyzes the implications of financial system’s development on country’s economic development. The aim of the article: to analyze the impact of financial system’s development on economic development. The following research methods were used: systemic, logical and comparative analysis of scientific literature, analysis of statistical data, time series model (Autoregressive Distributed Lag (ARDL) Model). The empirical results suggest about positive short and long term effect of stock market development on GDP per capita.
Keywords: Banking sector, economic development, financial system’s development, stock market, private bond market.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 212413975 Construction Unit Rate Factor Modelling Using Neural Networks
Authors: Balimu Mwiya, Mundia Muya, Chabota Kaliba, Peter Mukalula
Abstract:
Factors affecting construction unit cost vary depending on a country’s political, economic, social and technological inclinations. Factors affecting construction costs have been studied from various perspectives. Analysis of cost factors requires an appreciation of a country’s practices. Identified cost factors provide an indication of a country’s construction economic strata. The purpose of this paper is to identify the essential factors that affect unit cost estimation and their breakdown using artificial neural networks. Twenty five (25) identified cost factors in road construction were subjected to a questionnaire survey and employing SPSS factor analysis the factors were reduced to eight. The 8 factors were analysed using neural network (NN) to determine the proportionate breakdown of the cost factors in a given construction unit rate. NN predicted that political environment accounted 44% of the unit rate followed by contractor capacity at 22% and financial delays, project feasibility and overhead & profit each at 11%. Project location, material availability and corruption perception index had minimal impact on the unit cost from the training data provided. Quantified cost factors can be incorporated in unit cost estimation models (UCEM) to produce more accurate estimates. This can create improvements in the cost estimation of infrastructure projects and establish a benchmark standard to assist the process of alignment of work practises and training of new staff, permitting the on-going development of best practises in cost estimation to become more effective.
Keywords: Construction cost factors, neural networks, roadworks, Zambian Construction Industry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 382513974 Landslide Susceptibility Mapping: A Comparison between Logistic Regression and Multivariate Adaptive Regression Spline Models in the Municipality of Oudka, Northern of Morocco
Authors: S. Benchelha, H. C. Aoudjehane, M. Hakdaoui, R. El Hamdouni, H. Mansouri, T. Benchelha, M. Layelmam, M. Alaoui
Abstract:
The logistic regression (LR) and multivariate adaptive regression spline (MarSpline) are applied and verified for analysis of landslide susceptibility map in Oudka, Morocco, using geographical information system. From spatial database containing data such as landslide mapping, topography, soil, hydrology and lithology, the eight factors related to landslides such as elevation, slope, aspect, distance to streams, distance to road, distance to faults, lithology map and Normalized Difference Vegetation Index (NDVI) were calculated or extracted. Using these factors, landslide susceptibility indexes were calculated by the two mentioned methods. Before the calculation, this database was divided into two parts, the first for the formation of the model and the second for the validation. The results of the landslide susceptibility analysis were verified using success and prediction rates to evaluate the quality of these probabilistic models. The result of this verification was that the MarSpline model is the best model with a success rate (AUC = 0.963) and a prediction rate (AUC = 0.951) higher than the LR model (success rate AUC = 0.918, rate prediction AUC = 0.901).
Keywords: Landslide susceptibility mapping, regression logistic, multivariate adaptive regression spline, Oudka, Taounate, Morocco.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 98913973 Behavioral Modeling Accuracy for RF Power Amplifier with Memory Effects
Authors: Chokri Jebali, Noureddine Boulejfen, Ali Gharsallah, Fadhel M. Ghannouchi
Abstract:
In this paper, a system level behavioural model for RF power amplifier, which exhibits memory effects, and based on multibranch system is proposed. When higher order terms are included, the memory polynomial model (MPM) exhibits numerical instabilities. A set of memory orthogonal polynomial model (OMPM) is introduced to alleviate the numerical instability problem associated to MPM model. A data scaling and centring algorithm was applied to improve the power amplifier modeling accuracy. Simulation results prove that the numerical instability can be greatly reduced, as well as the model precision improved with nonlinear model.Keywords: power amplifier, orthogonal model, polynomialmodel , memory effects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 227713972 An Output Oriented Super-Efficiency Model for Considering Time Lag Effect
Authors: Yanshuang Zhang, Byungho Jeong
Abstract:
There exists some time lag between the consumption of inputs and the production of outputs. This time lag effect should be considered in calculating efficiency of decision making units (DMU). Recently, a couple of DEA models were developed for considering time lag effect in efficiency evaluation of research activities. However, these models can’t discriminate efficient DMUs because of the nature of basic DEA model in which efficiency scores are limited to ‘1’. This problem can be resolved a super-efficiency model. However, a super efficiency model sometimes causes infeasibility problem. This paper suggests an output oriented super-efficiency model for efficiency evaluation under the consideration of time lag effect. A case example using a long term research project is given to compare the suggested model with the MpO model.
Keywords: DEA, Super-efficiency, Time Lag.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 268213971 Teaching Attentive Literature Reading in Higher Education French as a Foreign Language: A Pilot Study of a Flipped Classroom Teaching Model
Authors: Malin Isaksson
Abstract:
Teaching French as a foreign language usually implies teaching French literature, especially in higher education. Training university students in literary reading in a foreign language requires addressing several aspects at the same time: the (foreign) language, the poetic language, the aesthetic aspects of the studied works, and various interpretations of them. A pilot study sought to test a teaching model that would support students in learning to perform competent readings and short analyses of French literary works, in a rather independent manner. This shared practice paper describes the use of a flipped classroom method in two French literature courses, a campus course and an online course, and suggests that the teaching model may provide efficient tools for teaching literary reading and analysis in a foreign language. The teaching model builds on a high level of student activity and focuses on attentive reading, meta-perspectives such as theoretical concepts, individual analyses by students where said concepts are applied, and group discussions of the studied texts and of possible interpretations.
Keywords: Shared practice, flipped classroom, literature in foreign language studies, teaching literature analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 77013970 Individual Learning and Collaborative Knowledge Building with Shared Digital Artifacts
Authors: Joachim Kimmerle, Johannes Moskaliuk, Ulrike Cress
Abstract:
The development of Internet technology in recent years has led to a more active role of users in creating Web content. This has significant effects both on individual learning and collaborative knowledge building. This paper will present an integrative framework model to describe and explain learning and knowledge building with shared digital artifacts on the basis of Luhmann-s systems theory and Piaget-s model of equilibration. In this model, knowledge progress is based on cognitive conflicts resulting from incongruities between an individual-s prior knowledge and the information which is contained in a digital artifact. Empirical support for the model will be provided by 1) applying it descriptively to texts from Wikipedia, 2) examining knowledge-building processes using a social network analysis, and 3) presenting a survey of a series of experimental laboratory studies.
Keywords: Individual learning, collaborative knowledge building, systems theory, equilibration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 163013969 Study on Applying Fuzzy AHP and GRA in Selection of Agent Construction Enterprise
Authors: Shirong Li, Huan Yan
Abstract:
To help the client to select a competent agent construction enterprise (ACE), this study aims to investigate the selection standards by using the Fuzzy Analytic Hierarchy Process (FAHP) and build an evaluation mathematical model with Grey Relational Analysis (GRA). According to the outputs of literature review, four orderly levels are established within the model, taking the consideration of various agent construction models in practice. Then, the process of applying FAHP and GRA is discussed in detailed. Finally, through a case study, this paper illustrates how to apply these methods in getting the weights of each standard and the final assessment result.Keywords: agent construction enterprise, agent constructionmodel, fuzzy analytic hierarchy process, grey relational analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 204313968 Hybrid Structure Learning Approach for Assessing the Phosphate Laundries Impact
Authors: Emna Benmohamed, Hela Ltifi, Mounir Ben Ayed
Abstract:
Bayesian Network (BN) is one of the most efficient classification methods. It is widely used in several fields (i.e., medical diagnostics, risk analysis, bioinformatics research). The BN is defined as a probabilistic graphical model that represents a formalism for reasoning under uncertainty. This classification method has a high-performance rate in the extraction of new knowledge from data. The construction of this model consists of two phases for structure learning and parameter learning. For solving this problem, the K2 algorithm is one of the representative data-driven algorithms, which is based on score and search approach. In addition, the integration of the expert's knowledge in the structure learning process allows the obtainment of the highest accuracy. In this paper, we propose a hybrid approach combining the improvement of the K2 algorithm called K2 algorithm for Parents and Children search (K2PC) and the expert-driven method for learning the structure of BN. The evaluation of the experimental results, using the well-known benchmarks, proves that our K2PC algorithm has better performance in terms of correct structure detection. The real application of our model shows its efficiency in the analysis of the phosphate laundry effluents' impact on the watershed in the Gafsa area (southwestern Tunisia).
Keywords: Classification, Bayesian network; structure learning, K2 algorithm, expert knowledge, surface water analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 51213967 Web Personalization to Build Trust in E-Commerce: A Design Science Approach
Authors: Choon Ling Sia, Yani Shi, Jiaqi Yan, Huaping Chen
Abstract:
With the development of the Internet, E-commerce is growing at an exponential rate, and lots of online stores are built up to sell their goods online. A major factor influencing the successful adoption of E-commerce is consumer-s trust. For new or unknown Internet business, consumers- lack of trust has been cited as a major barrier to its proliferation. As web sites provide key interface for consumer use of E-Commerce, we investigate the design of web site to build trust in E-Commerce from a design science approach. A conceptual model is proposed in this paper to describe the ontology of online transaction and human-computer interaction. Based on this conceptual model, we provide a personalized webpage design approach using Bayesian networks learning method. Experimental evaluation are designed to show the effectiveness of web personalization in improving consumer-s trust in new or unknown online store.Keywords: Trust, Web site design, Human-ComputerInteraction, E-Commerce, Design science, Bayesian network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 200313966 An Aggregate Production Planning Model for Brass Casting Industry in Fuzzy Environment
Authors: Ömer Faruk Baykoç, Ümit Sami Sakalli
Abstract:
In this paper, we propose a fuzzy aggregate production planning (APP) model for blending problem in a brass factory which is the problem of computing optimal amounts of raw materials for the total production of several types of brass in a period. The model has deterministic and imprecise parameters which follows triangular possibility distributions. The brass casting APP model can not always be solved by using common approaches used in the literature. Therefore a mathematical model is presented for solving this problem. In the proposed model, the Lai and Hwang-s fuzzy ranking concept is relaxed by using one constraint instead of three constraints. An application of the brass casting APP model in a brass factory shows that the proposed model successfully solves the multi-blend problem in casting process and determines the optimal raw material purchasing policies.Keywords: Aggregate production planning, Blending, brasscasting, possibilistic programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 190813965 Framework Study on Single Assembly Line to Improve Productivity with Six Sigma and Line Balancing Approach
Authors: Inaki Maulida Hakim, T. Yuri M. Zagloel, Astari Wulandari
Abstract:
Six sigma is a framework that is used to identify inefficiency so that the cause of inefficiency will be known and right improvement to overcome cause of inefficiency can be conducted. This paper presents result of implementing six sigma to improve piston assembly line in Manufacturing Laboratory, Universitas Indonesia. Six sigma framework will be used to analyze the significant factor of inefficiency that needs to be improved which causes bottleneck in assembly line. After analysis based on six sigma framework conducted, line balancing method was chosen for improvement to overcome causative factor of inefficiency which is differences time between workstation that causes bottleneck in assembly line. Then after line balancing conducted in piston assembly line, the result is increase in efficiency. Efficiency is shown in the decreasing of Defects per Million Opportunities (DPMO) from 900,000 to 700,000, the increasing of level of labor productivity from 0.0041 to 0.00742, the decreasing of idle time from 121.3 seconds to 12.1 seconds, and the increasing of output, which is from 1 piston in 5 minutes become 3 pistons in 5 minutes.
Keywords: Assembly line, efficiency, improvement, line balancing, productivity, six sigma, workstation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1820