Search results for: discrete choice models
7715 Creation and Management of Knowledge for Organization Sustainability and Learning
Authors: Deepa Kapoor, Rajshree Singh
Abstract:
This paper appreciates the emergence and growing importance as a new production factor makes the development of technologies, methodologies and strategies for measurement, creation, and diffusion into one of the main priorities of the organizations in the knowledge society. There are many models for creation and management of knowledge and diverse and varied perspectives for study, analysis, and understanding. In this article, we will conduct a theoretical approach to the type of models for the creation and management of knowledge; we will discuss some of them and see some of the difficulties and the key factors that determine the success of the processes for the creation and management of knowledge.Keywords: knowledge creation, knowledge management, organizational development, organization learning
Procedia PDF Downloads 3457714 Removal of Toxic Ni++ Ions from Wastewater by Nano-Bentonite
Authors: A. M. Ahmed, Mona A. Darwish
Abstract:
Removal of Ni++ ions from aqueous solution by sorption ontoNano-bentonite was investigated. Experiments were carried out as a function amount of Nano-bentonite, pH, concentration of metal, constant time, agitation speed and temperature. The adsorption parameter of metal ions followed the Langmuir Freundlich adsorption isotherm were applied to analyze adsorption data. The adsorption process has fit pseudo-second order kinetic models. Thermodynamics parameters e.g.ΔG*, ΔS °and ΔH ° of adsorption process have also been calculated and the sorption process was found to be endothermic. The adsorption process has fit pseudo-second order kinetic models. Langmuir and Freundich adsorption isotherm models were applied to analyze adsorption data and both were found to be applicable to the adsorption process. Thermodynamic parameters, e.g., ∆G °, ∆S ° and ∆H ° of the on-going adsorption process have also been calculated and the sorption process was found to be endothermic. Finally, it can be seen that Bentonite was found to be more effective for the removal of Ni (II) same with some experimental conditions.Keywords: waste water, nickel, bentonite, adsorption
Procedia PDF Downloads 2587713 Analyzing the Performance of Machine Learning Models to Predict Alzheimer's Disease and its Stages Addressing Missing Value Problem
Authors: Carlos Theran, Yohn Parra Bautista, Victor Adankai, Richard Alo, Jimwi Liu, Clement G. Yedjou
Abstract:
Alzheimer's disease (AD) is a neurodegenerative disorder primarily characterized by deteriorating cognitive functions. AD has gained relevant attention in the last decade. An estimated 24 million people worldwide suffered from this disease by 2011. In 2016 an estimated 40 million were diagnosed with AD, and for 2050 is expected to reach 131 million people affected by AD. Therefore, detecting and confirming AD at its different stages is a priority for medical practices to provide adequate and accurate treatments. Recently, Machine Learning (ML) models have been used to study AD's stages handling missing values in multiclass, focusing on the delineation of Early Mild Cognitive Impairment (EMCI), Late Mild Cognitive Impairment (LMCI), and normal cognitive (CN). But, to our best knowledge, robust performance information of these models and the missing data analysis has not been presented in the literature. In this paper, we propose studying the performance of five different machine learning models for AD's stages multiclass prediction in terms of accuracy, precision, and F1-score. Also, the analysis of three imputation methods to handle the missing value problem is presented. A framework that integrates ML model for AD's stages multiclass prediction is proposed, performing an average accuracy of 84%.Keywords: alzheimer's disease, missing value, machine learning, performance evaluation
Procedia PDF Downloads 2527712 Wireworms under the Sword of Damocles: Attraction to Maize Root Volatiles
Authors: Diana La Forgia, Jean Baptiste Thibord, François Verheggen
Abstract:
Volatiles Organic Compound (VOCs) are one of the many features of defense used by plants in their eternal fight against pests. Their main role is to attract the natural enemies of the herbivores. But on another hand, they can be used by the same herbivores to locate plants while foraging. In an attempt to fill a gap of knowledge in a complex web of interactions, we focused on wireworms (Coleoptera:Elateridae). Wireworms whose larvae feed on roots are one of the most spread pests of valuable crops such as maize and potatoes, causing important economical damage. Little is known about the root compounds that are playing a role in the attraction of the larvae. In order to know more about these compounds, we compared four different maize varieties (Zea mays mays) that are known to have different levels of attraction, from weak to strong, for wireworms in fields. We tested the attraction of larvae in laboratory conditions in dual-choice olfactometer assays where they were offered all possible combinations of the four maize varieties. Contemporary, we collected the VOCs of each variety during 24h using a push-and-pull system. The collected samples were then analyzed by gas chromatography coupled with a mass spectrometer (GC-MS) to identify their molecular profiles. The choice of the larvae was dependent on the offered combination and some varieties were preferred to others. Differences were also observed in terms of quantitative and qualitative emissions of volatile profiles between the maize varieties. Our aim is to develop traps based on VOCs from maize roots to open a new frontier in wireworms management.Keywords: integrated pest management, maize roots, plant defense, volatile organic compounds, wireworms
Procedia PDF Downloads 1557711 Persian Pistachio Nut (Pistacia vera L.) Dehydration in Natural and Industrial Conditions
Authors: Hamid Tavakolipour, Mohsen Mokhtarian, Ahmad Kalbasi Ashtari
Abstract:
In this study, the effect of various drying methods (sun drying, shade drying and industrial drying) on final moisture content, shell splitting degree, shrinkage and color change were studied. Sun drying resulted higher degree of pistachio nuts shell splitting on pistachio nuts relative other drying methods. The ANOVA results showed that the different drying methods did not significantly effects on color change of dried pistachio nut. The results illustrated that pistachio nut dried by industrial drying had the lowest moisture content. After the end of drying process, initially, the experimental drying data were fitted with five famous drying models namely Newton, Page, Silva et al., Peleg and Henderson and Pabis. The results indicated that Peleg and Page models gave better results compared with other models to monitor the moisture ratio’s pistachio nut in industrial drying and open sun (or shade drying) methods, respectively.Keywords: industrial drying, pistachio, quality properties, traditional drying
Procedia PDF Downloads 3357710 Existence and Uniqueness of Solutions to Singular Higher Order Two-Point BVPs on Time Scales
Authors: Zhenjie Liu
Abstract:
This paper investigates the existence and uniqueness of solutions for singular higher order boundary value problems on time scales by using mixed monotone method. The theorems obtained are very general. For the different time scale, the problem may be the corresponding continuous or discrete boundary value problem.Keywords: mixed monotone operator, boundary value problem, time scale, green's function, positive solution, singularity
Procedia PDF Downloads 2577709 Credit Risk Evaluation Using Genetic Programming
Authors: Ines Gasmi, Salima Smiti, Makram Soui, Khaled Ghedira
Abstract:
Credit risk is considered as one of the important issues for financial institutions. It provokes great losses for banks. To this objective, numerous methods for credit risk evaluation have been proposed. Many evaluation methods are black box models that cannot adequately reveal information hidden in the data. However, several works have focused on building transparent rules-based models. For credit risk assessment, generated rules must be not only highly accurate, but also highly interpretable. In this paper, we aim to build both, an accurate and transparent credit risk evaluation model which proposes a set of classification rules. In fact, we consider the credit risk evaluation as an optimization problem which uses a genetic programming (GP) algorithm, where the goal is to maximize the accuracy of generated rules. We evaluate our proposed approach on the base of German and Australian credit datasets. We compared our finding with some existing works; the result shows that the proposed GP outperforms the other models.Keywords: credit risk assessment, rule generation, genetic programming, feature selection
Procedia PDF Downloads 3537708 Introducing a Practical Model for Instructional System Design Based on Determining of the knowledge Level of the Organization: Case Study of Isfahan Public Transportation Co.
Authors: Mojtaba Aghajari, Alireza Aghasi
Abstract:
The first challenge which the current research faced has been the identification or determination of the level of knowledge in Isfahan public transportation corporation, and the second challenge has been the recognition and choice of a proper approach for the instructional system design. Responding these two challenges will present an appropriate model of instructional system design. In order to respond the first challenge or question, Nonaka and Takeuchi KM model has been utilized due to its universality among the 26 models proposed so far. The statistical population of this research included 2200 people, among which 200 persons were chosen as the sample of the research by the use of Morgan’s method. The data gathering has been carried out by the means of a questionnaire based on Nonaka and Takeuchi KM model, analysis of which has been done by SPSS program. The output of this questionnaire, yielding the point of 1.96 (out of 5 points), revealed that the general condition of Isfahan public transportation corporation is weak concerning its being knowledge-centered. After placing this output on Jonassen’s continuum, it was revealed that the appropriate approach for instructional system design is the system (or behavioral) approach. Accordingly, different steps of the general model of ADDIE, which covers all of the ISO10015 standards, were adopted in the act of designing. Such process in Isfahan public transportation corporation was designed and divided into three main steps, including: instructional designing and planning, instructional course planning, determination of the evaluation and the effectiveness of the instructional courses.Keywords: instructional system design, system approach, knowledge management, employees
Procedia PDF Downloads 3267707 An Infinite Mixture Model for Modelling Stutter Ratio in Forensic Data Analysis
Authors: M. A. C. S. Sampath Fernando, James M. Curran, Renate Meyer
Abstract:
Forensic DNA analysis has received much attention over the last three decades, due to its incredible usefulness in human identification. The statistical interpretation of DNA evidence is recognised as one of the most mature fields in forensic science. Peak heights in an Electropherogram (EPG) are approximately proportional to the amount of template DNA in the original sample being tested. A stutter is a minor peak in an EPG, which is not masking as an allele of a potential contributor, and considered as an artefact that is presumed to be arisen due to miscopying or slippage during the PCR. Stutter peaks are mostly analysed in terms of stutter ratio that is calculated relative to the corresponding parent allele height. Analysis of mixture profiles has always been problematic in evidence interpretation, especially with the presence of PCR artefacts like stutters. Unlike binary and semi-continuous models; continuous models assign a probability (as a continuous weight) for each possible genotype combination, and significantly enhances the use of continuous peak height information resulting in more efficient reliable interpretations. Therefore, the presence of a sound methodology to distinguish between stutters and real alleles is essential for the accuracy of the interpretation. Sensibly, any such method has to be able to focus on modelling stutter peaks. Bayesian nonparametric methods provide increased flexibility in applied statistical modelling. Mixture models are frequently employed as fundamental data analysis tools in clustering and classification of data and assume unidentified heterogeneous sources for data. In model-based clustering, each unknown source is reflected by a cluster, and the clusters are modelled using parametric models. Specifying the number of components in finite mixture models, however, is practically difficult even though the calculations are relatively simple. Infinite mixture models, in contrast, do not require the user to specify the number of components. Instead, a Dirichlet process, which is an infinite-dimensional generalization of the Dirichlet distribution, is used to deal with the problem of a number of components. Chinese restaurant process (CRP), Stick-breaking process and Pólya urn scheme are frequently used as Dirichlet priors in Bayesian mixture models. In this study, we illustrate an infinite mixture of simple linear regression models for modelling stutter ratio and introduce some modifications to overcome weaknesses associated with CRP.Keywords: Chinese restaurant process, Dirichlet prior, infinite mixture model, PCR stutter
Procedia PDF Downloads 3307706 Empirical Studies of Indigenous Career Choice in Taiwan
Authors: Zichun Chu
Abstract:
The issue of tribal poverty has always attracted attentions. Due to social and economic difficulties, the indigenous people's personal development and tribal development have been greatly restricted. Past studies have pointed out that poverty may come from a lack of education. The United Nations Sustainable Development Goals (SDGs) also stated that if we are to solve the poverty problem, providing education widely is an important key. According to the theory of intellectual capital adaptation, “being capable” and “willing to do” are the keys of development. Therefore, we can say that the "ability" and "will" of tribal residents for their tribal development is the core concern of the tribal development. This research was designed to investigate the career choice development model of indigenous tribe people by investigating the current status of human capital, social capital, and cultural capital of tribal residents. This study collected 327 questionnaires (70% of total households) from Truku tribe to answer the research question: Did education help them for job choosing decisions from the aspects of human capital, social capital, and cultural capital in tribal status. This project highlighted the ‘single tribal research approach’ to gain an in-depth understanding of the human capital formed under the unique culture of the tribe (Truku tribe). The results show that the education level of most research participants was high school, very few high school graduates chose to further their education to college level; due to the lack of education of their parents, the social capital was limited to support them for jobs choice, most of them work for labor and service industries; however, their culture capital was comparably rich for works, the sharing culture of Taiwanese indigenous people made their work status stable. The results suggested that we should emphasize more on the development of vocational education based on the tribe’s location and resources. The self-advocacy of indigenous people should be developed so that they would gain more power on making career decisions. This research project is part of a pilot project called “INDIGENOUS PEOPLES, POVERTY, AND DEVELOPMENT,” sponsored by the National Science and Technology Council of Taiwan. If this paper were accepted to present in the 2023 ICIP, it would be lovely if a panel is formed for me and other co-researchers (Chuanju Cheng, Chih-Yuan Weng, and YiXuan Chen), for the audience will be able to get a full picture of this pilot project.Keywords: career choices, career model, indegenous career development, indigenous education, tribe
Procedia PDF Downloads 827705 Adaptive Neuro Fuzzy Inference System Model Based on Support Vector Regression for Stock Time Series Forecasting
Authors: Anita Setianingrum, Oki S. Jaya, Zuherman Rustam
Abstract:
Forecasting stock price is a challenging task due to the complex time series of the data. The complexity arises from many variables that affect the stock market. Many time series models have been proposed before, but those previous models still have some problems: 1) put the subjectivity of choosing the technical indicators, and 2) rely upon some assumptions about the variables, so it is limited to be applied to all datasets. Therefore, this paper studied a novel Adaptive Neuro-Fuzzy Inference System (ANFIS) time series model based on Support Vector Regression (SVR) for forecasting the stock market. In order to evaluate the performance of proposed models, stock market transaction data of TAIEX and HIS from January to December 2015 is collected as experimental datasets. As a result, the method has outperformed its counterparts in terms of accuracy.Keywords: ANFIS, fuzzy time series, stock forecasting, SVR
Procedia PDF Downloads 2477704 Comparison of Fundamental Frequency Model and PWM Based Model for UPFC
Authors: S. A. Al-Qallaf, S. A. Al-Mawsawi, A. Haider
Abstract:
Among all FACTS devices, the unified power flow controller (UPFC) is considered to be the most versatile device. This is due to its capability to control all the transmission system parameters (impedance, voltage magnitude, and phase angle). With the growing interest in UPFC, the attention to develop a mathematical model has increased. Several models were introduced for UPFC in literature for different type of studies in power systems. In this paper a novel comparison study between two dynamic models of UPFC with their proposed control strategies.Keywords: FACTS, UPFC, dynamic modeling, PWM, fundamental frequency
Procedia PDF Downloads 3467703 dynr.mi: An R Program for Multiple Imputation in Dynamic Modeling
Authors: Yanling Li, Linying Ji, Zita Oravecz, Timothy R. Brick, Michael D. Hunter, Sy-Miin Chow
Abstract:
Assessing several individuals intensively over time yields intensive longitudinal data (ILD). Even though ILD provide rich information, they also bring other data analytic challenges. One of these is the increased occurrence of missingness with increased study length, possibly under non-ignorable missingness scenarios. Multiple imputation (MI) handles missing data by creating several imputed data sets, and pooling the estimation results across imputed data sets to yield final estimates for inferential purposes. In this article, we introduce dynr.mi(), a function in the R package, Dynamic Modeling in R (dynr). The package dynr provides a suite of fast and accessible functions for estimating and visualizing the results from fitting linear and nonlinear dynamic systems models in discrete as well as continuous time. By integrating the estimation functions in dynr and the MI procedures available from the R package, Multivariate Imputation by Chained Equations (MICE), the dynr.mi() routine is designed to handle possibly non-ignorable missingness in the dependent variables and/or covariates in a user-specified dynamic systems model via MI, with convergence diagnostic check. We utilized dynr.mi() to examine, in the context of a vector autoregressive model, the relationships among individuals’ ambulatory physiological measures, and self-report affect valence and arousal. The results from MI were compared to those from listwise deletion of entries with missingness in the covariates. When we determined the number of iterations based on the convergence diagnostics available from dynr.mi(), differences in the statistical significance of the covariate parameters were observed between the listwise deletion and MI approaches. These results underscore the importance of considering diagnostic information in the implementation of MI procedures.Keywords: dynamic modeling, missing data, mobility, multiple imputation
Procedia PDF Downloads 1647702 Monitoring Large-Coverage Forest Canopy Height by Integrating LiDAR and Sentinel-2 Images
Authors: Xiaobo Liu, Rakesh Mishra, Yun Zhang
Abstract:
Continuous monitoring of forest canopy height with large coverage is essential for obtaining forest carbon stocks and emissions, quantifying biomass estimation, analyzing vegetation coverage, and determining biodiversity. LiDAR can be used to collect accurate woody vegetation structure such as canopy height. However, LiDAR’s coverage is usually limited because of its high cost and limited maneuverability, which constrains its use for dynamic and large area forest canopy monitoring. On the other hand, optical satellite images, like Sentinel-2, have the ability to cover large forest areas with a high repeat rate, but they do not have height information. Hence, exploring the solution of integrating LiDAR data and Sentinel-2 images to enlarge the coverage of forest canopy height prediction and increase the prediction repeat rate has been an active research topic in the environmental remote sensing community. In this study, we explore the potential of training a Random Forest Regression (RFR) model and a Convolutional Neural Network (CNN) model, respectively, to develop two predictive models for predicting and validating the forest canopy height of the Acadia Forest in New Brunswick, Canada, with a 10m ground sampling distance (GSD), for the year 2018 and 2021. Two 10m airborne LiDAR-derived canopy height models, one for 2018 and one for 2021, are used as ground truth to train and validate the RFR and CNN predictive models. To evaluate the prediction performance of the trained RFR and CNN models, two new predicted canopy height maps (CHMs), one for 2018 and one for 2021, are generated using the trained RFR and CNN models and 10m Sentinel-2 images of 2018 and 2021, respectively. The two 10m predicted CHMs from Sentinel-2 images are then compared with the two 10m airborne LiDAR-derived canopy height models for accuracy assessment. The validation results show that the mean absolute error (MAE) for year 2018 of the RFR model is 2.93m, CNN model is 1.71m; while the MAE for year 2021 of the RFR model is 3.35m, and the CNN model is 3.78m. These demonstrate the feasibility of using the RFR and CNN models developed in this research for predicting large-coverage forest canopy height at 10m spatial resolution and a high revisit rate.Keywords: remote sensing, forest canopy height, LiDAR, Sentinel-2, artificial intelligence, random forest regression, convolutional neural network
Procedia PDF Downloads 927701 Experiments on Weakly-Supervised Learning on Imperfect Data
Authors: Yan Cheng, Yijun Shao, James Rudolph, Charlene R. Weir, Beth Sahlmann, Qing Zeng-Treitler
Abstract:
Supervised predictive models require labeled data for training purposes. Complete and accurate labeled data, i.e., a ‘gold standard’, is not always available, and imperfectly labeled data may need to serve as an alternative. An important question is if the accuracy of the labeled data creates a performance ceiling for the trained model. In this study, we trained several models to recognize the presence of delirium in clinical documents using data with annotations that are not completely accurate (i.e., weakly-supervised learning). In the external evaluation, the support vector machine model with a linear kernel performed best, achieving an area under the curve of 89.3% and accuracy of 88%, surpassing the 80% accuracy of the training sample. We then generated a set of simulated data and carried out a series of experiments which demonstrated that models trained on imperfect data can (but do not always) outperform the accuracy of the training data, e.g., the area under the curve for some models is higher than 80% when trained on the data with an error rate of 40%. Our experiments also showed that the error resistance of linear modeling is associated with larger sample size, error type, and linearity of the data (all p-values < 0.001). In conclusion, this study sheds light on the usefulness of imperfect data in clinical research via weakly-supervised learning.Keywords: weakly-supervised learning, support vector machine, prediction, delirium, simulation
Procedia PDF Downloads 1997700 Development of an Optimised, Automated Multidimensional Model for Supply Chains
Authors: Safaa H. Sindi, Michael Roe
Abstract:
This project divides supply chain (SC) models into seven Eras, according to the evolution of the market’s needs throughout time. The five earliest Eras describe the emergence of supply chains, while the last two Eras are to be created. Research objectives: The aim is to generate the two latest Eras with their respective models that focus on the consumable goods. Era Six contains the Optimal Multidimensional Matrix (OMM) that incorporates most characteristics of the SC and allocates them into four quarters (Agile, Lean, Leagile, and Basic SC). This will help companies, especially (SMEs) plan their optimal SC route. Era Seven creates an Automated Multidimensional Model (AMM) which upgrades the matrix of Era six, as it accounts for all the supply chain factors (i.e. Offshoring, sourcing, risk) into an interactive system with Heuristic Learning that helps larger companies and industries to select the best SC model for their market. Methodologies: The data collection is based on a Fuzzy-Delphi study that analyses statements using Fuzzy Logic. The first round of Delphi study will contain statements (fuzzy rules) about the matrix of Era six. The second round of Delphi contains the feedback given from the first round and so on. Preliminary findings: both models are applicable, Matrix of Era six reduces the complexity of choosing the best SC model for SMEs by helping them identify the best strategy of Basic SC, Lean, Agile and Leagile SC; that’s tailored to their needs. The interactive heuristic learning in the AMM of Era seven will help mitigate error and aid large companies to identify and re-strategize the best SC model and distribution system for their market and commodity, hence increasing efficiency. Potential contributions to the literature: The problematic issue facing many companies is to decide which SC model or strategy to incorporate, due to the many models and definitions developed over the years. This research simplifies this by putting most definition in a template and most models in the Matrix of era six. This research is original as the division of SC into Eras, the Matrix of Era six (OMM) with Fuzzy-Delphi and Heuristic Learning in the AMM of Era seven provides a synergy of tools that were not combined before in the area of SC. Additionally the OMM of Era six is unique as it combines most characteristics of the SC, which is an original concept in itself.Keywords: Leagile, automation, heuristic learning, supply chain models
Procedia PDF Downloads 3897699 Numerical Investigation of Two Turbulence Models for Predicting the Temperature Separation in Conical Vortex Tube
Authors: M. Guen
Abstract:
A three-dimensional numerical study is used to analyze the behavior of the flow inside a vortex tube. The vortex tube or Ranque-Hilsch vortex tube is a simple device which is capable of dividing compressed air from the inlet nozzle tangentially into two flow with different temperatures warm and cold. This phenomenon is known from literature by temperature separation. The K ω-SST and K-ε turbulence models are used to predict the turbulent flow behaviour inside the tube. The vortex tube is an Exair 708 slpm (25 scfm) commercial tube. The cold and hot exits areas are 30.2 and 95 mm2 respectively. The vortex nozzle consists of 6 straight slots; the height and the width of each slot are 0.97 mm and 1.41 mm. The total area normal to the flow associated with six nozzles is therefore 8.15 mm 2. The present study focuses on a comparison between two turbulence models K ω-SST, K-ε by using a new configuration of vortex tube (Conical Vortex Tube). The performance curves of the temperature separation versus cold outlet mass fraction were calculated and compared with experimental and numerical study of other researchers.Keywords: conical vortex tube, temperature separation, cold mass fraction, turbulence
Procedia PDF Downloads 2497698 Kinetics, Equilibrium and Thermodynamics of the Adsorption of Triphenyltin onto NanoSiO₂/Fly Ash/Activated Carbon Composite
Authors: Olushola S. Ayanda, Olalekan S. Fatoki, Folahan A. Adekola, Bhekumusa J. Ximba, Cecilia O. Akintayo
Abstract:
In the present study, the kinetics, equilibrium and thermodynamics of the adsorption of triphenyltin (TPT) from TPT-contaminated water onto nanoSiO2/fly ash/activated carbon composite was investigated in batch adsorption system. Equilibrium adsorption data were analyzed using Langmuir, Freundlich, Temkin and Dubinin–Radushkevich (D-R) isotherm models. Pseudo first- and second-order, Elovich and fractional power models were applied to test the kinetic data and in order to understand the mechanism of adsorption, thermodynamic parameters such as ΔG°, ΔSo and ΔH° were also calculated. The results showed a very good compliance with pseudo second-order equation while the Freundlich and D-R models fit the experiment data. Approximately 99.999 % TPT was removed from the initial concentration of 100 mg/L TPT at 80oC, contact time of 60 min, pH 8 and a stirring speed of 200 rpm. Thus, nanoSiO2/fly ash/activated carbon composite could be used as effective adsorbent for the removal of TPT from contaminated water and wastewater.Keywords: isotherm, kinetics, nanoSiO₂/fly ash/activated carbon composite, tributyltin
Procedia PDF Downloads 2937697 Decision-Making Under Uncertainty in Obsessive-Compulsive Disorder
Authors: Helen Pushkarskaya, David Tolin, Lital Ruderman, Ariel Kirshenbaum, J. MacLaren Kelly, Christopher Pittenger, Ifat Levy
Abstract:
Obsessive-Compulsive Disorder (OCD) produces profound morbidity. Difficulties with decision making and intolerance of uncertainty are prominent clinical features of OCD. The nature and etiology of these deficits are poorly understood. We used a well-validated choice task, grounded in behavioral economic theory, to investigate differences in valuation and value-based choice during decision making under uncertainty in 20 unmedicated participants with OCD and 20 matched healthy controls. Participants’ choices were used to assess individual decision-making characteristics. Compared to controls, individuals with OCD were less consistent in their choices and less able to identify options that were unambiguously preferable. These differences correlated with symptom severity. OCD participants did not differ from controls in how they valued uncertain options when outcome probabilities were known (risk) but were more likely than controls to avoid uncertain options when these probabilities were imprecisely specified (ambiguity). These results suggest that the underlying neural mechanisms of valuation and value-based choices during decision-making are abnormal in OCD. Individuals with OCD show elevated intolerance of uncertainty, but only when outcome probabilities are themselves uncertain. Future research focused on the neural valuation network, which is implicated in value-based computations, may provide new neurocognitive insights into the pathophysiology of OCD. Deficits in decision-making processes may represent a target for therapeutic intervention.Keywords: obsessive compulsive disorder, decision-making, uncertainty intolerance, risk aversion, ambiguity aversion, valuation
Procedia PDF Downloads 6157696 Strategy Management of Soybean (Glycine max L.) for Dealing with Extreme Climate through the Use of Cropsyst Model
Authors: Aminah Muchdar, Nuraeni, Eddy
Abstract:
The aims of the research are: (1) to verify the cropsyst plant model of experimental data in the field of soybean plants and (2) to predict planting time and potential yield soybean plant with the use of cropsyst model. This research is divided into several stages: (1) first calibration stage which conducted in the field from June until September 2015.(2) application models stage, where the data obtained from calibration in the field will be included in cropsyst models. The required data models are climate data, ground data/soil data,also crop genetic data. The relationship between the obtained result in field with simulation cropsyst model indicated by Efficiency Index (EF) which the value is 0,939.That is showing that cropsyst model is well used. From the calculation result RRMSE which the value is 1,922%.That is showing that comparative fault prediction results from simulation with result obtained in the field is 1,92%. The conclusion has obtained that the prediction of soybean planting time cropsyst based models that have been made valid for use. and the appropriate planting time for planting soybeans mainly on rain-fed land is at the end of the rainy season, in which the above study first planting time (June 2, 2015) which gives the highest production, because at that time there was still some rain. Tanggamus varieties more resistant to slow planting time cause the percentage decrease in the yield of each decade is lower than the average of all varieties.Keywords: soybean, Cropsyst, calibration, efficiency Index, RRMSE
Procedia PDF Downloads 1807695 Acceleration Techniques of DEM Simulation for Dynamics of Particle Damping
Authors: Masato Saeki
Abstract:
Presented herein is a novel algorithms for calculating the damping performance of particle dampers. The particle damper is a passive vibration control technique and has many practical applications due to simple design. It consists of granular materials constrained to move between two ends in the cavity of a primary vibrating system. The damping effect results from the exchange of momentum during the impact of granular materials against the wall of the cavity. This damping has the advantage of being independent of the environment. Therefore, particle damping can be applied in extreme temperature environments, where most conventional dampers would fail. It was shown experimentally in many papers that the efficiency of the particle dampers is high in the case of resonant vibration. In order to use the particle dampers effectively, it is necessary to solve the equations of motion for each particle, considering the granularity. The discrete element method (DEM) has been found to be effective for revealing the dynamics of particle damping. In this method, individual particles are assumed as rigid body and interparticle collisions are modeled by mechanical elements as springs and dashpots. However, the computational cost is significant since the equation of motion for each particle must be solved at each time step. In order to improve the computational efficiency of the DEM, the new algorithms are needed. In this study, new algorithms are proposed for implementing the high performance DEM. On the assumption that behaviors of the granular particles in the each divided area of the damper container are the same, the contact force of the primary system with all particles can be considered to be equal to the product of the divided number of the damper area and the contact force of the primary system with granular materials per divided area. This convenience makes it possible to considerably reduce the calculation time. The validity of this calculation method was investigated and the calculated results were compared with the experimental ones. This paper also presents the results of experimental studies of the performance of particle dampers. It is shown that the particle radius affect the noise level. It is also shown that the particle size and the particle material influence the damper performance.Keywords: particle damping, discrete element method (DEM), granular materials, numerical analysis, equivalent noise level
Procedia PDF Downloads 4537694 Birth Path and the Vitality of Caring Models in the Continuity of Midwifery
Authors: Elnaz Lalezari, Ramin Ghasemi Shaya
Abstract:
The birth way is influenced by a fracture within the quiet care handle, making a brokenness of this final one. The pregnant lady has got to interface with numerous experts, both amid the pregnancy, the childbirth, and the puerperium. Be that as it may, amid the final ten a long time, there has been an expanding of the pregnancy care worked by the midwife, who is considered to be the administrator with the correct competences, who can beware of each pregnancy and may profit herself of other professionals' commitments in arrange to make strides the results of maternal and neonatal health. To confirm whether there are proofs of viability that bolster the caseload birthing assistance care show, and in case it is conceivable to apply this show within the birth way in Italy. A amendment of writing has been done utilizing a few look motor (Google, Bing) and particular databases (MEDLINE, CINAHL, Embase, Domestic - ClinicalTrials.gov). There has, too, been a discussion of the Italian directions, the national rules, and the proposals of WHO. Results: The look string, legitimately adjusted to the three databases, has given the taking after comes about: MEDLINE 64 articles, CINAHL 94 articles, Embase 88 articles. From this choice, 14 articles have been extricated: 1 orderly survey, 3 controlled arbitrary trial, 7 observational ponders, 3 subjective studies. The caseload maternity care appears to be an successful and dependable organisational/caring strategy. It reacts to the criterions of quality and security, to the requirements of ladies not as it were amid the pregnancy but moreover amid the post-partum stage. For these reasons, it appears exceptionally valuable also for the birth way within the Italian reality.Keywords: midwifery, care, caseload, maternity
Procedia PDF Downloads 1327693 A Parallel Approach for 3D-Variational Data Assimilation on GPUs in Ocean Circulation Models
Authors: Rossella Arcucci, Luisa D'Amore, Simone Celestino, Giuseppe Scotti, Giuliano Laccetti
Abstract:
This work is the first dowel in a rather wide research activity in collaboration with Euro Mediterranean Center for Climate Changes, aimed at introducing scalable approaches in Ocean Circulation Models. We discuss designing and implementation of a parallel algorithm for solving the Variational Data Assimilation (DA) problem on Graphics Processing Units (GPUs). The algorithm is based on the fully scalable 3DVar DA model, previously proposed by the authors, which uses a Domain Decomposition approach (we refer to this model as the DD-DA model). We proceed with an incremental porting process consisting of 3 distinct stages: requirements and source code analysis, incremental development of CUDA kernels, testing and optimization. Experiments confirm the theoretic performance analysis based on the so-called scale up factor demonstrating that the DD-DA model can be suitably mapped on GPU architectures.Keywords: data assimilation, GPU architectures, ocean models, parallel algorithm
Procedia PDF Downloads 4127692 Kalman Filter for Bilinear Systems with Application
Authors: Abdullah E. Al-Mazrooei
Abstract:
In this paper, we present a new kind of the bilinear systems in the form of state space model. The evolution of this system depends on the product of state vector by its self. The well known Lotak Volterra and Lorenz models are special cases of this new model. We also present here a generalization of Kalman filter which is suitable to work with the new bilinear model. An application to real measurements is introduced to illustrate the efficiency of the proposed algorithm.Keywords: bilinear systems, state space model, Kalman filter, application, models
Procedia PDF Downloads 4417691 3D Numerical Study of Tsunami Loading and Inundation in a Model Urban Area
Authors: A. Bahmanpour, I. Eames, C. Klettner, A. Dimakopoulos
Abstract:
We develop a new set of diagnostic tools to analyze inundation into a model district using three-dimensional CFD simulations, with a view to generating a database against which to test simpler models. A three-dimensional model of Oregon city with different-sized groups of building next to the coastline is used to run calculations of the movement of a long period wave on the shore. The initial and boundary conditions of the off-shore water are set using a nonlinear inverse method based on Eulerian spatial information matching experimental Eulerian time series measurements of water height. The water movement is followed in time, and this enables the pressure distribution on every surface of each building to be followed in a temporal manner. The three-dimensional numerical data set is validated against published experimental work. In the first instance, we use the dataset as a basis to understand the success of reduced models - including 2D shallow water model and reduced 1D models - to predict water heights, flow velocity and forces. This is because models based on the shallow water equations are known to underestimate drag forces after the initial surge of water. The second component is to identify critical flow features, such as hydraulic jumps and choked states, which are flow regions where dissipation occurs and drag forces are large. Finally, we describe how future tsunami inundation models should be modified to account for the complex effects of buildings through drag and blocking.Financial support from UCL and HR Wallingford is greatly appreciated. The authors would like to thank Professor Daniel Cox and Dr. Hyoungsu Park for providing the data on the Seaside Oregon experiment.Keywords: computational fluid dynamics, extreme events, loading, tsunami
Procedia PDF Downloads 1157690 Housing Price Prediction Using Machine Learning Algorithms: The Case of Melbourne City, Australia
Authors: The Danh Phan
Abstract:
House price forecasting is a main topic in the real estate market research. Effective house price prediction models could not only allow home buyers and real estate agents to make better data-driven decisions but may also be beneficial for the property policymaking process. This study investigates the housing market by using machine learning techniques to analyze real historical house sale transactions in Australia. It seeks useful models which could be deployed as an application for house buyers and sellers. Data analytics show a high discrepancy between the house price in the most expensive suburbs and the most affordable suburbs in the city of Melbourne. In addition, experiments demonstrate that the combination of Stepwise and Support Vector Machine (SVM), based on the Mean Squared Error (MSE) measurement, consistently outperforms other models in terms of prediction accuracy.Keywords: house price prediction, regression trees, neural network, support vector machine, stepwise
Procedia PDF Downloads 2317689 Households’ Willingness to Pay for Watershed Management Practices in Lake Hawassa Watershed, Southern Ethiopia
Authors: Mulugeta Fola, Mengistu Ketema, Kumilachew Alamerie
Abstract:
Watershed provides vast economic benefits within and beyond the management area of interest. But most watersheds in Ethiopia are increasingly facing the threats of degradation due to both natural and man-made causes. To reverse these problems, communities’ participation in sustainable management programs is among the necessary measures. Hence, this study assessed the households’ willingness to pay for watershed management practices through a contingent valuation study approach. Double bounded dichotomous choice with open-ended follow-up format was used to elicit the households’ willingness to pay. Based on data collected from 275 randomly selected households, descriptive statistics results indicated that most households (79.64%) were willing to pay for watershed management practices. A bivariate Probit model was employed to identify determinants of households’ willingness to pay and estimate mean willingness to pay. Its result shows that age, gender, income, livestock size, perception of watershed degradation, social position, and offered bids were important variables affecting willingness to pay for watershed management practices. The study also revealed that the mean willingness to pay for watershed management practices was calculated to be 58.41 Birr and 47.27 Birr per year from the double bounded and open-ended format, respectively. The study revealed that the aggregate welfare gains from watershed management practices were calculated to be 931581.09 Birr and 753909.23 Birr per year from double bounded dichotomous choice and open-ended format, respectively. Therefore, the policymakers should make households to pay for the services of watershed management practices in the study area.Keywords: bivariate probit model, contingent valuation, watershed management practices, willingness to pay
Procedia PDF Downloads 2247688 The Compositional Effects on Electrospinning of Gelatin and Polyvinyl-alcohol Mixed Nanofibers
Authors: Yi-Chun Wu, Nai-Yun Chang, Chuan LI
Abstract:
This study investigates a feasible range of composition for the mixture of gelatin and polyvinyl alcohol to form nanofibers by electrospinning. Gelatin, one of the most available naturally derived hydrogels of amino acids, is a popular choice for food additives, cosmetic ingredients, biomedical implants, or dressing of its non-toxic and biodegradable nature. Nevertheless, synthetic hydrogel polyvinyl alcohol has long been used as a thickening agent for adhesion purposes. Many biomedical devices are also containing polyvinyl-alcohol as a major content, such as eye drops and contact lenses. To discover appropriate compositions of gelatin and polyvinyl-alcohol for electrospun nanofibers, polymer solutions of different volumetric ratios between gelatin and polyvinyl alcohol were prepared for electrospinning. The viscosity, surface tension, pH value, and electrical conductance of polymer solutions were measured. On the nanofibers, the vibrational modes of molecular structures in nanofibers were investigated by Fourier-transform infrared spectroscopy. The morphologies and surface chemical elements of fibers were examined by the scanning electron microscope and the energy-dispersive X-ray spectroscopy. The hydrophilicity of nanofiberswas evaluated by the water contact angles on the surface of the fibers. To further test the biotoxicity of nanofibers, an in-vitro 3T3 fibroblasts culture further tested the biotoxicity of the electrospun nanofibers. Throughstatistical analyses of the experimental data, it is found that the polyvinyl-alcohol rich composition (the volumetric ratio of gelatin/polyvinyl-alcohol < 1) would be a preferable choice for the formation of nanofibers by the current setup of electrospinning. These electrospun nanofibers tend to be hydrophilic with no biotoxicity threat to the 3T3 fibroblasts.Keywords: gelatin, polyvinyl-alcohol, nanofibers, electrospinning, spin coating
Procedia PDF Downloads 857687 Time Series Forecasting (TSF) Using Various Deep Learning Models
Authors: Jimeng Shi, Mahek Jain, Giri Narasimhan
Abstract:
Time Series Forecasting (TSF) is used to predict the target variables at a future time point based on the learning from previous time points. To keep the problem tractable, learning methods use data from a fixed-length window in the past as an explicit input. In this paper, we study how the performance of predictive models changes as a function of different look-back window sizes and different amounts of time to predict the future. We also consider the performance of the recent attention-based Transformer models, which have had good success in the image processing and natural language processing domains. In all, we compare four different deep learning methods (RNN, LSTM, GRU, and Transformer) along with a baseline method. The dataset (hourly) we used is the Beijing Air Quality Dataset from the UCI website, which includes a multivariate time series of many factors measured on an hourly basis for a period of 5 years (2010-14). For each model, we also report on the relationship between the performance and the look-back window sizes and the number of predicted time points into the future. Our experiments suggest that Transformer models have the best performance with the lowest Mean Average Errors (MAE = 14.599, 23.273) and Root Mean Square Errors (RSME = 23.573, 38.131) for most of our single-step and multi-steps predictions. The best size for the look-back window to predict 1 hour into the future appears to be one day, while 2 or 4 days perform the best to predict 3 hours into the future.Keywords: air quality prediction, deep learning algorithms, time series forecasting, look-back window
Procedia PDF Downloads 1547686 Generalized Hyperbolic Functions: Exponential-Type Quantum Interactions
Authors: Jose Juan Peña, J. Morales, J. García-Ravelo
Abstract:
In the search of potential models applied in the theoretical treatment of diatomic molecules, some of them have been constructed by using standard hyperbolic functions as well as from the so-called q-deformed hyperbolic functions (sc q-dhf) for displacing and modifying the shape of the potential under study. In order to transcend the scope of hyperbolic functions, in this work, a kind of generalized q-deformed hyperbolic functions (g q-dhf) is presented. By a suitable transformation, through the q deformation parameter, it is shown that these g q-dhf can be expressed in terms of their corresponding standard ones besides they can be reduced to the sc q-dhf. As a useful application of the proposed approach, and considering a class of exactly solvable multi-parameter exponential-type potentials, some new q-deformed quantum interactions models that can be used as interesting alternative in quantum physics and quantum states are presented. Furthermore, due that quantum potential models are conditioned on the q-dependence of the parameters that characterize to the exponential-type potentials, it is shown that many specific cases of q-deformed potentials are obtained as particular cases from the proposal.Keywords: diatomic molecules, exponential-type potentials, hyperbolic functions, q-deformed potentials
Procedia PDF Downloads 185