Search results for: Random Field Ising Model
24305 Finite Element Analysis of Piezolaminated Structures with Both Geometric and Electroelastic Material Nonlinearities
Authors: Shun-Qi Zhang, Shu-Yang Zhang, Min Chen, , Jing Bai
Abstract:
Piezoelectric laminated smart structures can be subjected to the strong driving electric field, which may result in large displacements and rotations. In one hand, piezoelectric materials usually behave very significant material nonlinear effects under strong electric fields. On the other hand, thin-walled structures undergoing large displacements and rotations exist nonnegligible geometric nonlinearity. In order to give a precise prediction of piezo laminated smart structures under the large electric field, this paper develops a finite element (FE) model accounting for material nonlinearity (piezoelectric part) and geometric nonlinearity based on the first order shear deformation (FSOD) hypothesis. The proposed FE model is first validated by both experimental and numerical examples from the literature. Afterwards, it is applied to simulate for plate and shell structures with multiple piezoelectric patches under the strong applied electric field. From the simulation results, it shows that large discrepancies occur between linear and nonlinear predictions for piezoelectric laminated structures driving at the strong electric field. Therefore, both material and geometric nonlinearities should be taken into account for piezoelectric structures under strong electric.Keywords: piezoelectric smart structures, finite element analysis, geometric nonlinearity, electroelastic material nonlinearities
Procedia PDF Downloads 31724304 Heart Attack Prediction Using Several Machine Learning Methods
Authors: Suzan Anwar, Utkarsh Goyal
Abstract:
Heart rate (HR) is a predictor of cardiovascular, cerebrovascular, and all-cause mortality in the general population, as well as in patients with cardio and cerebrovascular diseases. Machine learning (ML) significantly improves the accuracy of cardiovascular risk prediction, increasing the number of patients identified who could benefit from preventive treatment while avoiding unnecessary treatment of others. This research examines relationship between the individual's various heart health inputs like age, sex, cp, trestbps, thalach, oldpeaketc, and the likelihood of developing heart disease. Machine learning techniques like logistic regression and decision tree, and Python are used. The results of testing and evaluating the model using the Heart Failure Prediction Dataset show the chance of a person having a heart disease with variable accuracy. Logistic regression has yielded an accuracy of 80.48% without data handling. With data handling (normalization, standardscaler), the logistic regression resulted in improved accuracy of 87.80%, decision tree 100%, random forest 100%, and SVM 100%.Keywords: heart rate, machine learning, SVM, decision tree, logistic regression, random forest
Procedia PDF Downloads 13824303 River Habitat Modeling for the Entire Macroinvertebrate Community
Authors: Pinna Beatrice., Laini Alex, Negro Giovanni, Burgazzi Gemma, Viaroli Pierluigi, Vezza Paolo
Abstract:
Habitat models rarely consider macroinvertebrates as ecological targets in rivers. Available approaches mainly focus on single macroinvertebrate species, not addressing the ecological needs and functionality of the entire community. This research aimed to provide an approach to model the habitat of the macroinvertebrate community. The approach is based on the recently developed Flow-T index, together with a Random Forest (RF) regression, which is employed to apply the Flow-T index at the meso-habitat scale. Using different datasets gathered from both field data collection and 2D hydrodynamic simulations, the model has been calibrated in the Trebbia river (2019 campaign), and then validated in the Trebbia, Taro, and Enza rivers (2020 campaign). The three rivers are characterized by a braiding morphology, gravel riverbeds, and summer low flows. The RF model selected 12 mesohabitat descriptors as important for the macroinvertebrate community. These descriptors belong to different frequency classes of water depth, flow velocity, substrate grain size, and connectivity to the main river channel. The cross-validation R² coefficient (R²𝒸ᵥ) of the training dataset is 0.71 for the Trebbia River (2019), whereas the R² coefficient for the validation datasets (Trebbia, Taro, and Enza Rivers 2020) is 0.63. The agreement between the simulated results and the experimental data shows sufficient accuracy and reliability. The outcomes of the study reveal that the model can identify the ecological response of the macroinvertebrate community to possible flow regime alterations and to possible river morphological modifications. Lastly, the proposed approach allows extending the MesoHABSIM methodology, widely used for the fish habitat assessment, to a different ecological target community. Further applications of the approach can be related to flow design in both perennial and non-perennial rivers, including river reaches in which fish fauna is absent.Keywords: ecological flows, macroinvertebrate community, mesohabitat, river habitat modeling
Procedia PDF Downloads 9424302 Survival Data with Incomplete Missing Categorical Covariates
Authors: Madaki Umar Yusuf, Mohd Rizam B. Abubakar
Abstract:
The survival censored data with incomplete covariate data is a common occurrence in many studies in which the outcome is survival time. With model when the missing covariates are categorical, a useful technique for obtaining parameter estimates is the EM by the method of weights. The survival outcome for the class of generalized linear model is applied and this method requires the estimation of the parameters of the distribution of the covariates. In this paper, we propose some clinical trials with ve covariates, four of which have some missing values which clearly show that they were fully censored data.Keywords: EM algorithm, incomplete categorical covariates, ignorable missing data, missing at random (MAR), Weibull Distribution
Procedia PDF Downloads 40624301 Identifying and Evaluating the Effectiveness of Communication Channels between Employees and Management Based on the EFQM Excellence Model
Authors: Mehrdad Hosseinishakib, Mozhgan Chakani, Gholamreza Babaei
Abstract:
This study aims to investigate the relationship between the bilateral communication channels, communication technologies with effective communications and communication technologies, employee participation in motivated decision-making of employees using the EFQM excellence model in Education Organization of Area 4 in Karaj. This research is an applied research in terms of the purpose and is a descriptive survey research in terms of nature and method and assesses the current situation using field studies. The statistical population consists of all employees and managers of Education Organization of Area 4 in Karaj including 5442 persons and random sampling was used and sample size is 359 using Cochran formula. Measurement tool is a researcher-made questionnaire with 20 questions including two categories of expertise and general questions. The first category includes general questions about respondents' personal characteristics such as gender and level of education, work experience and courses of study. The second category includes expertise questions of the questionnaire that have been designed to test research hypotheses that its reliability was approved by Cronbach's alpha coefficient 0.916 and its validity was approved according to the vies of teachers and some senior managers of Education Organization of Area 4 in Karaj. The results of the analysis of the findings show that there is a significant relationship between mutual communication channels, communication technologies with effective communication between employees and management. There is also a significant relationship between communication technologies and employee motivation and employee participation in their motivated decision-making in Education Organization of Area 4 in Karaj.Keywords: communication channels, effective communication, EFQM model, ANOVA
Procedia PDF Downloads 24424300 Transcranial Electric Field Treatments on Redox-Toxic Iron Deposits in Transgenic Alzheimer’s Disease Mouse Models: The Electroceutical Targeting of Alzheimer’s Disease
Authors: Choi Younshick, Lee Wonseok, Lee Jaemeun, Park Sun-Hyun, Kim Sunwoung, Park Sua, Kim Eun Ho, Kim Jong-Ki
Abstract:
Iron accumulation in the brain accelerates Alzheimer’s disease progression. To cure iron toxicity, we assessed the therapeutic effects of noncontact transcranial electric field stimulation to the brain on toxic iron deposits in either the Aβ-fibril structure or the Aβ plaque in a mouse model of Alzheimer’s disease (AD). A capacitive electrode-based alternating electric field (AEF) was applied to a suspension of magnetite (Fe₃O₄) to measure the field-sensitized electro-Fenton effect and resultant reactive oxygen species (ROS) generation. The increase in ROS generation compared to the untreated control was both exposure-time and AEF-frequency dependent. The frequency-specific exposure of AEF to 0.7–1.4 V/cm on a magnetite-bound Aβ-fibril or a transgenic Alzheimer’s disease (AD) mouse model revealed the removal of intraplaque ferrous magnetite iron deposit and Aβ-plaque burden together at the same time compared to the untreated control. The results of the behavioral tests show an improvement in impaired cognitive function following AEF treatment on the AD mouse model. Western blot assay found some disease-modifying biological responses, including down-regulating ferroptosis, neuroinflammation and reactive astrocytes that eventually made cognitive improvement feasible. Tissue clearing and 3D-imaging analysis revealed no induced damage to the neuronal structures of normal brain tissue following AEF treatment. In conclusion, our results suggest that the effective degradation of magnetite-bound amyloid fibrils or plaques in the AD brain by the electro-Fenton effect from electric field-sensitized magnetite offers a potential electroceutical treatment option for AD.Keywords: electroceutical, intraplaque magnetite, alzheimer’s disease, transcranial electric field, electro-fenton effect
Procedia PDF Downloads 7124299 Deterministic and Stochastic Modeling of a Micro-Grid Management for Optimal Power Self-Consumption
Authors: D. Calogine, O. Chau, S. Dotti, O. Ramiarinjanahary, P. Rasoavonjy, F. Tovondahiniriko
Abstract:
Mafate is a natural circus in the north-western part of Reunion Island, without an electrical grid and road network. A micro-grid concept is being experimented in this area, composed of a photovoltaic production combined with electrochemical batteries, in order to meet the local population for self-consumption of electricity demands. This work develops a discrete model as well as a stochastic model in order to reach an optimal equilibrium between production and consumptions for a cluster of houses. The management of the energy power leads to a large linearized programming system, where the time interval of interest is 24 hours The experimental data are solar production, storage energy, and the parameters of the different electrical devices and batteries. The unknown variables to evaluate are the consumptions of the various electrical services, the energy drawn from and stored in the batteries, and the inhabitants’ planning wishes. The objective is to fit the solar production to the electrical consumption of the inhabitants, with an optimal use of the energies in the batteries by satisfying as widely as possible the users' planning requirements. In the discrete model, the different parameters and solutions of the linear programming system are deterministic scalars. Whereas in the stochastic approach, the data parameters and the linear programming solutions become random variables, then the distributions of which could be imposed or established by estimation from samples of real observations or from samples of optimal discrete equilibrium solutions.Keywords: photovoltaic production, power consumption, battery storage resources, random variables, stochastic modeling, estimations of probability distributions, mixed integer linear programming, smart micro-grid, self-consumption of electricity.
Procedia PDF Downloads 11024298 Application of Mathematical Models for Conducting Long-Term Metal Fume Exposure Assessments for Workers in a Shipbuilding Factory
Authors: Shu-Yu Chung, Ying-Fang Wang, Shih-Min Wang
Abstract:
To conduct long-term exposure assessments are important for workers exposed to chemicals with chronic effects. However, it usually encounters with several constrains, including cost, workers' willingness, and interference to work practice, etc., leading to inadequate long-term exposure data in the real world. In this study, an integrated approach was developed for conducting long-term exposure assessment for welding workers in a shipbuilding factory. A laboratory study was conducted to yield the fume generation rates under various operating conditions. The results and the measured environmental conditions were applied to the near field/far field (NF/FF) model for predicting long term fume exposures via the Monte Carlo simulation. Then, the predicted long-term concentrations were used to determine the prior distribution in Bayesian decision analysis (BDA). Finally, the resultant posterior distributions were used to assess the long-term exposure and serve as basis for initiating control strategies for shipbuilding workers. Results show that the NF/FF model was a suitable for predicting the exposures of metal contents containing in welding fume. The resultant posterior distributions could effectively assess the long-term exposures of shipbuilding welders. Welders' long-term Fe, Mn and Pb exposures were found with high possibilities to exceed the action level indicating preventive measures should be taken for reducing welders' exposures immediately. Though the resultant posterior distribution can only be regarded as the best solution based on the currently available predicting and monitoring data, the proposed integrated approach can be regarded as a possible solution for conducting long term exposure assessment in the field.Keywords: Bayesian decision analysis, exposure assessment, near field and far field model, shipbuilding industry, welding fume
Procedia PDF Downloads 14124297 DBN-Based Face Recognition System Using Light Field
Authors: Bing Gu
Abstract:
Abstract—Most of Conventional facial recognition systems are based on image features, such as LBP, SIFT. Recently some DBN-based 2D facial recognition systems have been proposed. However, we find there are few DBN-based 3D facial recognition system and relative researches. 3D facial images include all the individual biometric information. We can use these information to build more accurate features, So we present our DBN-based face recognition system using Light Field. We can see Light Field as another presentation of 3D image, and Light Field Camera show us a way to receive a Light Field. We use the commercially available Light Field Camera to act as the collector of our face recognition system, and the system receive a state-of-art performance as convenient as conventional 2D face recognition system.Keywords: DBN, face recognition, light field, Lytro
Procedia PDF Downloads 46424296 Interaction between Space Syntax and Agent-Based Approaches for Vehicle Volume Modelling
Authors: Chuan Yang, Jing Bie, Panagiotis Psimoulis, Zhong Wang
Abstract:
Modelling and understanding vehicle volume distribution over the urban network are essential for urban design and transport planning. The space syntax approach was widely applied as the main conceptual and methodological framework for contemporary vehicle volume models with the help of the statistical method of multiple regression analysis (MRA). However, the MRA model with space syntax variables shows a limitation in vehicle volume predicting in accounting for the crossed effect of the urban configurational characters and socio-economic factors. The aim of this paper is to construct models by interacting with the combined impact of the street network structure and socio-economic factors. In this paper, we present a multilevel linear (ML) and an agent-based (AB) vehicle volume model at an urban scale interacting with space syntax theoretical framework. The ML model allowed random effects of urban configurational characteristics in different urban contexts. And the AB model was developed with the incorporation of transformed space syntax components of the MRA models into the agents’ spatial behaviour. Three models were implemented in the same urban environment. The ML model exhibit superiority over the original MRA model in identifying the relative impacts of the configurational characters and macro-scale socio-economic factors that shape vehicle movement distribution over the city. Compared with the ML model, the suggested AB model represented the ability to estimate vehicle volume in the urban network considering the combined effects of configurational characters and land-use patterns at the street segment level.Keywords: space syntax, vehicle volume modeling, multilevel model, agent-based model
Procedia PDF Downloads 14524295 Numerical Simulation of Convective Flow of Nanofluids with an Oriented Magnetic Field in a Half Circular-Annulus
Authors: M. J. Uddin, M. M. Rahman
Abstract:
The unsteady convective heat transfer flow of nanofluids in a half circular-annulus shape enclosure using nonhomogeneous dynamic model has been investigated numerically. The round upper wall of the enclosure is maintained at constant low temperature whereas the bottom wall is heated by three different thermal conditions. The enclosure is permeated by a uniform magnetic field having variable orientation. The Brownian motion and thermophoretic phenomena of the nanoparticles are taken into account in model construction. The governing nonlinear momentum, energy, and concentration equations are solved numerically using Galerkin weighted residual finite element method. To discover the best performer, the average Nusselt number is demonstrated for different types of nanofluids. The heat transfer rate for different flow parameters, positions of the annulus, thicknesses of the half circular-annulus and thermal conditions is also exhibited.Keywords: nanofluid, convection, semicircular-annulus, nonhomogeneous dynamic model, finite element method
Procedia PDF Downloads 22124294 Influence of an External Magnetic Field on the Acoustomagnetoelectric Field in a Rectangular Quantum Wire with an Infinite Potential by Using a Quantum Kinetic Equation
Authors: N. Q. Bau, N. V. Nghia
Abstract:
The acoustomagnetoelectric (AME) field in a rectangular quantum wire with an infinite potential (RQWIP) is calculated in the presence of an external magnetic field (EMF) by using the quantum kinetic equation for the distribution function of electrons system interacting with external phonons and electrons scattering with internal acoustic phonon in a RQWIP. We obtained ananalytic expression for the AME field in the RQWIP in the presence of the EMF. The dependence of AME field on the frequency of external acoustic wave, the temperature T of system, the cyclotron frequency of the EMF and the intensity of the EMF is obtained. Theoretical results for the AME field are numerically evaluated, plotted and discussed for a specific RQWIP GaAs/GaAsAl. This result has shown that the dependence of the AME field on intensity of the EMF is nonlinearly and it is many distinct maxima in the quantized magnetic region. We also compared received fields with those for normal bulk semiconductors, quantum well and quantum wire to show the difference. The influence of an EMF on AME field in a RQWIP is newly developed.Keywords: rectangular quantum wire, acoustomagnetoelectric field, electron-phonon interaction, kinetic equation method
Procedia PDF Downloads 33624293 Combustion Analysis of Suspended Sodium Droplet
Authors: T. Watanabe
Abstract:
Combustion analysis of suspended sodium droplet is performed by solving numerically the Navier-Stokes equations and the energy conservation equations. The combustion model consists of the pre-ignition and post-ignition models. The reaction rate for the pre-ignition model is based on the chemical kinetics, while that for the post-ignition model is based on the mass transfer rate of oxygen. The calculated droplet temperature is shown to be in good agreement with the existing experimental data. The temperature field in and around the droplet is obtained as well as the droplet shape variation, and the present numerical model is confirmed to be effective for the combustion analysis.Keywords: analysis, combustion, droplet, sodium
Procedia PDF Downloads 21124292 The Impact of Inpatient New Boarding Policy on Emergency Department Overcrowding: A Discrete Event Simulation Study
Authors: Wheyming Tina Song, Chi-Hao Hong
Abstract:
In this study, we investigate the effect of a new boarding policy - short stay, on the overcrowding efficiency in emergency department (ED). The decision variables are no. of short stay beds for least acuity ED patients. The performance measurements used are national emergency department overcrowding score (NEDOCS) and ED retention rate (the percentage that patients stay in ED over than 48 hours in one month). Discrete event simulation (DES) is used as an analysis tool to evaluate the strategy. Also, common random number (CRN) technique is applied to enhance the simulation precision. The DES model was based on a census of 6 months' patients who were treated in the ED of the National Taiwan University Hospital Yunlin Branch. Our results show that the new short-stay boarding significantly impacts both the NEDOCS and ED retention rate when the no. of short stay beds is more than three.Keywords: emergency department (ED), common random number (CRN), national emergency department overcrowding score (NEDOCS), discrete event simulation (DES)
Procedia PDF Downloads 34824291 Human Intraocular Thermal Field in Action with Different Boundary Conditions Considering Aqueous Humor and Vitreous Humor Fluid Flow
Authors: Dara Singh, Keikhosrow Firouzbakhsh, Mohammad Taghi Ahmadian
Abstract:
In this study, a validated 3D finite volume model of human eye is developed to study the fluid flow and heat transfer in the human eye at steady state conditions. For this purpose, discretized bio-heat transfer equation coupled with Boussinesq equation is analyzed with different anatomical, environmental, and physiological conditions. It is demonstrated that the fluid circulation is formed as a result of thermal gradients in various regions of eye. It is also shown that posterior region of the human eye is less affected by the ambient conditions compared to the anterior segment which is sensitive to the ambient conditions and also to the way the gravitational field is defined compared to the geometry of the eye making the circulations and the thermal field complicated in transient states. The effect of variation in material and boundary conditions guides us to the conclusion that thermal field of a healthy and non-healthy eye can be distinguished via computer simulations.Keywords: bio-heat, boussinesq, conduction, convection, eye
Procedia PDF Downloads 34524290 A Proposed Training Program for the Development of the Kindergarten Teacher According To Her Contemporary Professionàĺ Needs
Authors: Abdulhakim Ali Mosleh Alzubidy
Abstract:
The study's aim was to establish a proposed training program for kindergarten teachers according to their modern professional demands so that they could effectively teach children through movement education and play. The sample, which consisted of (46) teachers and administrators selected at random from the Ibb governorate, represented the study population of kindergarten teachers and administrators. The researcher developed three survey forms as a tool for data collection, and the forms were used with the research sample. The researcher used the descriptive method due to its applicability and the nature of the study, and he also used the appropriate statistical treatment of the data, which is to extract the percentage and the percentage of agreement. The study came to the following conclusions: ● The proposed program is of great importance in preparing the kindergarten teacher in an appropriate manner that keeps pace with modern developments in this field. ● The field of movement education is a necessity for the kindergarten teacher, through which she will be able to prepare the child physically and kinetically and teach him effectively the principles of reading, writing, and numerical and arithmetic concepts.Keywords: training program, professional needs, kindergarten, kindergarten teacher
Procedia PDF Downloads 8524289 Use of SUDOKU Design to Assess the Implications of the Block Size and Testing Order on Efficiency and Precision of Dulce De Leche Preference Estimation
Authors: Jéssica Ferreira Rodrigues, Júlio Silvio De Sousa Bueno Filho, Vanessa Rios De Souza, Ana Carla Marques Pinheiro
Abstract:
This study aimed to evaluate the implications of the block size and testing order on efficiency and precision of preference estimation for Dulce de leche samples. Efficiency was defined as the inverse of the average variance of pairwise comparisons among treatments. Precision was defined as the inverse of the variance of treatment means (or effects) estimates. The experiment was originally designed to test 16 treatments as a series of 8 Sudoku 16x16 designs being 4 randomized independently and 4 others in the reverse order, to yield balance in testing order. Linear mixed models were assigned to the whole experiment with 112 testers and all their grades, as well as their partially balanced subgroups, namely: a) experiment with the four initial EU; b) experiment with EU 5 to 8; c) experiment with EU 9 to 12; and b) experiment with EU 13 to 16. To record responses we used a nine-point hedonic scale, it was assumed a mixed linear model analysis with random tester and treatments effects and with fixed test order effect. Analysis of a cumulative random effects probit link model was very similar, with essentially no different conclusions and for simplicity, we present the results using Gaussian assumption. R-CRAN library lme4 and its function lmer (Fit Linear Mixed-Effects Models) was used for the mixed models and libraries Bayesthresh (default Gaussian threshold function) and ordinal with the function clmm (Cumulative Link Mixed Model) was used to check Bayesian analysis of threshold models and cumulative link probit models. It was noted that the number of samples tested in the same session can influence the acceptance level, underestimating the acceptance. However, proving a large number of samples can help to improve the samples discrimination.Keywords: acceptance, block size, mixed linear model, testing order, testing order
Procedia PDF Downloads 32124288 Supervised-Component-Based Generalised Linear Regression with Multiple Explanatory Blocks: THEME-SCGLR
Authors: Bry X., Trottier C., Mortier F., Cornu G., Verron T.
Abstract:
We address component-based regularization of a Multivariate Generalized Linear Model (MGLM). A set of random responses Y is assumed to depend, through a GLM, on a set X of explanatory variables, as well as on a set T of additional covariates. X is partitioned into R conceptually homogeneous blocks X1, ... , XR , viewed as explanatory themes. Variables in each Xr are assumed many and redundant. Thus, Generalised Linear Regression (GLR) demands regularization with respect to each Xr. By contrast, variables in T are assumed selected so as to demand no regularization. Regularization is performed searching each Xr for an appropriate number of orthogonal components that both contribute to model Y and capture relevant structural information in Xr. We propose a very general criterion to measure structural relevance (SR) of a component in a block, and show how to take SR into account within a Fisher-scoring-type algorithm in order to estimate the model. We show how to deal with mixed-type explanatory variables. The method, named THEME-SCGLR, is tested on simulated data.Keywords: Component-Model, Fisher Scoring Algorithm, GLM, PLS Regression, SCGLR, SEER, THEME
Procedia PDF Downloads 39624287 Multilevel Regression Model - Evaluate Relationship Between Early Years’ Activities of Daily Living and Alzheimer’s Disease Onset Accounting for Influence of Key Sociodemographic Factors Using a Longitudinal Household Survey Data
Authors: Linyi Fan, C.J. Schumaker
Abstract:
Background: Biomedical efforts to treat Alzheimer’s disease (AD) have typically produced mixed to poor results, while more lifestyle-focused treatments such as exercise may fare better than existing biomedical treatments. A few promising studies have indicated that activities of daily life (ADL) may be a useful way of predicting AD. However, the existing cross-sectional studies fail to show how functional-related issues such as ADL in early years predict AD and how social factors influence health either in addition to or in interaction with individual risk factors. This study would helpbetterscreening and early treatments for the elderly population and healthcare practice. The findings have significance academically and practically in terms of creating positive social change. Methodology: The purpose of this quantitative historical, correlational study was to examine the relationship between early years’ ADL and the development of AD in later years. The studyincluded 4,526participantsderived fromRAND HRS dataset. The Health and Retirement Study (HRS) is a longitudinal household survey data set that is available forresearchof retirement and health among the elderly in the United States. The sample was selected by the completion of survey questionnaire about AD and dementia. The variablethat indicates whether the participant has been diagnosed with AD was the dependent variable. The ADL indices and changes in ADL were the independent variables. A four-step multilevel regression model approach was utilized to address the research questions. Results: Amongst 4,526 patients who completed the AD and dementia questionnaire, 144 (3.1%) were diagnosed with AD. Of the 4,526 participants, 3,465 (76.6%) have high school and upper education degrees,4,074 (90.0%) were above poverty threshold. The model evaluatedthe effect of ADL and change in ADL on onset of AD in late years while allowing the intercept of the model to vary by level of education. The results suggested that the only significant predictor of the onset of AD was changes in early years’ ADL (b = 20.253, z = 2.761, p < .05). However, the result of the sensitivity analysis (b = 7.562, z = 1.900, p =.058), which included more control variables and increased the observation period of ADL, are not supported this finding. The model also estimated whether the variances of random effect vary by Level-2 variables. The results suggested that the variances associated with random slopes were approximately zero, suggesting that the relationship between early years’ ADL were not influenced bysociodemographic factors. Conclusion: The finding indicated that an increase in changes in ADL leads to an increase in the probability of onset AD in the future. However, this finding is not support in a broad observation period model. The study also failed to reject the hypothesis that the sociodemographic factors explained significant amounts of variance in random effect. Recommendations were then made for future research and practice based on these limitations and the significance of the findings.Keywords: alzheimer’s disease, epidemiology, moderation, multilevel modeling
Procedia PDF Downloads 13524286 Bayesian Meta-Analysis to Account for Heterogeneity in Studies Relating Life Events to Disease
Authors: Elizabeth Stojanovski
Abstract:
Associations between life events and various forms of cancers have been identified. The purpose of a recent random-effects meta-analysis was to identify studies that examined the association between adverse events associated with changes to financial status including decreased income and breast cancer risk. The same association was studied in four separate studies which displayed traits that were not consistent between studies such as the study design, location and time frame. It was of interest to pool information from various studies to help identify characteristics that differentiated study results. Two random-effects Bayesian meta-analysis models are proposed to combine the reported estimates of the described studies. The proposed models allow major sources of variation to be taken into account, including study level characteristics, between study variance, and within study variance and illustrate the ease with which uncertainty can be incorporated using a hierarchical Bayesian modelling approach.Keywords: random-effects, meta-analysis, Bayesian, variation
Procedia PDF Downloads 16024285 Verification and Application of Finite Element Model Developed for Flood Routing in Rivers
Authors: A. L. Qureshi, A. A. Mahessar, A. Baloch
Abstract:
Flood wave propagation in river channel flow can be enunciated by nonlinear equations of motion for unsteady flow. However, it is difficult to find analytical solution of these complex non-linear equations. Hence, verification of the numerical model should be carried out against field data and numerical predictions. This paper presents the verification of developed finite element model applying for unsteady flow in the open channels. The results of a proposed model indicate a good matching with both Preissmann scheme and HEC-RAS model for a river reach of 29 km at both sites (15 km from upstream and at downstream end) for discharge hydrographs. It also has an agreeable comparison with the Preissemann scheme for the flow depth (stage) hydrographs. The proposed model has also been applying to forecast daily discharges at 400 km downstream from Sukkur barrage, which demonstrates accurate model predictions with observed daily discharges. Hence, this model may be utilized for predicting and issuing flood warnings about flood hazardous in advance.Keywords: finite element method, Preissmann scheme, HEC-RAS, flood forecasting, Indus river
Procedia PDF Downloads 50424284 Next Generation Radiation Risk Assessment and Prediction Tools Generation Applying AI-Machine (Deep) Learning Algorithms
Authors: Selim M. Khan
Abstract:
Indoor air quality is strongly influenced by the presence of radioactive radon (222Rn) gas. Indeed, exposure to high 222Rn concentrations is unequivocally linked to DNA damage and lung cancer and is a worsening issue in North American and European built environments, having increased over time within newer housing stocks as a function of as yet unclear variables. Indoor air radon concentration can be influenced by a wide range of environmental, structural, and behavioral factors. As some of these factors are quantitative while others are qualitative, no single statistical model can determine indoor radon level precisely while simultaneously considering all these variables across a complex and highly diverse dataset. The ability of AI- machine (deep) learning to simultaneously analyze multiple quantitative and qualitative features makes it suitable to predict radon with a high degree of precision. Using Canadian and Swedish long-term indoor air radon exposure data, we are using artificial deep neural network models with random weights and polynomial statistical models in MATLAB to assess and predict radon health risk to human as a function of geospatial, human behavioral, and built environmental metrics. Our initial artificial neural network with random weights model run by sigmoid activation tested different combinations of variables and showed the highest prediction accuracy (>96%) within the reasonable iterations. Here, we present details of these emerging methods and discuss strengths and weaknesses compared to the traditional artificial neural network and statistical methods commonly used to predict indoor air quality in different countries. We propose an artificial deep neural network with random weights as a highly effective method for assessing and predicting indoor radon.Keywords: radon, radiation protection, lung cancer, aI-machine deep learnng, risk assessment, risk prediction, Europe, North America
Procedia PDF Downloads 9624283 A Study on Big Data Analytics, Applications and Challenges
Authors: Chhavi Rana
Abstract:
The aim of the paper is to highlight the existing development in the field of big data analytics. Applications like bioinformatics, smart infrastructure projects, Healthcare, and business intelligence contain voluminous and incremental data, which is hard to organise and analyse and can be dealt with using the framework and model in this field of study. An organization's decision-making strategy can be enhanced using big data analytics and applying different machine learning techniques and statistical tools on such complex data sets that will consequently make better things for society. This paper reviews the current state of the art in this field of study as well as different application domains of big data analytics. It also elaborates on various frameworks in the process of Analysis using different machine-learning techniques. Finally, the paper concludes by stating different challenges and issues raised in existing research.Keywords: big data, big data analytics, machine learning, review
Procedia PDF Downloads 8324282 A Study on Big Data Analytics, Applications, and Challenges
Authors: Chhavi Rana
Abstract:
The aim of the paper is to highlight the existing development in the field of big data analytics. Applications like bioinformatics, smart infrastructure projects, healthcare, and business intelligence contain voluminous and incremental data which is hard to organise and analyse and can be dealt with using the framework and model in this field of study. An organisation decision-making strategy can be enhanced by using big data analytics and applying different machine learning techniques and statistical tools to such complex data sets that will consequently make better things for society. This paper reviews the current state of the art in this field of study as well as different application domains of big data analytics. It also elaborates various frameworks in the process of analysis using different machine learning techniques. Finally, the paper concludes by stating different challenges and issues raised in existing research.Keywords: big data, big data analytics, machine learning, review
Procedia PDF Downloads 9524281 Quantitative Analysis of the Trade Potential of the United States with Members of the European Union: A Gravity Model Approach
Authors: Zahid Ahmad, Nauman Ali
Abstract:
This study has estimated the trade between USA and individual members of European Union using Gravity Model of Trade as The USA has a complex trade relationship with the European countries consist of a large number of consumers, which make USA dependent on EU for major of its total world trade. However, among the member of EU, the trade potential of USA with individual members of EU is not known. Panel data techniques e.g. Random Effect, Fixed Effect and Pooled Panel have been applied to secondary quantitative data to analyze the Trade between USA and EU. Trade Potential of USA with individual members of EU has been obtained using the ratio of Actual trade of USA with EU members and the trade as predicted by Gravity Model. The Study concluded that the USA has greater trade potential with 16 members of EU, including Croatia, Portugal and United Kingdom on top. On the other hand, Finland, Ireland, and France are the top countries with which the USA has exhaustive trade potential.Keywords: analytical technique, economic, gravity, international trade, significant
Procedia PDF Downloads 30624280 Molecular Junctions between Graphene Strips: Electronic and Transport Properties
Authors: Adel Belayadi, Ahmed Mougari, Boualem Bourahla
Abstract:
Molecular junctions are currently considered a promising style in the miniaturization of electronic devices. In this contribution, we provide a tight-binding model to investigate the quantum transport properties across-molecular junctions sandwiched between 2D-graphene nanoribbons in the zigzag direction. We investigate, in particular, the effect of embedded atoms such as Gold and Silicon across the molecular junction. The results exhibit a resonance behavior in terms of incident Fermi levels, depending on the molecular junction type. Additionally, the transport properties under a perpendicular magnetic field exhibit an oscillation for the transmittance versus the magnetic field strength.Keywords: molecular junction, 2D-graphene nanoribbons, quantum transport properties, magnetic field
Procedia PDF Downloads 9624279 Assessment of Carbon Dioxide Separation by Amine Solutions Using Electrolyte Non-Random Two-Liquid and Peng-Robinson Models: Carbon Dioxide Absorption Efficiency
Authors: Arash Esmaeili, Zhibang Liu, Yang Xiang, Jimmy Yun, Lei Shao
Abstract:
A high pressure carbon dioxide (CO2) absorption from a specific gas in a conventional column has been evaluated by the Aspen HYSYS simulator using a wide range of single absorbents and blended solutions to estimate the outlet CO2 concentration, absorption efficiency and CO2 loading to choose the most proper solution in terms of CO2 capture for environmental concerns. The property package (Acid Gas-Chemical Solvent) which is compatible with all applied solutions for the simulation in this study, estimates the properties based on an electrolyte non-random two-liquid (E-NRTL) model for electrolyte thermodynamics and Peng-Robinson equation of state for the vapor and liquid hydrocarbon phases. Among all the investigated single amines as well as blended solutions, piperazine (PZ) and the mixture of piperazine and monoethanolamine (MEA) have been found as the most effective absorbents respectively for CO2 absorption with high reactivity based on the simulated operational conditions.Keywords: absorption, amine solutions, Aspen HYSYS, carbon dioxide, simulation
Procedia PDF Downloads 18624278 International Service Learning 3.0: Using Technology to Improve Outcomes and Sustainability
Authors: Anthony Vandarakis
Abstract:
Today’s International Service Learning practices require an update: modern technologies, fresh educational frameworks, and a new operating system to accountably prosper. This paper describes a model of International Service Learning (ISL), which combines current technological hardware, electronic platforms, and asynchronous communications that are grounded in inclusive pedagogy. This model builds on the work around collaborative field trip learning, extending the reach to international partnerships across continents. Mobile technology, 21st century skills and summit-basecamp modeling intersect to support novel forms of learning that tread lightly on fragile natural ecosystems, affirm local reciprocal partnership in projects, and protect traveling participants from common yet avoidable cultural pitfalls.Keywords: International Service Learning, ISL, field experiences, mobile technology, out there in here, summit basecamp pedagogy
Procedia PDF Downloads 17224277 Predicting Costs in Construction Projects with Machine Learning: A Detailed Study Based on Activity-Level Data
Authors: Soheila Sadeghi
Abstract:
Construction projects are complex and often subject to significant cost overruns due to the multifaceted nature of the activities involved. Accurate cost estimation is crucial for effective budget planning and resource allocation. Traditional methods for predicting overruns often rely on expert judgment or analysis of historical data, which can be time-consuming, subjective, and may fail to consider important factors. However, with the increasing availability of data from construction projects, machine learning techniques can be leveraged to improve the accuracy of overrun predictions. This study applied machine learning algorithms to enhance the prediction of cost overruns in a case study of a construction project. The methodology involved the development and evaluation of two machine learning models: Random Forest and Neural Networks. Random Forest can handle high-dimensional data, capture complex relationships, and provide feature importance estimates. Neural Networks, particularly Deep Neural Networks (DNNs), are capable of automatically learning and modeling complex, non-linear relationships between input features and the target variable. These models can adapt to new data, reduce human bias, and uncover hidden patterns in the dataset. The findings of this study demonstrate that both Random Forest and Neural Networks can significantly improve the accuracy of cost overrun predictions compared to traditional methods. The Random Forest model also identified key cost drivers and risk factors, such as changes in the scope of work and delays in material delivery, which can inform better project risk management. However, the study acknowledges several limitations. First, the findings are based on a single construction project, which may limit the generalizability of the results to other projects or contexts. Second, the dataset, although comprehensive, may not capture all relevant factors influencing cost overruns, such as external economic conditions or political factors. Third, the study focuses primarily on cost overruns, while schedule overruns are not explicitly addressed. Future research should explore the application of machine learning techniques to a broader range of projects, incorporate additional data sources, and investigate the prediction of both cost and schedule overruns simultaneously.Keywords: cost prediction, machine learning, project management, random forest, neural networks
Procedia PDF Downloads 5624276 Establishment of a Classifier Model for Early Prediction of Acute Delirium in Adult Intensive Care Unit Using Machine Learning
Authors: Pei Yi Lin
Abstract:
Objective: The objective of this study is to use machine learning methods to build an early prediction classifier model for acute delirium to improve the quality of medical care for intensive care patients. Background: Delirium is a common acute and sudden disturbance of consciousness in critically ill patients. After the occurrence, it is easy to prolong the length of hospital stay and increase medical costs and mortality. In 2021, the incidence of delirium in the intensive care unit of internal medicine was as high as 59.78%, which indirectly prolonged the average length of hospital stay by 8.28 days, and the mortality rate is about 2.22% in the past three years. Therefore, it is expected to build a delirium prediction classifier through big data analysis and machine learning methods to detect delirium early. Method: This study is a retrospective study, using the artificial intelligence big data database to extract the characteristic factors related to delirium in intensive care unit patients and let the machine learn. The study included patients aged over 20 years old who were admitted to the intensive care unit between May 1, 2022, and December 31, 2022, excluding GCS assessment <4 points, admission to ICU for less than 24 hours, and CAM-ICU evaluation. The CAMICU delirium assessment results every 8 hours within 30 days of hospitalization are regarded as an event, and the cumulative data from ICU admission to the prediction time point are extracted to predict the possibility of delirium occurring in the next 8 hours, and collect a total of 63,754 research case data, extract 12 feature selections to train the model, including age, sex, average ICU stay hours, visual and auditory abnormalities, RASS assessment score, APACHE-II Score score, number of invasive catheters indwelling, restraint and sedative and hypnotic drugs. Through feature data cleaning, processing and KNN interpolation method supplementation, a total of 54595 research case events were extracted to provide machine learning model analysis, using the research events from May 01 to November 30, 2022, as the model training data, 80% of which is the training set for model training, and 20% for the internal verification of the verification set, and then from December 01 to December 2022 The CU research event on the 31st is an external verification set data, and finally the model inference and performance evaluation are performed, and then the model has trained again by adjusting the model parameters. Results: In this study, XG Boost, Random Forest, Logistic Regression, and Decision Tree were used to analyze and compare four machine learning models. The average accuracy rate of internal verification was highest in Random Forest (AUC=0.86), and the average accuracy rate of external verification was in Random Forest and XG Boost was the highest, AUC was 0.86, and the average accuracy of cross-validation was the highest in Random Forest (ACC=0.77). Conclusion: Clinically, medical staff usually conduct CAM-ICU assessments at the bedside of critically ill patients in clinical practice, but there is a lack of machine learning classification methods to assist ICU patients in real-time assessment, resulting in the inability to provide more objective and continuous monitoring data to assist Clinical staff can more accurately identify and predict the occurrence of delirium in patients. It is hoped that the development and construction of predictive models through machine learning can predict delirium early and immediately, make clinical decisions at the best time, and cooperate with PADIS delirium care measures to provide individualized non-drug interventional care measures to maintain patient safety, and then Improve the quality of care.Keywords: critically ill patients, machine learning methods, delirium prediction, classifier model
Procedia PDF Downloads 76