Search results for: parameter inference
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2327

Search results for: parameter inference

1697 Improvement of the Q-System Using the Rock Engineering System: A Case Study of Water Conveyor Tunnel of Azad Dam

Authors: Sahand Golmohammadi, Sana Hosseini Shirazi

Abstract:

Because the status and mechanical parameters of discontinuities in the rock mass are included in the calculations, various methods of rock engineering classification are often used as a starting point for the design of different types of structures. The Q-system is one of the most frequently used methods for stability analysis and determination of support systems of underground structures in rock, including tunnel. In this method, six main parameters of the rock mass, namely, the rock quality designation (RQD), joint set number (Jn), joint roughness number (Jr), joint alteration number (Ja), joint water parameter (Jw) and stress reduction factor (SRF) are required. In this regard, in order to achieve a reasonable and optimal design, identifying the effective parameters for the stability of the mentioned structures is one of the most important goals and the most necessary actions in rock engineering. Therefore, it is necessary to study the relationships between the parameters of a system and how they interact with each other and, ultimately, the whole system. In this research, it has attempted to determine the most effective parameters (key parameters) from the six parameters of rock mass in the Q-system using the rock engineering system (RES) method to improve the relationships between the parameters in the calculation of the Q value. The RES system is, in fact, a method by which one can determine the degree of cause and effect of a system's parameters by making an interaction matrix. In this research, the geomechanical data collected from the water conveyor tunnel of Azad Dam were used to make the interaction matrix of the Q-system. For this purpose, instead of using the conventional methods that are always accompanied by defects such as uncertainty, the Q-system interaction matrix is coded using a technique that is actually a statistical analysis of the data and determining the correlation coefficient between them. So, the effect of each parameter on the system is evaluated with greater certainty. The results of this study show that the formed interaction matrix provides a reasonable estimate of the effective parameters in the Q-system. Among the six parameters of the Q-system, the SRF and Jr parameters have the maximum and minimum impact on the system, respectively, and also the RQD and Jw parameters have the maximum and minimum impact on the system, respectively. Therefore, by developing this method, we can obtain a more accurate relation to the rock mass classification by weighting the required parameters in the Q-system.

Keywords: Q-system, rock engineering system, statistical analysis, rock mass, tunnel

Procedia PDF Downloads 70
1696 Evaluation of Weather Risk Insurance for Agricultural Products Using a 3-Factor Pricing Model

Authors: O. Benabdeljelil, A. Karioun, S. Amami, R. Rouger, M. Hamidine

Abstract:

A model for preventing the risks related to climate conditions in the agricultural sector is presented. It will determine the yearly optimum premium to be paid by a producer in order to reach his required turnover. The model is based on both climatic stability and 'soft' responses of usually grown species to average climate variations at the same place and inside a safety ball which can be determined from past meteorological data. This allows the use of linear regression expression for dependence of production result in terms of driving meteorological parameters, the main ones of which are daily average sunlight, rainfall and temperature. By simple best parameter fit from the expert table drawn with professionals, optimal representation of yearly production is determined from records of previous years, and yearly payback is evaluated from minimum yearly produced turnover. The model also requires accurate pricing of commodity at N+1. Therefore, a pricing model is developed using 3 state variables, namely the spot price, the difference between the mean-term and the long-term forward price, and the long-term structure of the model. The use of historical data enables to calibrate the parameters of state variables, and allows the pricing of commodity. Application to beet sugar underlines pricer precision. Indeed, the percentage of accuracy between computed result and real world is 99,5%. Optimal premium is then deduced and gives the producer a useful bound for negotiating an offer by insurance companies to effectively protect its harvest. The application to beet production in French Oise department illustrates the reliability of present model with as low as 6% difference between predicted and real data. The model can be adapted to almost any agricultural field by changing state parameters and calibrating their associated coefficients.

Keywords: agriculture, production model, optimal price, meteorological factors, 3-factor model, parameter calibration, forward price

Procedia PDF Downloads 376
1695 Foamability and Foam Stability of Gelatine-Sodium Dodecyl Sulfate Solutions

Authors: Virginia Martin Torrejon, Song Hang

Abstract:

Gelatine foams are widely explored materials due to their biodegradability, biocompatibility, and availability. They exhibit outstanding properties and are currently subject to increasing scientific research due to their potential use in different applications, such as biocompatible cellular materials for biomedical products or biofoams as an alternative to fossil-fuel-derived packaging. Gelatine is a highly surface-active polymer, and its concentrated solutions usually do not require surfactants to achieve low surface tension. Still, anionic surfactants like sodium dodecyl sulfate (SDS) strongly interact with gelatine, impacting its viscosity and rheological properties and, in turn, their foaming behaviour. Foaming behaviour is a key parameter for cellular solids produced by mechanical foaming as it has a significant effect on the processing and properties of cellular materials. Foamability mainly impacts the density and the mechanical properties of the foams, while foam stability is crucial to achieving foams with low shrinkage and desirable pore morphology. This work aimed to investigate the influence of SDS on the foaming behaviour of concentrated gelatine foams by using a dynamic foam analyser. The study of maximum foam height created, foam formation behaviour, drainage behaviour, and foam structure with regard to bubble size and distribution were carried out in 10 wt% gelatine solutions prepared at different SDS/gelatine concentration ratios. Comparative rheological and viscometry measurements provided a good correlation with the data from the dynamic foam analyser measurements. SDS incorporation at optimum dosages and gelatine gelation led to highly stable foams at high expansion ratios. The viscosity increase of the hydrogel solution at SDS content increased was a key parameter for foam stabilization. In addition, the impact of SDS content on gelling time and gel strength also considerably impacted the foams' stability and pore structure.

Keywords: dynamic foam analyser, gelatine foams stability and foamability, gelatine-surfactant foams, gelatine-SDS rheology, gelatine-SDS viscosity

Procedia PDF Downloads 151
1694 Building a Parametric Link between Mapping and Planning: A Sunlight-Adaptive Urban Green System Plan Formation Process

Authors: Chenhao Zhu

Abstract:

Quantitative mapping is playing a growing role in guiding urban planning, such as using a heat map created by CFX, CFD2000, or Envi-met, to adjust the master plan. However, there is no effective quantitative link between the mappings and planning formation. So, in many cases, the decision-making is still based on the planner's subjective interpretation and understanding of these mappings, which limits the improvement of scientific and accuracy brought by the quantitative mapping. Therefore, in this paper, an effort has been made to give a methodology of building a parametric link between the mapping and planning formation. A parametric planning process based on radiant mapping has been proposed for creating an urban green system. In the first step, a script is written in Grasshopper to build a road network and form the block, while the Ladybug Plug-in is used to conduct a radiant analysis in the form of mapping. Then, the research creatively transforms the radiant mapping from a polygon into a data point matrix, because polygon is hard to engage in the design formation. Next, another script is created to select the main green spaces from the road network based on the criteria of radiant intensity and connect the green spaces' central points to generate a green corridor. After that, a control parameter is introduced to adjust the corridor's form based on the radiant intensity. Finally, a green system containing greenspace and green corridor is generated under the quantitative control of the data matrix. The designer only needs to modify the control parameter according to the relevant research results and actual conditions to realize the optimization of the green system. This method can also be applied to much other mapping-based analysis, such as wind environment analysis, thermal environment analysis, and even environmental sensitivity analysis. The parameterized link between the mapping and planning will bring about a more accurate, objective, and scientific planning.

Keywords: parametric link, mapping, urban green system, radiant intensity, planning strategy, grasshopper

Procedia PDF Downloads 140
1693 A Posterior Predictive Model-Based Control Chart for Monitoring Healthcare

Authors: Yi-Fan Lin, Peter P. Howley, Frank A. Tuyl

Abstract:

Quality measurement and reporting systems are used in healthcare internationally. In Australia, the Australian Council on Healthcare Standards records and reports hundreds of clinical indicators (CIs) nationally across the healthcare system. These CIs are measures of performance in the clinical setting, and are used as a screening tool to help assess whether a standard of care is being met. Existing analysis and reporting of these CIs incorporate Bayesian methods to address sampling variation; however, such assessments are retrospective in nature, reporting upon the previous six or twelve months of data. The use of Bayesian methods within statistical process control for monitoring systems is an important pursuit to support more timely decision-making. Our research has developed and assessed a new graphical monitoring tool, similar to a control chart, based on the beta-binomial posterior predictive (BBPP) distribution to facilitate the real-time assessment of health care organizational performance via CIs. The BBPP charts have been compared with the traditional Bernoulli CUSUM (BC) chart by simulation. The more traditional “central” and “highest posterior density” (HPD) interval approaches were each considered to define the limits, and the multiple charts were compared via in-control and out-of-control average run lengths (ARLs), assuming that the parameter representing the underlying CI rate (proportion of cases with an event of interest) required estimation. Preliminary results have identified that the BBPP chart with HPD-based control limits provides better out-of-control run length performance than the central interval-based and BC charts. Further, the BC chart’s performance may be improved by using Bayesian parameter estimation of the underlying CI rate.

Keywords: average run length (ARL), bernoulli cusum (BC) chart, beta binomial posterior predictive (BBPP) distribution, clinical indicator (CI), healthcare organization (HCO), highest posterior density (HPD) interval

Procedia PDF Downloads 200
1692 Fast-Forward Problem in Asymmetric Double-Well Potential

Authors: Iwan Setiawan, Bobby Eka Gunara, Katshuhiro Nakamura

Abstract:

The theory to accelerate system on quantum dynamics has been constructed to get the desired wave function on shorter time. This theory is developed on adiabatic quantum dynamics which any regulation is done on wave function that satisfies Schrödinger equation. We show accelerated manipulation of WFs with the use of a parameter-dependent in asymmetric double-well potential and also when it’s influenced by electromagnetic fields.

Keywords: driving potential, Adiabatic Quantum Dynamics, regulation, electromagnetic field

Procedia PDF Downloads 337
1691 Box Counting Dimension of the Union L of Trinomial Curves When α ≥ 1

Authors: Kaoutar Lamrini Uahabi, Mohamed Atounti

Abstract:

In the present work, we consider one category of curves denoted by L(p, k, r, n). These curves are continuous arcs which are trajectories of roots of the trinomial equation zn = αzk + (1 − α), where z is a complex number, n and k are two integers such that 1 ≤ k ≤ n − 1 and α is a real parameter greater than 1. Denoting by L the union of all trinomial curves L(p, k, r, n) and using the box counting dimension as fractal dimension, we will prove that the dimension of L is equal to 3/2.

Keywords: feasible angles, fractal dimension, Minkowski sausage, trinomial curves, trinomial equation

Procedia PDF Downloads 188
1690 Elucidation of the Sequential Transcriptional Activity in Escherichia coli Using Time-Series RNA-Seq Data

Authors: Pui Shan Wong, Kosuke Tashiro, Satoru Kuhara, Sachiyo Aburatani

Abstract:

Functional genomics and gene regulation inference has readily expanded our knowledge and understanding of gene interactions with regards to expression regulation. With the advancement of transcriptome sequencing in time-series comes the ability to study the sequential changes of the transcriptome. This method presented here works to augment existing regulation networks accumulated in literature with transcriptome data gathered from time-series experiments to construct a sequential representation of transcription factor activity. This method is applied on a time-series RNA-Seq data set from Escherichia coli as it transitions from growth to stationary phase over five hours. Investigations are conducted on the various metabolic activities in gene regulation processes by taking advantage of the correlation between regulatory gene pairs to examine their activity on a dynamic network. Especially, the changes in metabolic activity during phase transition are analyzed with focus on the pagP gene as well as other associated transcription factors. The visualization of the sequential transcriptional activity is used to describe the change in metabolic pathway activity originating from the pagP transcription factor, phoP. The results show a shift from amino acid and nucleic acid metabolism, to energy metabolism during the transition to stationary phase in E. coli.

Keywords: Escherichia coli, gene regulation, network, time-series

Procedia PDF Downloads 371
1689 Faster, Lighter, More Accurate: A Deep Learning Ensemble for Content Moderation

Authors: Arian Hosseini, Mahmudul Hasan

Abstract:

To address the increasing need for efficient and accurate content moderation, we propose an efficient and lightweight deep classification ensemble structure. Our approach is based on a combination of simple visual features, designed for high-accuracy classification of violent content with low false positives. Our ensemble architecture utilizes a set of lightweight models with narrowed-down color features, and we apply it to both images and videos. We evaluated our approach using a large dataset of explosion and blast contents and compared its performance to popular deep learning models such as ResNet-50. Our evaluation results demonstrate significant improvements in prediction accuracy, while benefiting from 7.64x faster inference and lower computation cost. While our approach is tailored to explosion detection, it can be applied to other similar content moderation and violence detection use cases as well. Based on our experiments, we propose a "think small, think many" philosophy in classification scenarios. We argue that transforming a single, large, monolithic deep model into a verification-based step model ensemble of multiple small, simple, and lightweight models with narrowed-down visual features can possibly lead to predictions with higher accuracy.

Keywords: deep classification, content moderation, ensemble learning, explosion detection, video processing

Procedia PDF Downloads 52
1688 Ill-Posed Inverse Problems in Molecular Imaging

Authors: Ranadhir Roy

Abstract:

Inverse problems arise in medical (molecular) imaging. These problems are characterized by large in three dimensions, and by the diffusion equation which models the physical phenomena within the media. The inverse problems are posed as a nonlinear optimization where the unknown parameters are found by minimizing the difference between the predicted data and the measured data. To obtain a unique and stable solution to an ill-posed inverse problem, a priori information must be used. Mathematical conditions to obtain stable solutions are established in Tikhonov’s regularization method, where the a priori information is introduced via a stabilizing functional, which may be designed to incorporate some relevant information of an inverse problem. Effective determination of the Tikhonov regularization parameter requires knowledge of the true solution, or in the case of optical imaging, the true image. Yet, in, clinically-based imaging, true image is not known. To alleviate these difficulties we have applied the penalty/modified barrier function (PMBF) method instead of Tikhonov regularization technique to make the inverse problems well-posed. Unlike the Tikhonov regularization method, the constrained optimization technique, which is based on simple bounds of the optical parameter properties of the tissue, can easily be implemented in the PMBF method. Imposing the constraints on the optical properties of the tissue explicitly restricts solution sets and can restore uniqueness. Like the Tikhonov regularization method, the PMBF method limits the size of the condition number of the Hessian matrix of the given objective function. The accuracy and the rapid convergence of the PMBF method require a good initial guess of the Lagrange multipliers. To obtain the initial guess of the multipliers, we use a least square unconstrained minimization problem. Three-dimensional images of fluorescence absorption coefficients and lifetimes were reconstructed from contact and noncontact experimentally measured data.

Keywords: constrained minimization, ill-conditioned inverse problems, Tikhonov regularization method, penalty modified barrier function method

Procedia PDF Downloads 269
1687 Merging Appeal to Ignorance, Composition, and Division Argument Schemes with Bayesian Networks

Authors: Kong Ngai Pei

Abstract:

The argument scheme approach to argumentation has two components. One is to identify the recurrent patterns of inferences used in everyday discourse. The second is to devise critical questions to evaluate the inferences in these patterns. Although this approach is intuitive and contains many insightful ideas, it has been noted to be not free of problems. One is that due to its disavowing the probability calculus, it cannot give the exact strength of an inference. In order to tackle this problem, thereby paving the way to a more complete normative account of argument strength, it has been proposed, the most promising way is to combine the scheme-based approach with Bayesian networks (BNs). This paper pursues this line of thought, attempting to combine three common schemes, Appeal to Ignorance, Composition, and Division, with BNs. In the first part, it is argued that most (if not all) formulations of the critical questions corresponding to these schemes in the current argumentation literature are incomplete and not very informative. To remedy these flaws, more thorough and precise formulations of these questions are provided. In the second part, how to use graphical idioms (e.g. measurement and synthesis idioms) to translate the schemes as well as their corresponding critical questions to graphical structure of BNs, and how to define probability tables of the nodes using functions of various sorts are shown. In the final part, it is argued that many misuses of these schemes, traditionally called fallacies with the same names as the schemes, can indeed be adequately accounted for by the BN models proposed in this paper.

Keywords: appeal to ignorance, argument schemes, Bayesian networks, composition, division

Procedia PDF Downloads 285
1686 PLO-AIM: Potential-Based Lane Organization in Autonomous Intersection Management

Authors: Berk Ecer, Ebru Akcapinar Sezer

Abstract:

Traditional management models of intersections, such as no-light intersections or signalized intersection, are not the most effective way of passing the intersections if the vehicles are intelligent. To this end, Dresner and Stone proposed a new intersection control model called Autonomous Intersection Management (AIM). In the AIM simulation, they were examining the problem from a multi-agent perspective, demonstrating that intelligent intersection control can be made more efficient than existing control mechanisms. In this study, autonomous intersection management has been investigated. We extended their works and added a potential-based lane organization layer. In order to distribute vehicles evenly to each lane, this layer triggers vehicles to analyze near lanes, and they change their lane if other lanes have an advantage. We can observe this behavior in real life, such as drivers, change their lane by considering their intuitions. Basic intuition on selecting the correct lane for traffic is selecting a less crowded lane in order to reduce delay. We model that behavior without any change in the AIM workflow. Experiment results show us that intersection performance is directly connected with the vehicle distribution in lanes of roads of intersections. We see the advantage of handling lane management with a potential approach in performance metrics such as average delay of intersection and average travel time. Therefore, lane management and intersection management are problems that need to be handled together. This study shows us that the lane through which vehicles enter the intersection is an effective parameter for intersection management. Our study draws attention to this parameter and suggested a solution for it. We observed that the regulation of AIM inputs, which are vehicles in lanes, was as effective as contributing to aim intersection management. PLO-AIM model outperforms AIM in evaluation metrics such as average delay of intersection and average travel time for reasonable traffic rates, which is in between 600 vehicle/hour per lane to 1300 vehicle/hour per lane. The proposed model reduced the average travel time reduced in between %0.2 - %17.3 and reduced the average delay of intersection in between %1.6 - %17.1 for 4-lane and 6-lane scenarios.

Keywords: AIM project, autonomous intersection management, lane organization, potential-based approach

Procedia PDF Downloads 138
1685 Parametric Studies of Ethylene Dichloride Purification Process

Authors: Sh. Arzani, H. Kazemi Esfeh, Y. Galeh Zadeh, V. Akbari

Abstract:

Ethylene dichloride is a colorless liquid with a smell like chloroform. EDC is classified in the simple hydrocarbon group which is obtained from chlorinating ethylene gas. Its chemical formula is C2H2Cl2 which is used as the main mediator in VCM production. Therefore, the purification process of EDC is important in the petrochemical process. In this study, the purification unit of EDC was simulated, and then validation was performed. Finally, the impact of process parameter was studied for the degree of EDC purity. The results showed that by increasing the feed flow, the reflux impure combinations increase and result in an EDC purity decrease.

Keywords: ethylene dichloride, purification, edc, simulation

Procedia PDF Downloads 315
1684 An Experimental Study on the Variability of Nonnative and Native Inference of Word Meanings in Timed and Untimed Conditions

Authors: Swathi M. Vanniarajan

Abstract:

Reading research suggests that online contextual vocabulary comprehension while reading is an interactive and integrative process. One’s success in it depends on a variety of factors including the amount and the nature of available linguistic and nonlinguistic cues, his/her analytical and integrative skills, schema memory (content familiarity), and processing speed characterized along the continuum of controlled to automatic processing. The experiment reported here, conducted with 30 native speakers as one group and 30 nonnative speakers as another group (all graduate students), hypothesized that while working on (24) tasks which required them to comprehend an unfamiliar word in real time without backtracking, due to the differences in the nature of their respective reading processes, the nonnative subjects would be less able to construct the meanings of the unknown words by integrating the multiple but sufficient contextual cues provided in the text but the native subjects would be able to. The results indicated that there were significant inter-group as well as intra-group differences in terms of the quality of definitions given. However, when given additional time, while the nonnative speakers could significantly improve the quality of their definitions, the native speakers in general would not, suggesting that all things being equal, time is a significant factor for success in nonnative vocabulary and reading comprehension processes and that accuracy precedes automaticity in the development of nonnative reading processes also.

Keywords: reading, second language processing, vocabulary comprehension

Procedia PDF Downloads 165
1683 A Novel Stator Resistance Estimation Method and Control Design of Speed-Sensorless Induction Motor Drives

Authors: N. Ben Si Ali, N. Benalia, N. Zarzouri

Abstract:

Speed sensorless systems are intensively studied during recent years; this is mainly due to their economical benefit and fragility of mechanical sensors and also the difficulty of installing this type of sensor in many applications. These systems suffer from instability problems and sensitivity to parameter mismatch at low speed operation. In this paper an analysis of adaptive observer stability with stator resistance estimation is given.

Keywords: motor drive, sensorless control, adaptive observer, stator resistance estimation

Procedia PDF Downloads 374
1682 Controller Design Using GA for SMC Systems

Authors: Susy Thomas, Sajju Thomas, Varghese Vaidyan

Abstract:

This paper considers SMCs using linear feedback with switched gains and proposes a method which can minimize the pole perturbation. The method is able to enhance the robustness property of the controller. A pre-assigned neighborhood of the ‘nominal’ positions is assigned and the system poles are not allowed to stray out of these bounds even when parameters variations/uncertainties act upon the system. A quasi SMM is maintained within the assigned boundaries of the sliding surface.

Keywords: parameter variations, pole perturbation, sliding mode control, switching surface, robust switching vector

Procedia PDF Downloads 361
1681 Bayesian Analysis of Topp-Leone Generalized Exponential Distribution

Authors: Najrullah Khan, Athar Ali Khan

Abstract:

The Topp-Leone distribution was introduced by Topp- Leone in 1955. In this paper, an attempt has been made to fit Topp-Leone Generalized exponential (TPGE) distribution. A real survival data set is used for illustrations. Implementation is done using R and JAGS and appropriate illustrations are made. R and JAGS codes have been provided to implement censoring mechanism using both optimization and simulation tools. The main aim of this paper is to describe and illustrate the Bayesian modelling approach to the analysis of survival data. Emphasis is placed on the modeling of data and the interpretation of the results. Crucial to this is an understanding of the nature of the incomplete or 'censored' data encountered. Analytic approximation and simulation tools are covered here, but most of the emphasis is on Markov chain based Monte Carlo method including independent Metropolis algorithm, which is currently the most popular technique. For analytic approximation, among various optimization algorithms and trust region method is found to be the best. In this paper, TPGE model is also used to analyze the lifetime data in Bayesian paradigm. Results are evaluated from the above mentioned real survival data set. The analytic approximation and simulation methods are implemented using some software packages. It is clear from our findings that simulation tools provide better results as compared to those obtained by asymptotic approximation.

Keywords: Bayesian Inference, JAGS, Laplace Approximation, LaplacesDemon, posterior, R Software, simulation

Procedia PDF Downloads 534
1680 Ultracapacitor State-of-Energy Monitoring System with On-Line Parameter Identification

Authors: N. Reichbach, A. Kuperman

Abstract:

The paper describes a design of a monitoring system for super capacitor packs in propulsion systems, allowing determining the instantaneous energy capacity under power loading. The system contains real-time recursive-least-squares identification mechanism, estimating the values of pack capacitance and equivalent series resistance. These values are required for accurate calculation of the state-of-energy.

Keywords: real-time monitoring, RLS identification algorithm, state-of-energy, super capacitor

Procedia PDF Downloads 530
1679 Impact of Modifying the Surface Materials on the Radiative Heat Transfer Phenomenon

Authors: Arkadiusz Urzędowski, Dorota Wójcicka-Migasiuk, Andrzej Sachajdak, Magdalena Paśnikowska-Łukaszuk

Abstract:

Due to the impact of climate changes and inevitability to reduce greenhouse gases, the need to use low-carbon and sustainable construction has increased. In this work, it is investigated how texture of the surface building materials and radiative heat transfer phenomenon in flat multilayer can be correlated. Attempts to test the surface emissivity are taken however, the trustworthiness of measurement results remains a concern since sensor size and thickness are common problems. This paper presents an experimental method to studies surface emissivity with use self constructed thermal sensors and thermal imaging technique. The surface of building materials was modified by mechanical and chemical treatment affecting the reduction of the emissivity. For testing the shaping surface of materials and mapping its three-dimensional structure, scanning profilometry were used in a laboratory. By comparing the results of laboratory tests and performed analysis of 3D computer fluid dynamics software, it can be shown that a change in the surface coverage of materials affects the heat transport by radiation between layers. Motivated by recent advancements in variational inference, this publication evaluates the potential use a dedicated data processing approach, and properly constructed temperature sensors, the influence of the surface emissivity on the phenomenon of radiation and heat transport in the entire partition can be determined.

Keywords: heat transfer, surface roughness, surface emissivity, radiation

Procedia PDF Downloads 94
1678 Numerical Simulation of Laser ‎Propagation through Turbulent ‎Atmosphere Using Zernike ‎Polynomials

Authors: Mohammad Moradi ‎

Abstract:

In this article, propagation of a laser beam through turbulent ‎atmosphere is evaluated. At first the laser beam is simulated and then ‎turbulent atmosphere will be simulated by using Zernike polynomials. ‎Some parameter like intensity, PSF will be measured for four ‎wavelengths in different Cn2.

Keywords: laser beam propagation, phase screen, turbulent atmosphere, Zernike ‎polynomials

Procedia PDF Downloads 510
1677 Joint Modeling of Longitudinal and Time-To-Event Data with Latent Variable

Authors: Xinyuan Y. Song, Kai Kang

Abstract:

Joint models for analyzing longitudinal and survival data are widely used to investigate the relationship between a failure time process and time-variant predictors. A common assumption in conventional joint models in the survival analysis literature is that all predictors are observable. However, this assumption may not always be supported because unobservable traits, namely, latent variables, which are indirectly observable and should be measured through multiple observed variables, are commonly encountered in the medical, behavioral, and financial research settings. In this study, a joint modeling approach to deal with this feature is proposed. The proposed model comprises three parts. The first part is a dynamic factor analysis model for characterizing latent variables through multiple observed indicators over time. The second part is a random coefficient trajectory model for describing the individual trajectories of latent variables. The third part is a proportional hazard model for examining the effects of time-invariant predictors and the longitudinal trajectories of time-variant latent risk factors on hazards of interest. A Bayesian approach coupled with a Markov chain Monte Carlo algorithm to perform statistical inference. An application of the proposed joint model to a study on the Alzheimer's disease neuroimaging Initiative is presented.

Keywords: Bayesian analysis, joint model, longitudinal data, time-to-event data

Procedia PDF Downloads 142
1676 Comparative Study Performance of the Induction Motor between SMC and NLC Modes Control

Authors: A. Oukaci, R. Toufouti, D. Dib, l. Atarsia

Abstract:

This article presents a multitude of alternative techniques to control the vector control, namely the nonlinear control and sliding mode control. Moreover, the implementation of their control law applied to the high-performance to the induction motor with the objective to improve the tracking control, ensure stability robustness to parameter variations and disturbance rejection. Tests are performed numerical simulations in the Matlab/Simulink interface, the results demonstrate the efficiency and dynamic performance of the proposed strategy.

Keywords: Induction Motor (IM), Non-linear Control (NLC), Sliding Mode Control (SMC), nonlinear sliding surface

Procedia PDF Downloads 570
1675 Bayesian Semiparametric Geoadditive Modelling of Underweight Malnutrition of Children under 5 Years in Ethiopia

Authors: Endeshaw Assefa Derso, Maria Gabriella Campolo, Angela Alibrandi

Abstract:

Objectives:Early childhood malnutrition can have long-term and irreversible effects on a child's health and development. This study uses the Bayesian method with spatial variation to investigate the flexible trends of metrical covariates and to identify communities at high risk of injury. Methods: Cross-sectional data on underweight are collected from the 2016 Ethiopian Demographic and Health Survey (EDHS). The Bayesian geo-additive model is performed. Appropriate prior distributions were provided for scall parameters in the models, and the inference is entirely Bayesian, using Monte Carlo Markov chain (MCMC) stimulation. Results: The results show that metrical covariates like child age, maternal body mass index (BMI), and maternal age affect a child's underweight non-linearly. Lower and higher maternal BMI seem to have a significant impact on the child’s high underweight. There was also a significant spatial heterogeneity, and based on IDW interpolation of predictive values, the western, central, and eastern parts of the country are hotspot areas. Conclusion: Socio-demographic and community- based programs development should be considered compressively in Ethiopian policy to combat childhood underweight malnutrition.

Keywords: bayesX, Ethiopia, malnutrition, MCMC, semi-parametric bayesian analysis, spatial distribution, P- splines

Procedia PDF Downloads 86
1674 Prediction of Ionic Liquid Densities Using a Corresponding State Correlation

Authors: Khashayar Nasrifar

Abstract:

Ionic liquids (ILs) exhibit particular properties exemplified by extremely low vapor pressure and high thermal stability. The properties of ILs can be tailored by proper selection of cations and anions. As such, ILs are appealing as potential solvents to substitute traditional solvents with high vapor pressure. One of the IL properties required in chemical and process design is density. In developing corresponding state liquid density correlations, scaling hypothesis is often used. The hypothesis expresses the temperature dependence of saturated liquid densities near the vapor-liquid critical point as a function of reduced temperature. Extending the temperature dependence, several successful correlations were developed to accurately correlate the densities of normal liquids from the triple point to a critical point. Applying mixing rules, the liquid density correlations are extended to liquid mixtures as well. ILs are not molecular liquids, and they are not classified among normal liquids either. Also, ILs are often used where the condition is far from equilibrium. Nevertheless, in calculating the properties of ILs, the use of corresponding state correlations would be useful if no experimental data were available. With well-known generalized saturated liquid density correlations, the accuracy in predicting the density of ILs is not that good. An average error of 4-5% should be expected. In this work, a data bank was compiled. A simplified and concise corresponding state saturated liquid density correlation is proposed by phenomena-logically modifying reduced temperature using the temperature-dependence for an interacting parameter of the Soave-Redlich-Kwong equation of state. This modification improves the temperature dependence of the developed correlation. Parametrization was next performed to optimize the three global parameters of the correlation. The correlation was then applied to the ILs in our data bank with satisfactory predictions. The correlation of IL density applied at 0.1 MPa and was tested with an average uncertainty of around 2%. No adjustable parameter was used. The critical temperature, critical volume, and acentric factor were all required. Methods to extend the predictions to higher pressures (200 MPa) were also devised. Compared to other methods, this correlation was found more accurate. This work also presents the chronological order of developing such correlations dealing with ILs. The pros and cons are also expressed.

Keywords: correlation, corresponding state principle, ionic liquid, density

Procedia PDF Downloads 126
1673 Corellation between Soil Electrical Resistivity and Metal Corrosion Based on Soil Types for Structure Designs

Authors: L. O. A. Oyinkanola, J.A. Fajemiroye

Abstract:

Soil resistivity measurements are an important parameter employed in the designing earthing installations. Thus, The knowledge of soil resistivity with respect to how it varies with related parameters such as moisture content, Temperature and depth at the intended site is very vital to determine how the desired earth resistance value can be attained and sustained over the life of the installation with the lowest cost and effort. The relationship between corrosion and soil resistivity has been investigated in this work. Varios soil samples: Sand, Gravel, Loam, Clay and Silt were collected from different spot within the vicinity.

Keywords: Corrosion, resistivity, clay, hydraulic conductivity

Procedia PDF Downloads 558
1672 Predicting and Optimizing the Mechanical Behavior of a Flax Reinforced Composite

Authors: Georgios Koronis, Arlindo Silva

Abstract:

This study seeks to understand the mechanical behavior of a natural fiber reinforced composite (epoxy/flax) in more depth, utilizing both experimental and numerical methods. It is attempted to identify relationships between the design parameters and the product performance, understand the effect of noise factors and reduce process variations. Optimization of the mechanical performance of manufactured goods has recently been implemented by numerous studies for green composites. However, these studies are limited and have explored in principal mass production processes. It is expected here to discover knowledge about composite’s manufacturing that can be used to design artifacts that are of low batch and tailored to niche markets. The goal is to reach greater consistency in the performance and further understand which factors play significant roles in obtaining the best mechanical performance. A prediction of response function (in various operating conditions) of the process is modeled by the DoE. Normally, a full factorial designed experiment is required and consists of all possible combinations of levels for all factors. An analytical assessment is possible though with just a fraction of the full factorial experiment. The outline of the research approach will comprise of evaluating the influence that these variables have and how they affect the composite mechanical behavior. The coupons will be fabricated by the vacuum infusion process defined by three process parameters: flow rate, injection point position and fiber treatment. Each process parameter is studied at 2-levels along with their interactions. Moreover, the tensile and flexural properties will be obtained through mechanical testing to discover the key process parameters. In this setting, an experimental phase will be followed in which a number of fabricated coupons will be tested to allow for a validation of the design of the experiment’s setup. Finally, the results are validated by performing the optimum set of in a final set of experiments as indicated by the DoE. It is expected that after a good agreement between the predicted and the verification experimental values, the optimal processing parameter of the biocomposite lamina will be effectively determined.

Keywords: design of experiments, flax fabrics, mechanical performance, natural fiber reinforced composites

Procedia PDF Downloads 203
1671 Characters of Developing Commercial Employment Sub-Centres and Employment Density in Ahmedabad City

Authors: Bhaumik Patel, Amit Gotecha

Abstract:

Commercial centres of different hierarchy and sizes play a vital role in the growth and development of the city. Economic uncertainty and demand for space leads to more urban sprawl and emerging more commercial spaces. The study was focused on the understanding of various indicators affecting the commercial development that can help to solve many issues related to commercial urban development and can guide for future employment growth centre development, Accessibility, Infrastructure, Planning and development regulations and Market forces. The aim of the study was to review characteristics and identifying employment density of Commercial Employment Sub-centres by achieving objectives Understanding various employment sub-centres, Identifying characteristics and deriving behaviour of employment densities and Evaluating and comparing employment sub-centres for the Ahmedabad city. Commercial employment sub-centres one in old city (Kalupur), second in highly developed commercial (C.G.road-Ashram road) and third in the latest developing commercial area (Prahladnagar) were identified by distance from city centre, Land use diversity, Access to Major roads and Public transport, Population density in proximity, Complimentary land uses in proximity and Land price. Commercial activities were categorised into retail, wholesale and service sector and sub categorised into various activities. From the study, Time period of establishment of the unit is a critical parameter for commercial activity, building height, and land-use diversity. Employment diversity is also one parameter for the commercial centre. The old city has retail, wholesale and trading and higher commercial density concerning units and employment both. Prahladnagar area functioned as commercial due to market pressure and developed as more units rather than a requirement. Employment density is higher in the centre of the city, as far as distance increases from city centre employment density and unit density decreases. Characters of influencing employment density and unit density are distance from city centre, development type, establishment time period, building density, unit density, public transport accessibility and road connectivity.

Keywords: commercial employment sub-centres, employment density, employment diversity, unit density

Procedia PDF Downloads 137
1670 Hexavalent Chromium-Induced Changes in Biochemical Parameters of Wistar Albino Rats

Authors: Ounassa Adjroud

Abstract:

Potassium dichromate (K2Cr2O7) is one of the most toxic elements to which man can be exposed at work or in the environment. The purpose of the current work is to compare the effect of K2Cr2O7 using variations in the dose, route of administration and duration of exposure in male and female Wistar albino rats with a special focus on biochemical parameters. K2Cr2O7 was subcutaneously administered alone (10, 50 and 100 mg/kg body weight) to female Wistar albino rats. Male rats received in their drinking water K2Cr2O7 30 mg/L/day) for 20 consecutive days. The Biochemical parameters were evaluated on days 3, 6 and 21 after subcutaneous (sc.) treatment in female rats and on days 10 and 20 after oral administration in male rats. The subcutaneous (s.c.) administration of 25 mg/kg of K2Cr2O7 to Wistar albino rats induced a slight change in plasma glucose levels during the experiment period. On the contrary, a significant decrease in plasma glucose levels was observed with 50 mg/kg mainly on days 3 (-26%) and 21 (-48%) after treatment compared to controls females rats. On the other hand, the higher dose provoked a significant increase in plasma glucose concentrations on days 6 (+31%) and 21 (+60%). similarly, the lower dose of chromium had no effect on the plasma urea levels. Conversely, a significant increase (122%) in this parameter was obtained during the first three days after treatment. In addition, a significant decrease in plasma glucose levels was observed with 50 mg/kg mainly on days 3 (-26%) and 21 (-48%) after treatment. On the other hand, the higher dose provoked a significant increase in plasma glucose concentrations on days 6 (+31%) and 21 (+60%). similarly, the lower dose of chromium had no effect on the plasma urea levels. Conversely, a significant increase in this parameter (122%) was obtained during the first three days after treatment. In addition, administration of 100 mg/kg of K2Cr2O7 by s.c markedly augmented the levels of plasma urea on days 3 (62%) and 6 (121%). Administration of 30 mg/L/day of K2Cr2O7 in the drinking water induced a significant augmentation in both of plasma glucose (27%) and urea (126%) during the first ten days of treatment. These results suggested that K2Cr2O7 administered subcutaneously or in the drinking water may induce harmful effects on biochemical parameters.

Keywords: glucose, potassium dichromate, Wistar albino rat, urea

Procedia PDF Downloads 283
1669 Retrospective Data Analysis of Penetrating Injuries Admitted to Jigme Dorji Wangchuck National Referral Hospital (JDWNRH), Thimphu, Bhutan, Due to Traditional Sports over a Period of 3 Years

Authors: Sonam Kelzang

Abstract:

Background: Penetrating injuries as a result of traditional sports (Archery and Khuru) are commonly seen in Bhutan. To our knowledge, there is no study carried out looking into the data of penetrating injuries due to traditional sports. Aim: This is a retrospective analysis of cases of penetrating injuries as a result of traditional sports admitted to JDWNRH over the last 3 years to draw an inference on the pattern of injury and associated morbidity and mortality. Method: Data on penetrating injuries related to traditional sports (Archery and Khuru) were collected and reviewed over the period of 3 years. Assault cases were excluded. For each year we analysed age, sex, parts of the body affected, agent of injury and whether admission was required or not. Results: Out of the total 44 victims of penetrating injury by traditional sports (Archery and Khuru) between 2013 and 2015 (average of 15 cases of penetrating injuries per year). Eighty-five percent were male and 15% were female. Their age ranged from 4 yrs to 62 years. Sixty-one percent of the victims were in the working age group of 19-58 years; 30% of the victims were referred from various district hospitals; 38% of the victims needed admission; 42 % of the victims suffered injury to the head; and 54% of the injuries were caused by Khuru. Conclusion: Penetrating injuries due to traditional sports admitted to JDWNRH, Thimphu, remained same over the three years period despite safety regulations in place. Although there were no deaths during the last three years, morbidity still remains high.

Keywords: archery, Bhutan, Khuru, darts

Procedia PDF Downloads 165
1668 Bayesian Inference for High Dimensional Dynamic Spatio-Temporal Models

Authors: Sofia M. Karadimitriou, Kostas Triantafyllopoulos, Timothy Heaton

Abstract:

Reduced dimension Dynamic Spatio-Temporal Models (DSTMs) jointly describe the spatial and temporal evolution of a function observed subject to noise. A basic state space model is adopted for the discrete temporal variation, while a continuous autoregressive structure describes the continuous spatial evolution. Application of such a DSTM relies upon the pre-selection of a suitable reduced set of basic functions and this can present a challenge in practice. In this talk, we propose an online estimation method for high dimensional spatio-temporal data based upon DSTM and we attempt to resolve this issue by allowing the basis to adapt to the observed data. Specifically, we present a wavelet decomposition in order to obtain a parsimonious approximation of the spatial continuous process. This parsimony can be achieved by placing a Laplace prior distribution on the wavelet coefficients. The aim of using the Laplace prior, is to filter wavelet coefficients with low contribution, and thus achieve the dimension reduction with significant computation savings. We then propose a Hierarchical Bayesian State Space model, for the estimation of which we offer an appropriate particle filter algorithm. The proposed methodology is illustrated using real environmental data.

Keywords: multidimensional Laplace prior, particle filtering, spatio-temporal modelling, wavelets

Procedia PDF Downloads 424