Search results for: daily probability model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19542

Search results for: daily probability model

16032 Mixtures of Length-Biased Weibull Distributions for Loss Severity Modelling

Authors: Taehan Bae

Abstract:

In this paper, a class of length-biased Weibull mixtures is presented to model loss severity data. The proposed model generalizes the Erlang mixtures with the common scale parameter, and it shares many important modelling features, such as flexibility to fit various data distribution shapes and weak-denseness in the class of positive continuous distributions, with the Erlang mixtures. We show that the asymptotic tail estimate of the length-biased Weibull mixture is Weibull-type, which makes the model effective to fit loss severity data with heavy-tailed observations. A method of statistical estimation is discussed with applications on real catastrophic loss data sets.

Keywords: Erlang mixture, length-biased distribution, transformed gamma distribution, asymptotic tail estimate, EM algorithm, expectation-maximization algorithm

Procedia PDF Downloads 220
16031 Timetabling for Interconnected LRT Lines: A Package Solution Based on a Real-world Case

Authors: Huazhen Lin, Ruihua Xu, Zhibin Jiang

Abstract:

In this real-world case, timetabling the LRT network as a whole is rather challenging for the operator: they are supposed to create a timetable to avoid various route conflicts manually while satisfying a given interval and the number of rolling stocks, but the outcome is not satisfying. Therefore, the operator adopts a computerised timetabling tool, the Train Plan Maker (TPM), to cope with this problem. However, with various constraints in the dual-line network, it is still difficult to find an adequate pairing of turnback time, interval and rolling stocks’ number, which requires extra manual intervention. Aiming at current problems, a one-off model for timetabling is presented in this paper to simplify the procedure of timetabling. Before the timetabling procedure starts, this paper presents how the dual-line system with a ring and several branches is turned into a simpler structure. Then, a non-linear programming model is presented in two stages. In the first stage, the model sets a series of constraints aiming to calculate a proper timing for coordinating two lines by adjusting the turnback time at termini. Then, based on the result of the first stage, the model introduces a series of inequality constraints to avoid various route conflicts. With this model, an analysis is conducted to reveal the relation between the ratio of trains in different directions and the possible minimum interval, observing that the more imbalance the ratio is, the less possible to provide frequent service under such strict constraints.

Keywords: light rail transit (LRT), non-linear programming, railway timetabling, timetable coordination

Procedia PDF Downloads 77
16030 The Effect of Mobile Technology Use in Education: A Meta-Analysis Study

Authors: Şirin Küçük, Ayşe Kök, İsmail Şahin

Abstract:

Mobile devices are very popular and useful tools for assisting people in daily life. With the advancement of mobile technologies, the issue of mobile learning has been widely investigated in education. Many researches consider that it is important to integrate pedagogical and technical strengths of mobile technology into learning environments. For this reason, the purpose of this research is to examine the effect of mobile technology use in education with meta-analysis method. Meta-analysis is a statistical technique which combines the findings of independent studies in a specific subject. In this respect, the articles will be examined by searching the databases for researches which are conducted between 2005 and 2014. It is expected that the results of this research will contribute to future research related to mobile technology use in education.

Keywords: mobile learning, meta-analysis, mobile technology, education

Procedia PDF Downloads 718
16029 In situ Real-Time Multivariate Analysis of Methanolysis Monitoring of Sunflower Oil Using FTIR

Authors: Pascal Mwenge, Tumisang Seodigeng

Abstract:

The combination of world population and the third industrial revolution led to high demand for fuels. On the other hand, the decrease of global fossil 8fuels deposits and the environmental air pollution caused by these fuels has compounded the challenges the world faces due to its need for energy. Therefore, new forms of environmentally friendly and renewable fuels such as biodiesel are needed. The primary analytical techniques for methanolysis yield monitoring have been chromatography and spectroscopy, these methods have been proven reliable but are more demanding, costly and do not provide real-time monitoring. In this work, the in situ monitoring of biodiesel from sunflower oil using FTIR (Fourier Transform Infrared) has been studied; the study was performed using EasyMax Mettler Toledo reactor equipped with a DiComp (Diamond) probe. The quantitative monitoring of methanolysis was performed by building a quantitative model with multivariate calibration using iC Quant module from iC IR 7.0 software. 15 samples of known concentrations were used for the modelling which were taken in duplicate for model calibration and cross-validation, data were pre-processed using mean centering and variance scale, spectrum math square root and solvent subtraction. These pre-processing methods improved the performance indexes from 7.98 to 0.0096, 11.2 to 3.41, 6.32 to 2.72, 0.9416 to 0.9999, RMSEC, RMSECV, RMSEP and R2Cum, respectively. The R2 value of 1 (training), 0.9918 (test), 0.9946 (cross-validation) indicated the fitness of the model built. The model was tested against univariate model; small discrepancies were observed at low concentration due to unmodelled intermediates but were quite close at concentrations above 18%. The software eliminated the complexity of the Partial Least Square (PLS) chemometrics. It was concluded that the model obtained could be used to monitor methanol of sunflower oil at industrial and lab scale.

Keywords: biodiesel, calibration, chemometrics, methanolysis, multivariate analysis, transesterification, FTIR

Procedia PDF Downloads 143
16028 Numerical Simulation of the Dynamic Behavior of a LaNi5 Water Pumping System

Authors: Miled Amel, Ben Maad Hatem, Askri Faouzi, Ben Nasrallah Sassi

Abstract:

Metal hydride water pumping system uses hydrogen as working fluid to pump water for low head and high discharge. The principal operation of this pump is based on the desorption of hydrogen at high pressure and its absorption at low pressure by a metal hydride. This work is devoted to study a concept of the dynamic behavior of a metal hydride pump using unsteady model and LaNi5 as hydriding alloy. This study shows that with MHP, it is possible to pump 340l/kg-cycle of water in 15 000s using 1 Kg of LaNi5 at a desorption temperature of 360 K, a pumping head equal to 5 m and a desorption gear ratio equal to 33. This study reveals also that the error given by the steady model, using LaNi5 is about 2%.A dimensional mathematical model and the governing equations of the pump were presented to predict the coupled heat and mass transfer within the MHP. Then, a numerical simulation is carried out to present the time evolution of the specific water discharge and to test the effect of different parameters (desorption temperature, absorption temperature, desorption gear ratio) on the performance of the water pumping system (specific water discharge, pumping efficiency and pumping time). In addition, a comparison between results obtained with steady and unsteady model is performed with different hydride mass. Finally, a geometric configuration of the reactor is simulated to optimize the pumping time.

Keywords: dynamic behavior, LaNi5, performance of water pumping system, unsteady model

Procedia PDF Downloads 197
16027 The Effect of Initial Sample Size and Increment in Simulation Samples on a Sequential Selection Approach

Authors: Mohammad H. Almomani

Abstract:

In this paper, we argue the effect of the initial sample size, and the increment in simulation samples on the performance of a sequential approach that used in selecting the top m designs when the number of alternative designs is very large. The sequential approach consists of two stages. In the first stage the ordinal optimization is used to select a subset that overlaps with the set of actual best k% designs with high probability. Then in the second stage the optimal computing budget is used to select the top m designs from the selected subset. We apply the selection approach on a generic example under some parameter settings, with a different choice of initial sample size and the increment in simulation samples, to explore the impacts on the performance of this approach. The results show that the choice of initial sample size and the increment in simulation samples does affect the performance of a selection approach.

Keywords: Large Scale Problems, Optimal Computing Budget Allocation, ordinal optimization, simulation optimization

Procedia PDF Downloads 352
16026 Evidence of Conditional and Unconditional Cooperation in a Public Goods Game: Experimental Evidence from Mali

Authors: Maria Laura Alzua, Maria Adelaida Lopera

Abstract:

This paper measures the relative importance of conditional cooperation and unconditional cooperation in a large public goods experiment conducted in Mali. We use expectations about total public goods provision to estimate a structural choice model with heterogeneous preferences. While unconditional cooperation can be captured by common preferences shared by all participants, conditional cooperation is much more heterogeneous and depends on unobserved individual factors. This structural model, in combination with two experimental treatments, suggests that leadership and group communication incentivize public goods provision through different channels. First, We find that participation of local leaders effectively changes individual choices through unconditional cooperation. A simulation exercise predicts that even in the most pessimistic scenario in which all participants expect zero public good provision, 60% would still choose to cooperate. Second, allowing participants to communicate fosters conditional cooperation. The simulations suggest that expectations are responsible for around 24% of the observed public good provision and that group communication does not necessarily ameliorate public good provision. In fact, communication may even worsen the outcome when expectations are low.

Keywords: conditional cooperation, discrete choice model, expectations, public goods game, random coefficients model

Procedia PDF Downloads 303
16025 High-Speed Particle Image Velocimetry of the Flow around a Moving Train Model with Boundary Layer Control Elements

Authors: Alexander Buhr, Klaus Ehrenfried

Abstract:

Trackside induced airflow velocities, also known as slipstream velocities, are an important criterion for the design of high-speed trains. The maximum permitted values are given by the Technical Specifications for Interoperability (TSI) and have to be checked in the approval process. For train manufactures it is of great interest to know in advance, how new train geometries would perform in TSI tests. The Reynolds number in moving model experiments is lower compared to full-scale. Especially the limited model length leads to a thinner boundary layer at the rear end. The hypothesis is that the boundary layer rolls up to characteristic flow structures in the train wake, in which the maximum flow velocities can be observed. The idea is to enlarge the boundary layer using roughness elements at the train model head so that the ratio between the boundary layer thickness and the car width at the rear end is comparable to a full-scale train. This may lead to similar flow structures in the wake and better prediction accuracy for TSI tests. In this case, the design of the roughness elements is limited by the moving model rig. Small rectangular roughness shapes are used to get a sufficient effect on the boundary layer, while the elements are robust enough to withstand the high accelerating and decelerating forces during the test runs. For this investigation, High-Speed Particle Image Velocimetry (HS-PIV) measurements on an ICE3 train model have been realized in the moving model rig of the DLR in Göttingen, the so called tunnel simulation facility Göttingen (TSG). The flow velocities within the boundary layer are analysed in a plain parallel to the ground. The height of the plane corresponds to a test position in the EN standard (TSI). Three different shapes of roughness elements are tested. The boundary layer thickness and displacement thickness as well as the momentum thickness and the form factor are calculated along the train model. Conditional sampling is used to analyse the size and dynamics of the flow structures at the time of maximum velocity in the train wake behind the train. As expected, larger roughness elements increase the boundary layer thickness and lead to larger flow velocities in the boundary layer and in the wake flow structures. The boundary layer thickness, displacement thickness and momentum thickness are increased by using larger roughness especially when applied in the height close to the measuring plane. The roughness elements also cause high fluctuations in the form factors of the boundary layer. Behind the roughness elements, the form factors rapidly are approaching toward constant values. This indicates that the boundary layer, while growing slowly along the second half of the train model, has reached a state of equilibrium.

Keywords: boundary layer, high-speed PIV, ICE3, moving train model, roughness elements

Procedia PDF Downloads 302
16024 Lab Bench for Synthetic Aperture Radar Imaging System

Authors: Karthiyayini Nagarajan, P. V. Ramakrishna

Abstract:

Radar Imaging techniques provides extensive applications in the field of remote sensing, majorly Synthetic Aperture Radar (SAR) that provide high resolution target images. This paper work puts forward the effective and realizable signal generation and processing for SAR images. The major units in the system include camera, signal generation unit, signal processing unit and display screen. The real radio channel is replaced by its mathematical model based on optical image to calculate a reflected signal model in real time. Signal generation realizes the algorithm and forms the radar reflection model. Signal processing unit provides range and azimuth resolution through matched filtering and spectrum analysis procedure to form radar image on the display screen. The restored image has the same quality as that of the optical image. This SAR imaging system has been designed and implemented using MATLAB and Quartus II tools on Stratix III device as a System (Lab Bench) that works in real time to study/investigate on radar imaging rudiments and signal processing scheme for educational and research purposes.

Keywords: synthetic aperture radar, radio reflection model, lab bench, imaging engineering

Procedia PDF Downloads 492
16023 Evaluating the Effectiveness of Plantar Sensory Insoles and Remote Patient Monitoring for Early Intervention in Diabetic Foot Ulcer Prevention in Patients with Peripheral Neuropathy

Authors: Brock Liden, Eric Janowitz

Abstract:

Introduction: Diabetic peripheral neuropathy (DPN) affects 70% of individuals with diabetes1. DPN causes a loss of protective sensation, which can lead to tissue damage and diabetic foot ulcer (DFU) formation2. These ulcers can result in infections and lower-extremity amputations of toes, the entire foot, and the lower leg. Even after a DFU is healed, recurrence is common, with 49% of DFU patients developing another ulcer within a year and 68% within 5 years3. This case series examines the use of sensory insoles and newly available plantar data (pressure, temperature, step count, adherence) and remote patient monitoring in patients at risk of DFU. Methods: Participants were provided with custom-made sensory insoles to monitor plantar pressure, temperature, step count, and daily use and were provided with real-time cues for pressure offloading as they went about their daily activities. The sensory insoles were used to track subject compliance, ulceration, and response to feedback from real-time alerts. Patients were remotely monitored by a qualified healthcare professional and were contacted when areas of concern were seen and provided coaching on reducing risk factors and overall support to improve foot health. Results: Of the 40 participants provided with the sensory insole system, 4 presented with a DFU. Based on flags generated from the available plantar data, patients were contacted by the remote monitor to address potential concerns. A standard clinical escalation protocol detailed when and how concerns should be escalated to the provider by the remote monitor. Upon escalation to the provider, patients were brought into the clinic as needed, allowing for any issues to be addressed before more serious complications might arise. Conclusion: This case series explores the use of innovative sensory technology to collect plantar data (pressure, temperature, step count, and adherence) for DFU detection and early intervention. The results from this case series suggest the importance of sensory technology and remote patient monitoring in providing proactive, preventative care for patients at risk of DFU. This robust plantar data, with the addition of remote patient monitoring, allow for patients to be seen in the clinic when concerns arise, giving providers the opportunity to intervene early and prevent more serious complications, such as wounds, from occurring.

Keywords: diabetic foot ulcer, DFU prevention, digital therapeutics, remote patient monitoring

Procedia PDF Downloads 72
16022 Gaussian Mixture Model Based Identification of Arterial Wall Movement for Computation of Distension Waveform

Authors: Ravindra B. Patil, P. Krishnamoorthy, Shriram Sethuraman

Abstract:

This work proposes a novel Gaussian Mixture Model (GMM) based approach for accurate tracking of the arterial wall and subsequent computation of the distension waveform using Radio Frequency (RF) ultrasound signal. The approach was evaluated on ultrasound RF data acquired using a prototype ultrasound system from an artery mimicking flow phantom. The effectiveness of the proposed algorithm is demonstrated by comparing with existing wall tracking algorithms. The experimental results show that the proposed method provides 20% reduction in the error margin compared to the existing approaches in tracking the arterial wall movement. This approach coupled with ultrasound system can be used to estimate the arterial compliance parameters required for screening of cardiovascular related disorders.

Keywords: distension waveform, Gaussian Mixture Model, RF ultrasound, arterial wall movement

Procedia PDF Downloads 498
16021 Impact of Workers’ Remittances on Poverty in Pakistan: A Time Series Analysis by Ardl

Authors: Syed Aziz Rasool, Ayesha Zaman

Abstract:

Poverty is one of the most important problems for any developing nation. Workers’ remittances and investment plays a crucial role in development of any country by reducing the poverty level in Pakistan. This research studies the relationship between workers’ remittances and poverty alleviation. It also focused the significant effect on poverty reduction. This study uses time series data for the period of 1972-2013. Autoregressive Distributed Lag (ARDL)Model and Error Correction (ECM)Model has been used in order to find out the long run and short run relationship between the worker’s remittances and poverty level respectively. Thus, inflow of remittances showed the significant and negative impact on poverty level. Moreover, coefficient of error correction model explains the adjustment towards convergence and it has highly significant and negative value. According to this research, Policy makers should strongly focus on positive and effective policies to attract more remittances. JELCODE: JEL: J61

Keywords: ECM, ARDL, AIC, SC

Procedia PDF Downloads 280
16020 Service Quality Improvement in Ghana's Healthcare Supply Chain

Authors: Ammatu Alhassan

Abstract:

Quality healthcare delivery is a crucial indicator in assessing the overall developmental status of a country. There are many limitations in the Ghanaian healthcare supply chain due to the lack of studies about the correlation between quality health service and the healthcare supply chain. Patients who visit various healthcare providers face unpleasant experiences such as delays in the availability of their medications. In this study, an assessment of the quality of services provided to Ghanaian outpatients who visit public healthcare providers was investigated to establish its effect on the healthcare supply chain using a conceptual model. The Donabedian’s structure, process, and outcome theory for service quality evaluation were used to analyse 20 Ghanaian hospitals. The data obtained was tested using the structural equation model (SEM). The findings from this research will help us to improve the overall quality of the Ghanaian healthcare supply chain. The model which will be developed will help us to understand better the linkage between quality healthcare and the healthcare supply chain as well as serving as a reference tool for future healthcare research in Ghana.

Keywords: Ghana, healthcare, outpatients, supply chain

Procedia PDF Downloads 179
16019 A Regression Model for Predicting Sugar Crystal Size in a Fed-Batch Vacuum Evaporative Crystallizer

Authors: Sunday B. Alabi, Edikan P. Felix, Aniediong M. Umo

Abstract:

Crystal size distribution is of great importance in the sugar factories. It determines the market value of granulated sugar and also influences the cost of production of sugar crystals. Typically, sugar is produced using fed-batch vacuum evaporative crystallizer. The crystallization quality is examined by crystal size distribution at the end of the process which is quantified by two parameters: the average crystal size of the distribution in the mean aperture (MA) and the width of the distribution of the coefficient of variation (CV). Lack of real-time measurement of the sugar crystal size hinders its feedback control and eventual optimisation of the crystallization process. An attractive alternative is to use a soft sensor (model-based method) for online estimation of the sugar crystal size. Unfortunately, the available models for sugar crystallization process are not suitable as they do not contain variables that can be measured easily online. The main contribution of this paper is the development of a regression model for estimating the sugar crystal size as a function of input variables which are easy to measure online. This has the potential to provide real-time estimates of crystal size for its effective feedback control. Using 7 input variables namely: initial crystal size (Lo), temperature (T), vacuum pressure (P), feed flowrate (Ff), steam flowrate (Fs), initial super-saturation (S0) and crystallization time (t), preliminary studies were carried out using Minitab 14 statistical software. Based on the existing sugar crystallizer models, and the typical ranges of these 7 input variables, 128 datasets were obtained from a 2-level factorial experimental design. These datasets were used to obtain a simple but online-implementable 6-input crystal size model. It seems the initial crystal size (Lₒ) does not play a significant role. The goodness of the resulting regression model was evaluated. The coefficient of determination, R² was obtained as 0.994, and the maximum absolute relative error (MARE) was obtained as 4.6%. The high R² (~1.0) and the reasonably low MARE values are an indication that the model is able to predict sugar crystal size accurately as a function of the 6 easy-to-measure online variables. Thus, the model can be used as a soft sensor to provide real-time estimates of sugar crystal size during sugar crystallization process in a fed-batch vacuum evaporative crystallizer.

Keywords: crystal size, regression model, soft sensor, sugar, vacuum evaporative crystallizer

Procedia PDF Downloads 203
16018 Optimal Bayesian Chart for Controlling Expected Number of Defects in Production Processes

Authors: V. Makis, L. Jafari

Abstract:

In this paper, we develop an optimal Bayesian chart to control the expected number of defects per inspection unit in production processes with long production runs. We formulate this control problem in the optimal stopping framework. The objective is to determine the optimal stopping rule minimizing the long-run expected average cost per unit time considering partial information obtained from the process sampling at regular epochs. We prove the optimality of the control limit policy, i.e., the process is stopped and the search for assignable causes is initiated when the posterior probability that the process is out of control exceeds a control limit. An algorithm in the semi-Markov decision process framework is developed to calculate the optimal control limit and the corresponding average cost. Numerical examples are presented to illustrate the developed optimal control chart and to compare it with the traditional u-chart.

Keywords: Bayesian u-chart, economic design, optimal stopping, semi-Markov decision process, statistical process control

Procedia PDF Downloads 570
16017 Analyzing the Effects of Supply and Demand Shocks in the Spanish Economy

Authors: José M Martín-Moreno, Rafaela Pérez, Jesús Ruiz

Abstract:

In this paper we use a small open economy Dynamic Stochastic General Equilibrium Model (DSGE) for the Spanish economy to search for a deeper characterization of the determinants of Spain’s macroeconomic fluctuations throughout the period 1970-2008. In order to do this, we distinguish between tradable and non-tradable goods to take into account the fact that the presence of non-tradable goods in this economy is one of the largest in the world. We estimate a DSGE model with supply and demand shocks (sectorial productivity, public spending, international real interest rate and preferences) using Kalman Filter techniques. We find the following results. First of all, our variance decomposition analysis suggests that 1) the preference shock basically accounts for private consumption volatility, 2) the idiosyncratic productivity shock accounts for non-tradable output volatility, and 3) the sectorial productivity shock along with the international interest rate both greatly account for tradable output. Secondly, the model closely replicates the time path observed in the data for the Spanish economy and finally, the model captures the main cyclical qualitative features of this economy reasonably well.

Keywords: business cycle, DSGE models, Kalman filter estimation, small open economy

Procedia PDF Downloads 414
16016 The Acquisition of Case in Biological Domain Based on Text Mining

Authors: Shen Jian, Hu Jie, Qi Jin, Liu Wei Jie, Chen Ji Yi, Peng Ying Hong

Abstract:

In order to settle the problem of acquiring case in biological related to design problems, a biometrics instance acquisition method based on text mining is presented. Through the construction of corpus text vector space and knowledge mining, the feature selection, similarity measure and case retrieval method of text in the field of biology are studied. First, we establish a vector space model of the corpus in the biological field and complete the preprocessing steps. Then, the corpus is retrieved by using the vector space model combined with the functional keywords to obtain the biological domain examples related to the design problems. Finally, we verify the validity of this method by taking the example of text.

Keywords: text mining, vector space model, feature selection, biologically inspired design

Procedia PDF Downloads 258
16015 Adjustment of Parents of Children with Autism: A Multivariate Model

Authors: Ayelet Siman-Tov, Shlomo Kaniel

Abstract:

Objectives: The research validates a multivariate model that predicts parental adjustment to coping successfully with an autistic child. The model comprises four elements: parental stress, parental resources, parental adjustment and the child's autism symptoms. Background and aims: The purpose of the current study is the construction and validation of a model for the adjustment of parents and a child with autism. The suggested model is based on theoretical views on stress and links personal resources, stress, perception, parental mental health and quality of marriage and child adjustment with autism. The family stress approach focuses on the family as a system made up of a dynamic interaction between its members, who constitute interdependent parts of the system, and thus, a change in one family member brings about changes in the processes of the entire family system. From this perspective, a rise of new demands in the family and stress in the role of one family member affects the family system as a whole. Materials and methods: 176 parents of children aged between 6 to 16 diagnosed with ASD answered several questionnaires measuring parental stress, personal resources (sense of coherence, locus of control, social support), adjustment (mental health and marriage quality) and the child's autism symptoms. Results: Path analysis showed that a sense of coherence, internal locus of control, social support and quality of marriage increase the ability to cope with the stress of parenting an autistic child. Directions for further research are suggested.

Keywords: stress, adjustment, resources, Autism, parents, coherence

Procedia PDF Downloads 132
16014 Methadone Maintenance Treatment Patients' and Medical Students' Common Trait: Low Mindfulness Trait Associated with High Perceived Stress

Authors: Einat Peles, Anat Sason, Ariel Claman, Gabriel Barkay, Miriam Adelson

Abstract:

Individuals with opioid addiction are characterized as suffering from stress responses disturbance, including the hypothalamic-pituitary-adrenal (HPA) axis, and autonomic nervous system function. HPA axis is known to be stabilized during methadone maintenance treatment (MMT). Mindfulness (present-oriented, nonjudgmental awareness of cognitions, emotions, perceptions, and habitual behavioral reactions in daily life) counteracts stress. To our knowledge, the relation between perceived stress and mindfulness trait among MMT patients has never been studied. To measure indices of mindfulness and their relation to perceived stress among MMT patients, a cross-sectional random sample of current MMT patients was performed using questionnaires for perceived stress (PSS) and mindfulness trait (FFMQ- yields a total score and individual scores for five internally consistent mindfulness factors: Observing, Describing, Acting with awareness and consciousness, Non-judging the inner experience, Non-reactivity to the inner experience). Two additional groups were studied to serve as reference groups; Medical students that are known to suffer from stress, and Axis II psychiatric diagnosis patients that are known to characterized with poor mindfulness trait. Results: Groups included 41 MMT patients, 27 Axis II patients and 36 medical students. High perceived stressed (PSS≥18) defined among 61% of the MMT patients and 50% of the medical students. Highest mindfulness score observed among non-stressed MMT patients (153.5±17.2) followed by the groups of stressed MMT and non-stressed student (128.9±17.0 and 130.5±13.3 respectively), with the lowest score among stressed students (116.3±17.9) (multivariate analyses, corrected model p (F=14.3) < 0.0005, p (group) < 0.0005, p (stress) < 0.0005, p (interaction) =0.2). Linear inverse correlations were found between perceived stress score and mindfulness score among MMT patients (R=-0.65, p < 0.0005) and students (R=-0.51, p=0.002). Axis II patients had the lowest mindfulness score (103.4±25.3). Conclusion: High prevalence of high perceived stressed which characterized with poor mindfulness trait observed in both MMT patients and medical students, two different population groups. The effectiveness of mindfulness treatment in reducing stress and improve mindfulness trait should be evaluated to improve rehabilitation of MMT patients, and students success.

Keywords: mindfulness, stress, methadone maintenance treatment, medical students

Procedia PDF Downloads 180
16013 Thermodynamics of Aqueous Solutions of Organic Molecule and Electrolyte: Use Cloud Point to Obtain Better Estimates of Thermodynamic Parameters

Authors: Jyoti Sahu, Vinay A. Juvekar

Abstract:

Electrolytes are often used to bring about salting-in and salting-out of organic molecules and polymers (e.g. polyethylene glycols/proteins) from the aqueous solutions. For quantification of these phenomena, a thermodynamic model which can accurately predict activity coefficient of electrolyte as a function of temperature is needed. The thermodynamics models available in the literature contain a large number of empirical parameters. These parameters are estimated using lower/upper critical solution temperature of the solution in the electrolyte/organic molecule at different temperatures. Since the number of parameters is large, inaccuracy can bethe creep in during their estimation, which can affect the reliability of prediction beyond the range in which these parameters are estimated. Cloud point of solution is related to its free energy through temperature and composition derivative. Hence, the Cloud point measurement can be used for accurate estimation of the temperature and composition dependence of parameters in the model for free energy. Hence, if we use a two pronged procedure in which we first use cloud point of solution to estimate some of the parameters of the thermodynamic model and determine the rest using osmotic coefficient data, we gain on two counts. First, since the parameters, estimated in each of the two steps, are fewer, we achieve higher accuracy of estimation. The second and more important gain is that the resulting model parameters are more sensitive to temperature. This is crucial when we wish to use the model outside temperatures window within which the parameter estimation is sought. The focus of the present work is to prove this proposition. We have used electrolyte (NaCl/Na2CO3)-water-organic molecule (Iso-propanol/ethanol) as the model system. The model of Robinson-Stokes-Glukauf is modified by incorporating the temperature dependent Flory-Huggins interaction parameters. The Helmholtz free energy expression contains, in addition to electrostatic and translational entropic contributions, three Flory-Huggins pairwise interaction contributions viz., and (w-water, p-polymer, s-salt). These parameters depend both on temperature and concentrations. The concentration dependence is expressed in the form of a quadratic expression involving the volume fractions of the interacting species. The temperature dependence is expressed in the form .To obtain the temperature-dependent interaction parameters for organic molecule-water and electrolyte-water systems, Critical solution temperature of electrolyte -water-organic molecules is measured using cloud point measuring apparatus The temperature and composition dependent interaction parameters for electrolyte-water-organic molecule are estimated through measurement of cloud point of solution. The model is used to estimate critical solution temperature (CST) of electrolyte water-organic molecules solution. We have experimentally determined the critical solution temperature of different compositions of electrolyte-water-organic molecule solution and compared the results with the estimates based on our model. The two sets of values show good agreement. On the other hand when only osmotic coefficients are used for estimation of the free energy model, CST predicted using the resulting model show poor agreement with the experiments. Thus, the importance of the CST data in the estimation of parameters of the thermodynamic model is confirmed through this work.

Keywords: concentrated electrolytes, Debye-Hückel theory, interaction parameters, Robinson-Stokes-Glueckauf model, Flory-Huggins model, critical solution temperature

Procedia PDF Downloads 389
16012 Ottoman Archaeology in Kostence (Constanta, Romania): A Locality on the Periphery of the Ottoman World

Authors: Margareta Simina Stanc, Aurel Mototolea, Tiberiu Potarniche

Abstract:

The city of Constanta (former Köstence) is located in the Dobrogea region, on the west shore of the Black Sea. Between 1420-1878, Dobrogea was a possession of the Ottoman Empire. Archaeological researches starting with the second half of the 20th century revealed various traces of the Ottoman period in this region. Between 2016-2018, preventive archaeological research conducted in the perimeter of the old Ottoman city of Köstence led to the discovery of structures of habitation as well as of numerous artifacts of the Ottoman period (pottery, coins, buckles, etc.). This study uses the analysis of these new discoveries to complete the picture of daily life in the Ottoman period. In 2017, in the peninsular area of Constanta, preventive archaeological research began at a point in the former Ottoman area. In the range between the current ironing level and the -1.5m depth, the Ottoman period materials appeared constantly. It is worth noting the structure of a large building that has been repaired at least once but could not be fully investigated. In parallel to this wall, there was arranged a transversally arranged brick-lined drainage channel. The drainage channel is poured into a tank (hazna), filled with various vintage materials, but mainly gilded ceramics and iron objects. This type of hazna is commonly found in Constanta for the pre-modern and modern period due to the lack of a sewage system in the peninsular area. A similar structure, probably fountain, was discovered in 2016 in another part of the old city. An interesting piece is that of a cup (probably) Persians and a bowl belonging to Kütahya style, both of the 17th century, proof of commercial routes passing through Constanta during that period and indirectly confirming the documentary testimonies of the time. Also, can be mentioned the discovery, in the year 2016, on the occasion of underwater research carried out by specialists of the department of the Constanta Museum, at a depth of 15 meters, a Turkish oil lamp (17th - the beginning of the 18th century), among other objects of a sunken ship. The archaeological pieces, in a fragmentary or integral state, found in research campaigns 2016-2018, are undergoing processing or restoration, leaving out all the available information, and establishing exact analogies. These discoveries bring new data to the knowledge of daily life during the Ottoman administration in the former Köstence, a locality on the periphery of the Islamic world.

Keywords: habitation, material culture, Ottoman administration, Ottoman archaeology, periphery

Procedia PDF Downloads 128
16011 Influence of Iron Ore Mineralogy on Cluster Formation inside the Shaft Furnace

Authors: M. Bahgat, H. A. Hanafy, S. Lakdawala

Abstract:

Clustering phenomenon of pellets was observed frequently in shaft processes operating at higher temperatures. Clustering is a result of the growth of fibrous iron precipitates (iron whiskers) that become hooked to each other and finally become crystallized during the initial stages of metallization. If the pellet clustering is pronounced, sometimes leads to blocking inside the furnace and forced shutdown takes place. This work clarifies further the relation between metallic iron whisker growth and iron ore mineralogy. Various pellet sizes (6 – 12.0 & +12.0 mm) from three different ores (A, B & C) were (completely and partially) reduced at 985 oC with H2/CO gas mixture using thermos-gravimetric technique. It was found that reducibility increases by decreasing the iron ore pellet’s size. Ore (A) has the highest reducibility than ore (B) and ore (C). Increasing the iron ore pellet’s size leads to increase the probability of metallic iron whisker formation. Ore (A) has the highest tendency for metallic iron whisker formation than ore (B) and ore (C). The reduction reactions for all iron ores A, B and C are mainly controlled by diffusion reaction mechanism.

Keywords: shaft furnace, cluster, metallic iron whisker, mineralogy, ferrous metallurgy

Procedia PDF Downloads 464
16010 Assessing Denitrification-Disintegration Model’s Efficacy in Simulating Greenhouse Gas Emissions, Crop Growth, Yield, and Soil Biochemical Processes in Moroccan Context

Authors: Mohamed Boullouz, Mohamed Louay Metougui

Abstract:

Accurate modeling of greenhouse gas (GHG) emissions, crop growth, soil productivity, and biochemical processes is crucial considering escalating global concerns about climate change and the urgent need to improve agricultural sustainability. The application of the denitrification-disintegration (DNDC) model in the context of Morocco's unique agro-climate is thoroughly investigated in this study. Our main research hypothesis is that the DNDC model offers an effective and powerful tool for precisely simulating a wide range of significant parameters, including greenhouse gas emissions, crop growth, yield potential, and complex soil biogeochemical processes, all consistent with the intricate features of environmental Moroccan agriculture. In order to verify these hypotheses, a vast amount of field data covering Morocco's various agricultural regions and encompassing a range of soil types, climatic factors, and crop varieties had to be gathered. These experimental data sets will serve as the foundation for careful model calibration and subsequent validation, ensuring the accuracy of simulation results. In conclusion, the prospective research findings add to the global conversation on climate-resilient agricultural practices while encouraging the promotion of sustainable agricultural models in Morocco. A policy architect's and an agricultural actor's ability to make informed decisions that not only advance food security but also environmental stability may be strengthened by the impending recognition of the DNDC model as a potent simulation tool tailored to Moroccan conditions.

Keywords: greenhouse gas emissions, DNDC model, sustainable agriculture, Moroccan cropping systems

Procedia PDF Downloads 59
16009 Design and Implementation of a Lab Bench for Synthetic Aperture Radar Imaging System

Authors: Karthiyayini Nagarajan, P. V. RamaKrishna

Abstract:

Radar Imaging techniques provides extensive applications in the field of remote sensing, majorly Synthetic Aperture Radar(SAR) that provide high resolution target images. This paper work puts forward the effective and realizable signal generation and processing for SAR images. The major units in the system include camera, signal generation unit, signal processing unit and display screen. The real radio channel is replaced by its mathematical model based on optical image to calculate a reflected signal model in real time. Signal generation realizes the algorithm and forms the radar reflection model. Signal processing unit provides range and azimuth resolution through matched filtering and spectrum analysis procedure to form radar image on the display screen. The restored image has the same quality as that of the optical image. This SAR imaging system has been designed and implemented using MATLAB and Quartus II tools on Stratix III device as a System(lab bench) that works in real time to study/investigate on radar imaging rudiments and signal processing scheme for educational and research purposes.

Keywords: synthetic aperture radar, radio reflection model, lab bench

Procedia PDF Downloads 462
16008 Seismic Performance of a Framed Structure Retrofitted with Damped Cable Systems

Authors: Asad Naeem, Minsung Kim, Jinkoo Kim

Abstract:

In this work, the effectiveness of damped cable systems (DCS) on the mitigation of earthquake-induced response of a framed structure is investigated. The seismic performance of DCS is investigated using fragility analysis and life cycle cost evaluation of an existing building retrofitted with DCS, and the results are compared with those of the structure retrofitted with viscous dampers. The comparison of the analysis results reveals that, due to the self-centering capability of the DCS, residual displacement becomes nearly zero in the structure retrofitted with the DCS. According to the fragility analysis, the structure retrofitted with the DCS has smaller probability of reaching a limit states compared to the structure with viscous dampers. It is also observed that both the initial and life cycle costs of the DCS method required for the seismic retrofit is smaller than those of the structure retrofitted with viscous dampers. Acknowledgment: This research was supported by a grant (17CTAP-C132889-01) from Technology Advancement Research Program (TARP) funded by Ministry of Land, Infrastructure, and Transport of Korean government.

Keywords: damped cable system, seismic retrofit, self centering, fragility analysis

Procedia PDF Downloads 448
16007 Evaluating Value of Users' Personal Information Based on Cost-Benefit Analysis

Authors: Jae Hyun Park, Sangmi Chai, Minkyun Kim

Abstract:

As users spend more time on the Internet, the probability of their personal information being exposed has been growing. This research has a main purpose of investigating factors and examining relationships when Internet users recognize their value of private information with a perspective of an economic asset. The study is targeted on Internet users, and the value of their private information will be converted into economic figures. Moreover, how economic value changes in relation with individual attributes, dealer’s traits, circumstantial properties will be studied. In this research, the changes in factors on private information value responding to different situations will be analyzed in an economic perspective. Additionally, this study examines the associations between users’ perceived risk and value of their personal information. By using the cost-benefit analysis framework, the hypothesis that the user’s sense in private information value can be influenced by individual attributes and situational properties will be tested. Therefore, this research will attempt to provide answers for three research objectives. First, this research will identify factors that affect value recognition of users’ personal information. Second, it provides evidences that there are differences on information system users’ economic value of information responding to personal, trade opponent, and situational attributes. Third, it investigates the impact of those attributes on individuals’ perceived risk. Based on the assumption that personal, trade opponent and situation attributes make an impact on the users’ value recognition on private information, this research will present the understandings on the different impacts of those attributes in recognizing the value of information with the economic perspective and prove the associative relationships between perceived risk and decision on the value of users’ personal information. In order to validate our research model, this research used the regression methodology. Our research results support that information breach experience and information security systems is associated with users’ perceived risk. Information control and uncertainty are also related to users’ perceived risk. Therefore, users’ perceived risk is considered as a significant factor on evaluating the value of personal information. It can be differentiated by trade opponent and situational attributes. This research presents new perspective on evaluating the value of users’ personal information in the context of perceived risk, personal, trade opponent and situational attributes. It fills the gap in the literature by providing how users’ perceived risk are associated with personal, trade opponent and situation attitudes in conducting business transactions with providing personal information. It adds to previous literature that the relationship exists between perceived risk and the value of users’ private information in the economic perspective. It also provides meaningful insights to the managers that in order to minimize the cost of information breach, managers need to recognize the value of individuals’ personal information and decide the proper amount of investments on protecting users’ online information privacy.

Keywords: private information, value, users, perceived risk, online information privacy, attributes

Procedia PDF Downloads 232
16006 Failure Analysis and Verification Using an Integrated Method for Automotive Electric/Electronic Systems

Authors: Lei Chen, Jian Jiao, Tingdi Zhao

Abstract:

Failures of automotive electric/electronic systems, which are universally considered to be safety-critical and software-intensive, may cause catastrophic accidents. Analysis and verification of failures in these kinds of systems is a big challenge with increasing system complexity. Model-checking is often employed to allow formal verification by ensuring that the system model conforms to specified safety properties. The system-level effects of failures are established, and the effects on system behavior are observed through the formal verification. A hazard analysis technique, called Systems-Theoretic Process Analysis, is capable of identifying design flaws which may cause potential failure hazardous, including software and system design errors and unsafe interactions among multiple system components. This paper provides a concept on how to use model-checking integrated with Systems-Theoretic Process Analysis to perform failure analysis and verification of automotive electric/electronic systems. As a result, safety requirements are optimized, and failure propagation paths are found. Finally, an automotive electric/electronic system case study is used to verify the effectiveness and practicability of the method.

Keywords: failure analysis and verification, model checking, system-theoretic process analysis, automotive electric/electronic system

Procedia PDF Downloads 114
16005 Measuring Business Strategy and Information Systems Alignment

Authors: Amit Saraswat, Ruchi Tewari

Abstract:

Purpose: The research paper aims at understanding the alignment of business and IT in the Indian context and the business value attached to such an alignment. Methodology: The study is conducted in two stages. Stage one: Bibliographic research was conducted to evolve the parameters for defining alignment. Stage two: Evolving a model for strategic alignment to conduct an empirical study. The model is defined in terms of four fundamental domains of strategic management choice – business strategy, information strategy, organizational structure, and information technology structure. A survey through a questionnaire was conducted across organizations from 4 different industries and Structure Equation Modelling (SEM) technique is used for validating the model. Findings: In the Indian scenario all the subscales of alignment could not be validated. It could be validated that organizational strategy impacts information strategy and information technology structure. Research Limitations: The study is limited to the Indian context. Business IT alignment may be culture dependent so further research is required to validate the model in other cultures. Originality/Value: In the western world several models of alignment of business strategy and information systems is available but they do not measure the extent of alignment which the current study in the Indian context. Findings of the study can be used by managers in strategizing and understanding their business and information systems needs holistically and cohesively leading to efficient use of resources and output.

Keywords: business strategy, information technology (IT), business IT alignment, SEM

Procedia PDF Downloads 382
16004 A Mathematical Model for Reliability Redundancy Optimization Problem of K-Out-Of-N: G System

Authors: Gak-Gyu Kim, Won Il Jung

Abstract:

According to a remarkable development of science and technology, function and role of the system of engineering fields has recently been diversified. The system has become increasingly more complex and precise, and thus, system designers intended to maximize reliability concentrate more effort at the design stage. This study deals with the reliability redundancy optimization problem (RROP) for k-out-of-n: G system configuration with cold standby and warm standby components. This paper further intends to present the optimal mathematical model through which the following three elements of (i) multiple components choices, (ii) redundant components quantity and (iii) the choice of redundancy strategies may be combined in order to maximize the reliability of the system. Therefore, we focus on the following three issues. First, we consider RROP that there exists warm standby state as well as cold standby state of the component. Second, as eliminating an approximation approach of the previous RROP studies, we construct a precise model for system reliability. Third, given transition time when the state of components changes, we present not simply a workable solution but the advanced method. For the wide applicability of RROPs, moreover, we use absorbing continuous time Markov chain and matrix analytic methods in the suggested mathematical model.

Keywords: RROP, matrix analytic methods, k-out-of-n: G system, MTTF, absorbing continuous time Markov Chain

Procedia PDF Downloads 252
16003 A Damage-Plasticity Concrete Model for Damage Modeling of Reinforced Concrete Structures

Authors: Thanh N. Do

Abstract:

This paper addresses the modeling of two critical behaviors of concrete material in reinforced concrete components: (1) the increase in strength and ductility due to confining stresses from surrounding transverse steel reinforcements, and (2) the progressive deterioration in strength and stiffness due to high strain and/or cyclic loading. To improve the state-of-the-art, the author presents a new 3D constitutive model of concrete material based on plasticity and continuum damage mechanics theory to simulate both the confinement effect and the strength deterioration in reinforced concrete components. The model defines a yield function of the stress invariants and a compressive damage threshold based on the level of confining stresses to automatically capture the increase in strength and ductility when subjected to high compressive stresses. The model introduces two damage variables to describe the strength and stiffness deterioration under tensile and compressive stress states. The damage formulation characterizes well the degrading behavior of concrete material, including the nonsymmetric strength softening in tension and compression, as well as the progressive strength and stiffness degradation under primary and follower load cycles. The proposed damage model is implemented in a general purpose finite element analysis program allowing an extensive set of numerical simulations to assess its ability to capture the confinement effect and the degradation of the load-carrying capacity and stiffness of structural elements. It is validated against a collection of experimental data of the hysteretic behavior of reinforced concrete columns and shear walls under different load histories. These correlation studies demonstrate the ability of the model to describe vastly different hysteretic behaviors with a relatively consistent set of parameters. The model shows excellent consistency in response determination with very good accuracy. Its numerical robustness and computational efficiency are also very good and will be further assessed with large-scale simulations of structural systems.

Keywords: concrete, damage-plasticity, shear wall, confinement

Procedia PDF Downloads 166