Search results for: real time kernel preemption
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20469

Search results for: real time kernel preemption

14349 Anomaly Detection with ANN and SVM for Telemedicine Networks

Authors: Edward Guillén, Jeisson Sánchez, Carlos Omar Ramos

Abstract:

In recent years, a wide variety of applications are developed with Support Vector Machines -SVM- methods and Artificial Neural Networks -ANN-. In general, these methods depend on intrusion knowledge databases such as KDD99, ISCX, and CAIDA among others. New classes of detectors are generated by machine learning techniques, trained and tested over network databases. Thereafter, detectors are employed to detect anomalies in network communication scenarios according to user’s connections behavior. The first detector based on training dataset is deployed in different real-world networks with mobile and non-mobile devices to analyze the performance and accuracy over static detection. The vulnerabilities are based on previous work in telemedicine apps that were developed on the research group. This paper presents the differences on detections results between some network scenarios by applying traditional detectors deployed with artificial neural networks and support vector machines.

Keywords: anomaly detection, back-propagation neural networks, network intrusion detection systems, support vector machines

Procedia PDF Downloads 342
14348 The Effects of the Interaction between Prenatal Stress and Diet on Maternal Insulin Resistance and Inflammatory Profile

Authors: Karen L. Lindsay, Sonja Entringer, Claudia Buss, Pathik D. Wadhwa

Abstract:

Maternal nutrition and stress are independently recognized as among the most important factors that influence prenatal biology, with implications for fetal development and poor pregnancy outcomes. While there is substantial evidence from non-pregnancy human and animal studies that a complex, bi-directional relationship exists between nutrition and stress, to the author’s best knowledge, their interaction in the context of pregnancy has been significantly understudied. The aim of this study is to assess the interaction between maternal psychological stress and diet quality across pregnancy and its effects on biomarkers of prenatal insulin resistance and inflammation. This is a prospective longitudinal study of N=235 women carrying a healthy, singleton pregnancy, recruited from prenatal clinics of the University of California, Irvine Medical Center. Participants completed a 4-day ambulatory assessment in early, middle and late pregnancy, which included multiple daily electronic diary entries using Ecological Momentary Assessment (EMA) technology on a dedicated study smartphone. The EMA diaries gathered moment-level data on maternal perceived stress, negative mood, positive mood and quality of social interactions. The numerical scores for these variables were averaged across each study time-point and converted to Z-scores. A single composite variable for 'STRESS' was computed as follows: (Negative mood+Perceived stress)–(Positive mood+Social interaction quality). Dietary intakes were assessed by three 24-hour dietary recalls conducted within two weeks of each 4-day assessment. Daily nutrient and food group intakes were averaged across each study time-point. The Alternative Healthy Eating Index adapted for pregnancy (AHEI-P) was computed for early, middle and late pregnancy as a validated summary measure of diet quality. At the end of each 4-day ambulatory assessment, women provided a fasting blood sample, which was assayed for levels of glucose, insulin, Interleukin (IL)-6 and Tumor Necrosis Factor (TNF)-α. Homeostasis Model Assessment of Insulin Resistance (HOMA-IR) was computed. Pearson’s correlation was used to explore the relationship between maternal STRESS and AHEI-P within and between each study time-point. Linear regression was employed to test the association of the stress-diet interaction (STRESS*AHEI-P) with the biological markers HOMA-IR, IL-6 and TNF-α at each study time-point, adjusting for key covariates (pre-pregnancy body mass index, maternal education level, race/ethnicity). Maternal STRESS and AHEI-P were significantly inversely correlated in early (r=-0.164, p=0.018) and mid-pregnancy (-0.160, p=0.019), and AHEI-P from earlier gestational time-points correlated with later STRESS (early AHEI-P x mid STRESS: r=-0.168, p=0.017; mid AHEI-P x late STRESS: r=-0.142, p=0.041). In regression models, the interaction term was not associated with HOMA-IR or IL-6 at any gestational time-point. The stress-diet interaction term was significantly associated with TNF-α according to the following patterns: early AHEI-P*early STRESS vs early TNF-α (p=0.005); early AHEI-P*early STRESS vs mid TNF-α (p=0.002); early AHEI-P*mid STRESS vs mid TNF-α (p=0.005); mid AHEI-P*mid STRESS vs mid TNF-α (p=0.070); mid AHEI-P*late STRESS vs late TNF-α (p=0.011). Poor diet quality is significantly related to higher psychosocial stress levels in pregnant women across gestation, which may promote inflammation via TNF-α. Future prenatal studies should consider the combined effects of maternal stress and diet when evaluating either one of these factors on pregnancy or infant outcomes.

Keywords: diet quality, inflammation, insulin resistance, nutrition, pregnancy, stress, tumor necrosis factor-alpha

Procedia PDF Downloads 187
14347 Investigation of Ignition Delay for Low Molecular Hydrocarbon Fuel and Oxygen Mixture behind the Reflected Shock

Authors: K. R. Guna, Aldin Justin Sundararaj, B. C. Pillai, A. N. Subash

Abstract:

A systematic study has been made for ignition delay times measurement behind a reflected shock wave for the low molecular weight hydrocarbon fuel in argon simulated gas mixtures. The low molecular hydrocarbon fuel–oxygen was diluted with argon for desired concentration is taken for the study. The suitability of the shock tube for measuring the ignition delay time is demonstrated by measuring the ignition delay for the liquefied petroleum gas for equivalence ratios (ф=0.5 & 1) in the temperature range 1150-1650 K. The pressure range was fixed from 5-15 bar. The ignition delay was measured by recording the ignition-induced pressure jump and emission from CH radical simultaneously. From conducting experiments, it was found that the ignition delay time for liquefied petroleum gas reduces with increase in temperature. The shock tube was calibrated for ethane-oxygen gas mixture and the results obtained from this study is compared with the earlier reported values and found to be comparably well suited for the measurement of ignition delay times. The above work was carried out using the shock tube facility at propulsion and high enthalpy laboratory, Karunya University.

Keywords: ignition delay, LPG, reflected shock, shock wave

Procedia PDF Downloads 240
14346 The Influence of Clayey Pellet Size on Adsorption Efficiency of Metal Ions Removal from Waste Printing Developer

Authors: Kiurski S. Jelena, Ranogajec G. Jonjaua, Oros B. Ivana, Kecić S. Vesna

Abstract:

The adsorption efficiency of fired clayey pellets of 5 and 8 mm diameter size for Cu(II) and Zn(II) ions removal from a waste printing developer was studied. In order to investigate the influence of contact time, adsorbent mass and pellet size on the adsorption efficiency the batch mode was carried out. Faster uptake of copper ions was obtained with the fired clay pellets of 5 mm diameter size within 30 minutes. The pellets of 8 mm diameter size showed the higher equilibrium time (60 to 75 minutes) for copper and zinc ions. The results pointed out that adsorption efficiency increases with the increase of adsorbent mass. The maximal efficiency is different for Cu(II) and Zn(II) ions due to the pellet size. Therefore, the fired clay pellets of 5 mm diameter size present an effective adsorbent for Cu(II) ions removal (adsorption efficiency is 63.6%), whereas the fired clay pellets of 8 mm diameter size are the best alternative for Zn(II) ions removal (adsorption efficiency is 92.8%) from a waste printing developer.

Keywords: adsorption efficiency, clayey pellet, metal ions, waste printing developer

Procedia PDF Downloads 284
14345 A Nonlinear Dynamical System with Application

Authors: Abdullah Eqal Al Mazrooei

Abstract:

In this paper, a nonlinear dynamical system is presented. This system is a bilinear class. The bilinear systems are very important kind of nonlinear systems because they have many applications in real life. They are used in biology, chemistry, manufacturing, engineering, and economics where linear models are ineffective or inadequate. They have also been recently used to analyze and forecast weather conditions. Bilinear systems have three advantages: First, they define many problems which have a great applied importance. Second, they give us approximations to nonlinear systems. Thirdly, they have a rich geometric and algebraic structures, which promises to be a fruitful field of research for scientists and applications. The type of nonlinearity that is treated and analyzed consists of bilinear interaction between the states vectors and the system input. By using some properties of the tensor product, these systems can be transformed to linear systems. But, here we discuss the nonlinearity when the state vector is multiplied by itself. So, this model will be able to handle evolutions according to the Lotka-Volterra models or the Lorenz weather models, thus enabling a wider and more flexible application of such models. Here we apply by using an estimator to estimate temperatures. The results prove the efficiency of the proposed system.

Keywords: Lorenz models, nonlinear systems, nonlinear estimator, state-space model

Procedia PDF Downloads 243
14344 MLOps Scaling Machine Learning Lifecycle in an Industrial Setting

Authors: Yizhen Zhao, Adam S. Z. Belloum, Goncalo Maia Da Costa, Zhiming Zhao

Abstract:

Machine learning has evolved from an area of academic research to a real-word applied field. This change comes with challenges, gaps and differences exist between common practices in academic environments and the ones in production environments. Following continuous integration, development and delivery practices in software engineering, similar trends have happened in machine learning (ML) systems, called MLOps. In this paper we propose a framework that helps to streamline and introduce best practices that facilitate the ML lifecycle in an industrial setting. This framework can be used as a template that can be customized to implement various machine learning experiment. The proposed framework is modular and can be recomposed to be adapted to various use cases (e.g. data versioning, remote training on cloud). The framework inherits practices from DevOps and introduces other practices that are unique to the machine learning system (e.g.data versioning). Our MLOps practices automate the entire machine learning lifecycle, bridge the gap between development and operation.

Keywords: cloud computing, continuous development, data versioning, DevOps, industrial setting, MLOps

Procedia PDF Downloads 250
14343 A Study on the Shear-Induced Crystallization of Aliphatic-Aromatic Copolyester

Authors: Ramin Hosseinnezhad, Iurii Vozniak, Andrzej Galeski

Abstract:

Shear-induced crystallization, originated from orientation of chains along the flow direction, is an inevitable part of most polymer processing technologies. It plays a dominant role in determining the final product properties and is affected by many factors such as shear rate, cooling rate, total strain, etc. Investigation of the shear-induced crystallization process become of great importance for preparation of nanocomposite, which requires crystallization of nanofibrous sheared inclusions at higher temperatures. Thus, the effects of shear time, shear rate, and also thermal condition of cooling on crystallization of two aliphatic-aromatic copolyesters have been investigated. This was performed using Linkam optical shearing system (CSS450) for both Ecoflex® F Blend C1200 produced by BASF and synthesized copolyester of butylene terephthalate and a mixture of butylene esters: adipate, succinate, and glutarate, (PBASGT), containing 60% of aromatic comonomer. Crystallization kinetics of these biodegradable copolyesters was studied at two different conditions of shearing. First, sample with a thickness of 60µm was heated to 60˚C above its melting point and subsequently subjected to different shear rates (100–800 sec-1) while cooling with specific rates. Second, the same type of sample was cooled down when shearing at constant temperature was finished. The intensity of transmitted depolarized light, recorded by a camera attached to the optical microscope, was used as a measure to follow the crystallization. Temperature dependencies of conversion degree of samples during cooling were collected and used to determine the half-temperature (Th), at which 50% conversion degree was reached. Shearing ecoflex films for 45 seconds with a shear rate of 100 sec-1 resulted in significant increase of Th from 56˚C to 70˚C. Moreover, the temperature range for the transition of molten samples to crystallized state decreased from 42˚C to 20˚C. Comparatively low shift of 10˚C in Th towards higher temperature was observed for PBASGT films at shear rate of 600 sec-1 for 45 seconds. However, insufficient melt flow strength and non-laminar flow due to Taylor vortices was a hindrance to reach more elevated Th at very high shear rates (600–800 sec-1). The shift in Th was smaller for the samples sheared at a constant temperature and subsequently cooled down. This may be attributed to the longer time gap between cessation of shearing and the onset of crystallization. The longer this time gap, the more possibility for crystal nucleus to re-melt at temperatures above Tm and for polymer chains to recoil and relax. It is found that the crystallization temperature, crystallization induction time and spherulite growth of aliphatic-aromatic copolyesters are dramatically influenced by both the cooling rate and the shear imposed during the process.

Keywords: induced crystallization, shear rate, aliphatic-aromatic copolyester, ecoflex

Procedia PDF Downloads 438
14342 The Optimum Operating Conditions for the Synthesis of Zeolite from Waste Incineration Fly Ash by Alkali Fusion and Hydrothermal Methods

Authors: Yi-Jie Lin, Jyh-Cherng Chen

Abstract:

The fly ash of waste incineration processes is usually hazardous and the disposal or reuse of waste incineration fly ash is difficult. In this study, the waste incineration fly ash was converted to useful zeolites by the alkali fusion and hydrothermal synthesis method. The influence of different operating conditions (the ratio of Si/Al, the ratio of hydrolysis liquid to solid, and hydrothermal time) was investigated to seek the optimum operating conditions for the synthesis of zeolite from waste incineration fly ash. The results showed that concentrations of heavy metals in the leachate of Toxicity Characteristic Leaching Procedure (TCLP) were all lower than the regulatory limits except lead. The optimum operating conditions for the synthesis of zeolite from waste incineration fly ash by the alkali fusion and hydrothermal synthesis method were Si/Al=40, NaOH/ash=1.5, alkali fusion at 400 oC for 40 min, hydrolysis with Liquid to Solid ratio (L/S)= 200 at 105 oC for 24 h, and hydrothermal synthesis at 105 oC for 24 h. The specific surface area of fly ash could be significantly increased from 8.59 m2/g to 651.51 m2/g (synthesized zeolite). The influence of different operating conditions on the synthesis of zeolite from waste incineration fly ash followed the sequence of Si/Al ratio > hydrothermal time > hydrolysis L/S ratio. The synthesized zeolites can be reused as good adsorbents to control the air or wastewater pollutants. The purpose of fly ash detoxification, reduction and waste recycling/reuse is achieved successfully.

Keywords: alkali fusion, hydrothermal, fly ash, zeolite

Procedia PDF Downloads 223
14341 Antidiabetic Evaluation of Pig (Sus scrofa) Bile on Alloxan-Induced BALB/c Mice

Authors: John Lyndon C. Lunnay

Abstract:

This study discerns to evaluate the antidiabetic efficacy of pig bile on alloxan-induced BALB/c mice. The experimental animals were divided and selected using RCBD into 5 groups (n= 4): T1 (negative control), T2 (1ml/kg), T3 (2ml/kg), T4 (3ml/kg) and T5 (Glibenclamide). Hyperglycemia was induced by injecting 1% alloxan monohydrate intraperitoneally. A glucose tolerance test was performed using a 2g/kg glucose solution, and blood glucose levels were measured at different time intervals. 14 days of monitoring was also done to ensure effectivity and efficacy of the different treatments. Bodyweight was also determined. Results show that administration of treatments on test animals significantly reverted the blood glucose levels of mice in 60 minutes and 120 minutes using an oral glucose tolerance test. After 14 days of monitoring, normal blood glucose levels were seen significantly on T2 (1ml/kg), T3 (2ml/kg), T4 (3ml/kg), and T5 (Glibenclamide), which only suggests the efficacy of pig bile on lowering glucose levels on alloxan-induced diabetic mice. Bodyweight analysis shows no significant difference. Duncan’s multiple range test (DMRT) shows comparable efficacy and effectivity between T4 (3ml/kg) and T5 (Glibenclamide) on lowering BGL at different day and time intervals.

Keywords: pig bile, BALB/c mice, blood glucose, Gllibenclamide

Procedia PDF Downloads 131
14340 Impacts of Building Design Factors on Auckland School Energy Consumptions

Authors: Bin Su

Abstract:

This study focuses on the impact of school building design factors on winter extra energy consumption which mainly includes space heating, water heating and other appliances related to winter indoor thermal conditions. A number of Auckland schools were randomly selected for the study which introduces a method of using real monthly energy consumption data for a year to calculate winter extra energy data of school buildings. The study seeks to identify the relationships between winter extra energy data related to school building design data related to the main architectural features, building envelope and elements of the sample schools. The relationships can be used to estimate the approximate saving in winter extra energy consumption which would result from a changed design datum for future school development, and identify any major energy-efficient design problems. The relationships are also valuable for developing passive design guides for school energy efficiency.

Keywords: building energy efficiency, building thermal design, building thermal performance, school building design

Procedia PDF Downloads 426
14339 Measurement of Fatty Acid Changes in Post-Mortem Belowground Carcass (Sus-scrofa) Decomposition: A Semi-Quantitative Methodology for Determining the Post-Mortem Interval

Authors: Nada R. Abuknesha, John P. Morgan, Andrew J. Searle

Abstract:

Information regarding post-mortem interval (PMI) in criminal investigations is vital to establish a time frame when reconstructing events. PMI is defined as the time period that has elapsed between the occurrence of death and the discovery of the corpse. Adipocere, commonly referred to as ‘grave-wax’, is formed when post-mortem adipose tissue is converted into a solid material that is heavily comprised of fatty acids. Adipocere is of interest to forensic anthropologists, as its formation is able to slow down the decomposition process. Therefore, analysing the changes in the patterns of fatty acids during the early decomposition process may be able to estimate the period of burial, and hence the PMI. The current study concerned the investigation of the fatty acid composition and patterns in buried pig fat tissue. This was in an attempt to determine whether particular patterns of fatty acid composition can be shown to be associated with the duration of the burial, and hence may be used to estimate PMI. The use of adipose tissue from the abdominal region of domestic pigs (Sus-scrofa), was used to model the human decomposition process. 17 x 20cm piece of pork belly was buried in a shallow artificial grave, and weekly samples (n=3) from the buried pig fat tissue were collected over an 11-week period. Marker fatty acids: palmitic (C16:0), oleic (C18:1n-9) and linoleic (C18:2n-6) acid were extracted from the buried pig fat tissue and analysed as fatty acid methyl esters using the gas chromatography system. Levels of the marker fatty acids were quantified from their respective standards. The concentrations of C16:0 (69.2 mg/mL) and C18:1n-9 (44.3 mg/mL) from time zero exhibited significant fluctuations during the burial period. Levels rose (116 and 60.2 mg/mL, respectively) and fell starting from the second week to reach 19.3 and 18.3 mg/mL, respectively at week 6. Levels showed another increase at week 9 (66.3 and 44.1 mg/mL, respectively) followed by gradual decrease at week 10 (20.4 and 18.5 mg/mL, respectively). A sharp increase was observed in the final week (131.2 and 61.1 mg/mL, respectively). Conversely, the levels of C18:2n-6 remained more or less constant throughout the study. In addition to fluctuations in the concentrations, several new fatty acids appeared in the latter weeks. Other fatty acids which were detectable in the time zero sample, were lost in the latter weeks. There are several probable opportunities to utilise fatty acid analysis as a basic technique for approximating PMI: the quantification of marker fatty acids and the detection of selected fatty acids that either disappear or appear during the burial period. This pilot study indicates that this may be a potential semi-quantitative methodology for determining the PMI. Ideally, the analysis of particular fatty acid patterns in the early stages of decomposition could be an additional tool to the already available techniques or methods in improving the overall processes in estimating PMI of a corpse.

Keywords: adipocere, fatty acids, gas chromatography, post-mortem interval

Procedia PDF Downloads 115
14338 Palliative Orthovoltage Radiotherapy and Subcutaneous Infusion of Carboplatin for Treatment of Appendicular Osteosarcoma in Dogs

Authors: Kathryn L. Duncan, Charles A. Kuntz, Alessandra C. Santamaria, James O. Simcock

Abstract:

Access to megavoltage radiation therapy for small animals is limited in many locations around the world. This can preclude the use of palliative radiation therapy for the treatment of appendicular osteosarcoma in dogs. The objective of this study was to retrospectively assess the adverse effects and survival times of dogs with appendicular osteosarcoma that were treated with hypofractionated orthovoltage radiation therapy and adjunctive carboplatin chemotherapy administered via a single subcutaneous infusion. Medical records were reviewed retrospectively to identify client-owned dogs with spontaneously occurring appendicular osteosarcoma that was treated with palliative orthovoltage radiation therapy and a single subcutaneous infusion of carboplatin. Data recorded included signalment, tumour location, results of diagnostic imaging, haematologic and serum biochemical analyses, adverse effects of radiation therapy and chemotherapy, and survival times. Kaplan-Meier survival analysis was performed, and log-rank analysis was used to determine the impact of specific patient variables on survival time. Twenty-three dogs were identified that met the inclusion criteria. Median survival time for dogs was 182 days. Eleven dogs had adverse haematologic effects, 3 had adverse gastrointestinal effects, 6 had adverse effects at the radiation site and 7 developed infections at the carboplatin infusion site. No statistically significant differences were identified in survival times based on sex, tumour location, development of infection, or pretreatment serum alkaline phosphatase. Median survival time and incidence of adverse effects were comparable to those previously reported in dogs undergoing palliative radiation therapy with megavoltage or cobalt radiation sources and conventional intravenous carboplatin chemotherapy. The use of orthovoltage palliative radiation therapy may be a reasonable alternative to megavoltage radiation in locations where access is limited.

Keywords: radiotherapy, veterinary oncology, chemotherapy, osteosarcoma

Procedia PDF Downloads 59
14337 A Solution for Production Facility Assignment: An Automotive Subcontract Case

Authors: Cihan Çetinkaya, Eren Özceylan, Kerem Elibal

Abstract:

This paper presents a solution method for selection of production facility. The motivation has been taken from a real life case, an automotive subcontractor which has two production facilities at different cities and parts. The problem is to decide which part(s) should be produced at which facility. To the best of our knowledge, until this study, there was no scientific approach about this problem at the firm and decisions were being given intuitively. In this study, some logistic cost parameters have been defined and with these parameters a mathematical model has been constructed. Defined and collected cost parameters are handling cost of parts, shipment cost of parts and shipment cost of welding fixtures. Constructed multi-objective mathematical model aims to minimize these costs while aims to balance the workload between two locations. Results showed that defined model can give optimum solutions in reasonable computing times. Also, this result gave encouragement to develop the model with addition of new logistic cost parameters.

Keywords: automotive subcontract, facility assignment, logistic costs, multi-objective models

Procedia PDF Downloads 356
14336 An Extended X-Ray Absorption Fine Structure Study of CoTi Thin Films

Authors: Jose Alberto Duarte Moller, Cynthia Deisy Gomez Esparza

Abstract:

The cobalt-titanium system was grown as thin films in an INTERCOVAMEX V3 sputtering system, equipped with four magnetrons assisted by DC pulsed and direct DC. A polished highly oriented (400) silicon wafer was used as substrate and the growing temperature was 500 oC. Xray Absorption Spectroscopy experiments were carried out in the SSRL in the 4-3 beam line. The Extenden X-Ray Absorption Fine Structure spectra have been numerically processed by WINXAS software from the background subtraction until the normalization and FFT adjustment. Analyzing the absorption spectra of cobalt in the CoTi2 phase we can appreciate that they agree in energy with the reference spectra that corresponds to the CoO, which indicates that the valence where upon working is Co2+. The RDF experimental results were then compared with those RDF´s generated theoretically by using FEFF software, from a model compound of CoTi2 phase obtained by XRD. The fitting procedure is a highly iterative process. Fits are also checked in R-space using both the real and imaginary parts of Fourier transform. Finally, the presence of overlapping coordination shells and the correctness of the assumption about the nature of the coordinating atom were checked.

Keywords: XAS, EXAFS, FEFF, CoTi

Procedia PDF Downloads 277
14335 The Application of FSI Techniques in Modeling of Realist Pulmonary Systems

Authors: Abdurrahim Bolukbasi, Hassan Athari, Dogan Ciloglu

Abstract:

The modeling lung respiratory system which has complex anatomy and biophysics presents several challenges including tissue-driven flow patterns and wall motion. Also, the lung pulmonary system because of that they stretch and recoil with each breath, has not static walls and structures. The direct relationship between air flow and tissue motion in the lung structures naturally prefers an FSI simulation technique. Therefore, in order to toward the realistic simulation of pulmonary breathing mechanics the development of a coupled FSI computational model is an important step. A simple but physiologically-relevant three dimensional deep long geometry is designed and fluid-structure interaction (FSI) coupling technique is utilized for simulating the deformation of the lung parenchyma tissue which produces airflow fields. The real understanding of respiratory tissue system as a complex phenomenon have been investigated with respect to respiratory patterns, fluid dynamics and tissue visco-elasticity and tidal breathing period.

Keywords: lung deformation and mechanics; Tissue mechanics; Viscoelasticity; Fluid-structure interactions; ANSYS

Procedia PDF Downloads 303
14334 Water Sources in 3 Local Municipalities of O. R. Tambo District Municipality, South Africa: A Comparative Study

Authors: Betek Cecilia Kunseh, Musampa Christopher

Abstract:

Despite significant investment and important progress, access to safe potable water continues to be one of the most pressing challenges for rural communities in O R Tambo District Municipality. This is coupled with the low income of most residents and government's policy which obliges municipalities to supply basic water usually set at 6 kilolitres per month to each household free of charge. During the research, data was collected from three local municipalities of O. R. Tambo, i.e. King Sabata Dalindyebo, Mhlontlo and Ingquza Hill local municipalities. According to the result, significant differences exist between the sources of water in the different local municipalities from which data was collected. The chi square was use to calculated the differences between the sources of water and the calculated critical value of the District Municipality was 18.77 which is more than the stipulated critical value of 3.84. More people in Mhlontlo Local Municipality got water from the taps while a greater percentage of households in King Sataba Dalindyebo and Ingquza hill local municipalities got their water from the natural sources. 77% of the sample population complained that there have been no improvements in water provision because they still get water from natural sources and even the remaining 33% that were getting water from the taps still have to depend on natural sources because the taps are most of the time broken and it takes a long time to fix them.

Keywords: availability, water, sources, supply

Procedia PDF Downloads 324
14333 Evaluating Forecasts Through Stochastic Loss Order

Authors: Wilmer Osvaldo Martinez, Manuel Dario Hernandez, Juan Manuel Julio

Abstract:

We propose to assess the performance of k forecast procedures by exploring the distributions of forecast errors and error losses. We argue that non systematic forecast errors minimize when their distributions are symmetric and unimodal, and that forecast accuracy should be assessed through stochastic loss order rather than expected loss order, which is the way it is customarily performed in previous work. Moreover, since forecast performance evaluation can be understood as a one way analysis of variance, we propose to explore loss distributions under two circumstances; when a strict (but unknown) joint stochastic order exists among the losses of all forecast alternatives, and when such order happens among subsets of alternative procedures. In spite of the fact that loss stochastic order is stronger than loss moment order, our proposals are at least as powerful as competing tests, and are robust to the correlation, autocorrelation and heteroskedasticity settings they consider. In addition, since our proposals do not require samples of the same size, their scope is also wider, and provided that they test the whole loss distribution instead of just loss moments, they can also be used to study forecast distributions as well. We illustrate the usefulness of our proposals by evaluating a set of real world forecasts.

Keywords: forecast evaluation, stochastic order, multiple comparison, non parametric test

Procedia PDF Downloads 77
14332 Violations of Press Freedom

Authors: Khalid Achaat

Abstract:

It is difficult to speak about freedom of the press in Algeria without first talking to fifty-seven journalists killed in the country between 1993 and 1997 and the five missing journalists. No serious investigation was conducted to find the culprits. When a State is not able to guarantee law, there is no justice and violations of the law become "systematic". How to claim the freedom of press in Algeria, when death becomes "banal"? In these circumstances, can we talk of rights of the Algerian press? It is impossible to understand the problems of the press in Algeria, focusing solely legal issues. Take into account technical, financial and political. Their respective roles varies depending on whether one focuses on the collection of information, the regime of the newspaper company or publication and dissemination. Can we say that the Algerian press is "the freest in the Arab world", while the latter reflects only partially the real problems facing the country? While any newspaper company is subject, de facto, to an authorization scheme, permanently subjected to the constant threat of withdrawal of the authorization, suspension, prohibition or closure without it has the right to a remedy? Can it be free when the majority of "media owners", head of the largest daily newspapers are derived from the single party in power since independence? Some of this release does not it serves the interests of the Algerian power?

Keywords: freedom, press, power, closure, suspension

Procedia PDF Downloads 340
14331 Non-Standard Monetary Policy Measures and Their Consequences

Authors: Aleksandra Nocoń (Szunke)

Abstract:

The study is a review of the literature concerning the consequences of non-standard monetary policy, which are used by central banks during unconventional periods, threatening instability of the banking sector. In particular, the attention was paid to the effects of non-standard monetary policy tools for financial markets. However, the empirical evidence about their effects and real consequences for the financial markets are still not final. The main aim of the study is to survey the consequences of standard and non-standard monetary policy instruments, implemented during the global financial crisis in the United States, United Kingdom and Euroland, with particular attention to the results for the stabilization of global financial markets. The study analyses the consequences for short and long-term market interest rates, interbank interest rates and LIBOR-OIS spread. The study consists mainly of the empirical review, indicating the impact of the implementation of these tools for the financial markets. The following research methods were used in the study: literature studies, including domestic and foreign literature, cause and effect analysis and statistical analysis.

Keywords: asset purchase facility, consequences of monetary policy instruments, non-standard monetary policy, quantitative easing

Procedia PDF Downloads 315
14330 Thermal Ageing of a 316 Nb Stainless Steel: From Mechanical and Microstructural Analyses to Thermal Ageing Models for Long Time Prediction

Authors: Julien Monnier, Isabelle Mouton, Francois Buy, Adrien Michel, Sylvain Ringeval, Joel Malaplate, Caroline Toffolon, Bernard Marini, Audrey Lechartier

Abstract:

Chosen to design and assemble massive components for nuclear industry, the 316 Nb austenitic stainless steel (also called 316 Nb) suits well this function thanks to its mechanical, heat and corrosion handling properties. However, these properties might change during steel’s life due to thermal ageing causing changes within its microstructure. Our main purpose is to determine if the 316 Nb will keep its mechanical properties after an exposition to industrial temperatures (around 300 °C) during a long period of time (< 10 years). The 316 Nb is composed by different phases, which are austenite as main phase, niobium-carbides, and ferrite remaining from the ferrite to austenite transformation during the process. Our purpose is to understand thermal ageing effects on the material microstructure and properties and to submit a model predicting the evolution of 316 Nb properties as a function of temperature and time. To do so, based on Fe-Cr and 316 Nb phase diagrams, we studied the thermal ageing of 316 Nb steel alloys (1%v of ferrite) and welds (10%v of ferrite) for various temperatures (350, 400, and 450 °C) and ageing time (from 1 to 10.000 hours). Higher temperatures have been chosen to reduce thermal treatment time by exploiting a kinetic effect of temperature on 316 Nb ageing without modifying reaction mechanisms. Our results from early times of ageing show no effect on steel’s global properties linked to austenite stability, but an increase of ferrite hardness during thermal ageing has been observed. It has been shown that austenite’s crystalline structure (cfc) grants it a thermal stability, however, ferrite crystalline structure (bcc) favours iron-chromium demixion and formation of iron-rich and chromium-rich phases within ferrite. Observations of thermal ageing effects on ferrite’s microstructure were necessary to understand the changes caused by the thermal treatment. Analyses have been performed by using different techniques like Atomic Probe Tomography (APT) and Differential Scanning Calorimetry (DSC). A demixion of alloy’s elements leading to formation of iron-rich (α phase, bcc structure), chromium-rich (α’ phase, bcc structure), and nickel-rich (fcc structure) phases within the ferrite have been observed and associated to the increase of ferrite’s hardness. APT results grant information about phases’ volume fraction and composition, allowing to associate hardness measurements to the volume fractions of the different phases and to set up a way to calculate α’ and nickel-rich particles’ growth rate depending on temperature. The same methodology has been applied to DSC results, which allowed us to measure the enthalpy of α’ phase dissolution between 500 and 600_°C. To resume, we started from mechanical and macroscopic measurements and explained the results through microstructural study. The data obtained has been match to CALPHAD models’ prediction and used to improve these calculations and employ them to predict 316 Nb properties’ change during the industrial process.

Keywords: stainless steel characterization, atom probe tomography APT, vickers hardness, differential scanning calorimetry DSC, thermal ageing

Procedia PDF Downloads 82
14329 Inverse Heat Conduction Analysis of Cooling on Run-Out Tables

Authors: M. S. Gadala, Khaled Ahmed, Elasadig Mahdi

Abstract:

In this paper, we introduced a gradient-based inverse solver to obtain the missing boundary conditions based on the readings of internal thermocouples. The results show that the method is very sensitive to measurement errors, and becomes unstable when small time steps are used. The artificial neural networks are shown to be capable of capturing the whole thermal history on the run-out table, but are not very effective in restoring the detailed behavior of the boundary conditions. Also, they behave poorly in nonlinear cases and where the boundary condition profile is different. GA and PSO are more effective in finding a detailed representation of the time-varying boundary conditions, as well as in nonlinear cases. However, their convergence takes longer. A variation of the basic PSO, called CRPSO, showed the best performance among the three versions. Also, PSO proved to be effective in handling noisy data, especially when its performance parameters were tuned. An increase in the self-confidence parameter was also found to be effective, as it increased the global search capabilities of the algorithm. RPSO was the most effective variation in dealing with noise, closely followed by CRPSO. The latter variation is recommended for inverse heat conduction problems, as it combines the efficiency and effectiveness required by these problems.

Keywords: inverse analysis, function specification, neural net works, particle swarm, run-out table

Procedia PDF Downloads 226
14328 Beggar-Thy-Neighbor's Beach: Pricing Adaptation to Sea-Level Rise

Authors: Arlan Zandro Brucal, John Lynham

Abstract:

With the accelerated sea-level rise (SLR) increasingly becoming a concern, demand for coastal management and protection is expected to grow. Among the coastal management and protection methods, building seawalls are among the most controversial due to the negative externalities they impose on beachgoers and neighboring properties. This paper provides estimates of the external cost associated with building seawalls on the island of Oahu in Hawaii. Using hedonic pricing approach on real properties sold between 1980-2010 and aerial photographs of seawalls in 1995, the paper finds that (1) while seawalls do increase the value of protected properties, the share of armored properties appear to be negatively correlated with property sale prices, suggesting that the positive effect of seawalls tend to decline as more and more rely on this coastal management method; and (2) the value of beachfront properties tend to decline as they get approach seawalls. Results suggest that policymakers should devise a policy that would internalize the externalities associated with private-sector adaptation to climate change.

Keywords: private sector climate change adaptation, externalities, sea-level rise, hedonic pricing

Procedia PDF Downloads 276
14327 Compression-Extrusion Test to Assess Texture of Thickened Liquids for Dysphagia

Authors: Jesus Salmeron, Carmen De Vega, Maria Soledad Vicente, Mireia Olabarria, Olaia Martinez

Abstract:

Dysphagia or difficulty in swallowing affects mostly elder people: 56-78% of the institutionalized and 44% of the hospitalized. Liquid food thickening is a necessary measure in this situation because it reduces the risk of penetration-aspiration. Until now, and as proposed by the American Dietetic Association in 2002, possible consistencies have been categorized in three groups attending to their viscosity: nectar (50-350 mPa•s), honey (350-1750 mPa•s) and pudding (>1750 mPa•s). The adequate viscosity level should be identified for every patient, according to her/his impairment. Nevertheless, a systematic review on dysphagia diet performed recently indicated that there is no evidence to suggest that there is any transition of clinical relevance between the three levels proposed. It was also stated that other physical properties of the bolus (slipperiness, density or cohesiveness, among others) could influence swallowing in affected patients and could contribute to the amount of remaining residue. Texture parameters need to be evaluated as possible alternative to viscosity. The aim of this study was to evaluate the instrumental extrusion-compression test as a possible tool to characterize changes along time in water thickened with various products and in the three theoretical consistencies. Six commercial thickeners were used: NM® (NM), Multi-thick® (M), Nutilis Powder® (Nut), Resource® (R), Thick&Easy® (TE) and Vegenat® (V). All of them with a modified starch base. Only one of them, Nut, also had a 6,4% of gum (guar, tara and xanthan). They were prepared as indicated in the instructions of each product and dispensing the correspondent amount for nectar, honey and pudding consistencies in 300 mL of tap water at 18ºC-20ºC. The mixture was stirred for about 30 s. Once it was homogeneously spread, it was dispensed in 30 mL plastic glasses; always to the same height. Each of these glasses was used as a measuring point. Viscosity was measured using a rotational viscometer (ST-2001, Selecta, Barcelona). Extrusion-compression test was performed using a TA.XT2i texture analyzer (Stable Micro Systems, UK) with a 25 mm diameter cylindrical probe (SMSP/25). Penetration distance was set at 10 mm and a speed of 3 mm/s. Measurements were made at 1, 5, 10, 20, 30, 40, 50 and 60 minutes from the moment samples were mixed. From the force (g)–time (s) curves obtained in the instrumental assays, maximum force peak (F) was chosen a reference parameter. Viscosity (mPa•s) and F (g) showed to be highly correlated and had similar development along time, following time-dependent quadratic models. It was possible to predict viscosity using F as an independent variable, as they were linearly correlated. In conclusion, compression-extrusion test could be an alternative and a useful tool to assess physical characteristics of thickened liquids.

Keywords: compression-extrusion test, dysphagia, texture analyzer, thickener

Procedia PDF Downloads 353
14326 Investigations on the Application of Avalanche Simulations: A Survey Conducted among Avalanche Experts

Authors: Korbinian Schmidtner, Rudolf Sailer, Perry Bartelt, Wolfgang Fellin, Jan-Thomas Fischer, Matthias Granig

Abstract:

This study focuses on the evaluation of snow avalanche simulations, based on a survey that has been carried out among avalanche experts. In the last decades, the application of avalanche simulation tools has gained recognition within the realm of hazard management. Traditionally, avalanche runout models were used to predict extreme avalanche runout and prepare avalanche maps. This has changed rather dramatically with the application of numerical models. For safety regulations such as road safety simulation tools are now being coupled with real-time meteorological measurements to predict frequent avalanche hazard. That places new demands on model accuracy and requires the simulation of physical processes that previously could be ignored. These simulation tools are based on a deterministic description of the avalanche movement allowing to predict certain quantities (e.g. pressure, velocities, flow heights, runout lengths etc.) of the avalanche flow. Because of the highly variable regimes of the flowing snow, no uniform rheological law describing the motion of an avalanche is known. Therefore, analogies to fluid dynamical laws of other materials are stated. To transfer these constitutional laws to snow flows, certain assumptions and adjustments have to be imposed. Besides these limitations, there exist high uncertainties regarding the initial and boundary conditions. Further challenges arise when implementing the underlying flow model equations into an algorithm executable by a computer. This implementation is constrained by the choice of adequate numerical methods and their computational feasibility. Hence, the model development is compelled to introduce further simplifications and the related uncertainties. In the light of these issues many questions arise on avalanche simulations, on their assets and drawbacks, on potentials for improvements as well as their application in practice. To address these questions a survey among experts in the field of avalanche science (e.g. researchers, practitioners, engineers) from various countries has been conducted. In the questionnaire, special attention is drawn on the expert’s opinion regarding the influence of certain variables on the simulation result, their uncertainty and the reliability of the results. Furthermore, it was tested to which degree a simulation result influences the decision making for a hazard assessment. A discrepancy could be found between a large uncertainty of the simulation input parameters as compared to a relatively high reliability of the results. This contradiction can be explained taking into account how the experts employ the simulations. The credibility of the simulations is the result of a rather thoroughly simulation study, where different assumptions are tested, comparing the results of different flow models along with the use of supplemental data such as chronicles, field observation, silent witnesses i.a. which are regarded as essential for the hazard assessment and for sanctioning simulation results. As the importance of avalanche simulations grows within the hazard management along with their further development studies focusing on the modeling fashion could contribute to a better understanding how knowledge of the avalanche process can be gained by running simulations.

Keywords: expert interview, hazard management, modeling, simulation, snow avalanche

Procedia PDF Downloads 313
14325 Minimizing the Impact of Covariate Detection Limit in Logistic Regression

Authors: Shahadut Hossain, Jacek Wesolowski, Zahirul Hoque

Abstract:

In many epidemiological and environmental studies covariate measurements are subject to the detection limit. In most applications, covariate measurements are usually truncated from below which is known as left-truncation. Because the measuring device, which we use to measure the covariate, fails to detect values falling below the certain threshold. In regression analyses, it causes inflated bias and inaccurate mean squared error (MSE) to the estimators. This paper suggests a response-based regression calibration method to correct the deleterious impact introduced by the covariate detection limit in the estimators of the parameters of simple logistic regression model. Compared to the maximum likelihood method, the proposed method is computationally simpler, and hence easier to implement. It is robust to the violation of distributional assumption about the covariate of interest. In producing correct inference, the performance of the proposed method compared to the other competing methods has been investigated through extensive simulations. A real-life application of the method is also shown using data from a population-based case-control study of non-Hodgkin lymphoma.

Keywords: environmental exposure, detection limit, left truncation, bias, ad-hoc substitution

Procedia PDF Downloads 222
14324 Analytical Terahertz Characterization of In0.53Ga0.47As Transistors and Homogenous Diodes

Authors: Abdelmadjid Mammeri, Fatima Zohra Mahi, Luca Varani, H. Marinchoi

Abstract:

We propose an analytical model for the admittance and the noise calculations of the InGaAs transistor and diode. The development of the small-signal admittance takes into account the longitudinal and transverse electric fields through a pseudo two-dimensional approximation of the Poisson equation. The frequency-dependent of the small-signal admittance response is determined by the total currents and the potentials matrix relation between the gate and the drain terminals. The noise is evaluated by using the real part of the transistor/diode admittance under a small-signal perturbation. The analytical results show that the admittance spectrum exhibits a series of resonant peaks corresponding to the excitation of plasma waves. The appearance of the resonance is discussed and analyzed as functions of the channel length and the temperature. The model can be used, on one hand; to control the appearance of the plasma resonances, and on other hand; can give significant information about the noise frequency dependence in the InGaAs transistor and diode.

Keywords: InGaAs transistors, InGaAs diode, admittance, resonant peaks, plasma waves, analytical model

Procedia PDF Downloads 294
14323 Optimizing the Passenger Throughput at an Airport Security Checkpoint

Authors: Kun Li, Yuzheng Liu, Xiuqi Fan

Abstract:

High-security standard and high efficiency of screening seem to be contradictory to each other in the airport security check process. Improving the efficiency as far as possible while maintaining the same security standard is significantly meaningful. This paper utilizes the knowledge of Operation Research and Stochastic Process to establish mathematical models to explore this problem. We analyze the current process of airport security check and use the M/G/1 and M/G/k models in queuing theory to describe the process. Then we find the least efficient part is the pre-check lane, the bottleneck of the queuing system. To improve passenger throughput and reduce the variance of passengers’ waiting time, we adjust our models and use Monte Carlo method, then put forward three modifications: adjust the ratio of Pre-Check lane to regular lane flexibly, determine the optimal number of security check screening lines based on cost analysis and adjust the distribution of arrival and service time based on Monte Carlo simulation results. We also analyze the impact of cultural differences as the sensitivity analysis. Finally, we give the recommendations for the current process of airport security check process.

Keywords: queue theory, security check, stochatic process, Monte Carlo simulation

Procedia PDF Downloads 191
14322 Impact of Endogenous Risk Factors on Risk Cost in KSA PPP Projects

Authors: Saleh Alzahrani, Halim Boussabaine

Abstract:

The Public Private Partnership (PPP) contracts are produced taking into account the reason that the configuration, development, operation, and financing of an open undertaking is to be recompensed to a private gathering inside a solitary contractual structure. PPP venture dangers are ordinarily connected with the improvement and development of another resource and in addition its operation for a considerable length of time. Without a doubt, the most genuine outcomes of dangers amid the development period are value and time overwhelms. These occasions are amongst the most extensively utilized situations as a part of worth for cash investigation dangers. The wellsprings of danger change over the life cycle of a PPP venture. In customary acquirement, the general population segment ordinarily needs to cover all value trouble from these dangers. At any rate there is bounty confirmation to recommend that cost pain is a standard in a percentage of the tasks that are conveyed under customary obtainment. This paper means to research the effect of endogenous dangers on expense overwhelm in KSA PPP ventures. The paper displays a brief writing survey on PPP danger evaluating systems, and after that presents an affiliation model between danger occasions and expense invade in KSA. The paper finishes up with considerations for future examination.

Keywords: PPP, risk pricing, impact of risk, Endogenous risks

Procedia PDF Downloads 439
14321 Literature Review on the Controversies and Changes in the Insanity Defense since the Wild Beast Standard in 1723 until the Federal Insanity Defense Reform Act of 1984

Authors: Jane E. Hill

Abstract:

Many variables led to the changes in the insanity defense since the Wild Beast Standard of 1723 until the Federal Insanity Defense Reform Act of 1984. The insanity defense is used in criminal trials and argued that the defendant is ‘not guilty by reason of insanity’ because the individual was unable to distinguish right from wrong during the time they were breaking the law. The issue that surrounds whether or not to use the insanity defense in the criminal court depends on the mental state of the defendant at the time the criminal act was committed. This leads us to the question of did the defendant know right from wrong when they broke the law? In 1723, The Wild Beast Test stated that to be exempted from punishment the individual is totally deprived of their understanding and memory and doth not know what they are doing. The Wild Beast Test became the standard in England for over seventy-five years. In 1800, James Hadfield attempted to assassinate King George III. He only made the attempt because he was having delusional beliefs. The jury and the judge gave a verdict of not guilty. However, to legal confine him; the Criminal Lunatics Act was enacted. Individuals that were deemed as ‘criminal lunatics’ and were given a verdict of not guilty would be taken into custody and not be freed into society. In 1843, the M'Naghten test required that the individual did not know the quality or the wrongfulness of the offense at the time they committed the criminal act(s). Daniel M'Naghten was acquitted on grounds of insanity. The M'Naghten Test is still a modern concept of the insanity defense used in many courts today. The Irresistible Impulse Test was enacted in the United States in 1887. The Irresistible Impulse Test suggested that offenders that could not control their behavior while they were committing a criminal act were not deterrable by the criminal sanctions in place; therefore no purpose would be served by convicting the offender. Due to the criticisms of the latter two contentions, the federal District of Columbia Court of Appeals ruled in 1954 to adopt the ‘product test’ by Sir Isaac Ray for insanity. The Durham Rule also known as the ‘product test’, stated an individual is not criminally responsible if the unlawful act was the product of mental disease or defect. Therefore, the two questions that need to be asked and answered are (1) did the individual have a mental disease or defect at the time they broke the law? and (2) was the criminal act the product of their disease or defect? The Durham courts failed to clearly define ‘mental disease’ or ‘product.’ Therefore, trial courts had difficulty defining the meaning of the terms and the controversy continued until 1972 when the Durham rule was overturned in most places. Therefore, the American Law Institute combined the M'Naghten test with the irresistible impulse test and The United States Congress adopted an insanity test for the federal courts in 1984.

Keywords: insanity defense, psychology law, The Federal Insanity Defense Reform Act of 1984, The Wild Beast Standard in 1723

Procedia PDF Downloads 127
14320 Impact of Safety and Quality Considerations of Housing Clients on the Construction Firms’ Intention to Adopt Quality Function Deployment: A Case of Construction Sector

Authors: Saif Ul Haq

Abstract:

The current study intends to examine the safety and quality considerations of clients of housing projects and their impact on the adoption of Quality Function Deployment (QFD) by the construction firm. Mixed method research technique has been used to collect and analyze the data wherein a survey was conducted to collect the data from 220 clients of housing projects in Saudi Arabia. Then, the telephonic and Skype interviews were conducted to collect data of 15 professionals working in the top ten real estate companies of Saudi Arabia. Data were analyzed by using partial least square (PLS) and thematic analysis techniques. Findings reveal that today’s customer prioritizes the safety and quality requirements of their houses and as a result, construction firms adopt QFD to address the needs of customers. The findings are of great importance for the clients of housing projects as well as for the construction firms as they could apply QFD in housing projects to address the safety and quality concerns of their clients.

Keywords: construction industry, quality considerations, quality function deployment, safety considerations

Procedia PDF Downloads 106