Search results for: probability of detection (PD)
2264 Predictive Maintenance Based on Oil Analysis Applicable to Transportation Fleets
Authors: Israel Ibarra Solis, Juan Carlos Rodriguez Sierra, Ma. del Carmen Salazar Hernandez, Isis Rodriguez Sanchez, David Perez Guerrero
Abstract:
At the present paper we try to explain the analysis techniques use for the lubricating oil in a maintenance period of a city bus (Mercedes Benz Boxer 40), which is call ‘R-24 route’, line Coecillo Centro SA de CV in Leon Guanajuato, to estimate the optimal time for the oil change. Using devices such as the rotational viscometer and the atomic absorption spectrometer, they can detect the incipient form when the oil loses its lubricating properties and, therefore, cannot protect the mechanical components of diesel engines such these trucks. Timely detection of lost property in the oil, it allows us taking preventive plan maintenance for the fleet.Keywords: atomic absorption spectrometry, maintenance, predictive velocity rate, lubricating oils
Procedia PDF Downloads 5692263 On Pooling Different Levels of Data in Estimating Parameters of Continuous Meta-Analysis
Authors: N. R. N. Idris, S. Baharom
Abstract:
A meta-analysis may be performed using aggregate data (AD) or an individual patient data (IPD). In practice, studies may be available at both IPD and AD level. In this situation, both the IPD and AD should be utilised in order to maximize the available information. Statistical advantages of combining the studies from different level have not been fully explored. This study aims to quantify the statistical benefits of including available IPD when conducting a conventional summary-level meta-analysis. Simulated meta-analysis were used to assess the influence of the levels of data on overall meta-analysis estimates based on IPD-only, AD-only and the combination of IPD and AD (mixed data, MD), under different study scenario. The percentage relative bias (PRB), root mean-square-error (RMSE) and coverage probability were used to assess the efficiency of the overall estimates. The results demonstrate that available IPD should always be included in a conventional meta-analysis using summary level data as they would significantly increased the accuracy of the estimates. On the other hand, if more than 80% of the available data are at IPD level, including the AD does not provide significant differences in terms of accuracy of the estimates. Additionally, combining the IPD and AD has moderating effects on the biasness of the estimates of the treatment effects as the IPD tends to overestimate the treatment effects, while the AD has the tendency to produce underestimated effect estimates. These results may provide some guide in deciding if significant benefit is gained by pooling the two levels of data when conducting meta-analysis.Keywords: aggregate data, combined-level data, individual patient data, meta-analysis
Procedia PDF Downloads 3752262 Comparing Xbar Charts: Conventional versus Reweighted Robust Estimation Methods for Univariate Data Sets
Authors: Ece Cigdem Mutlu, Burak Alakent
Abstract:
Maintaining the quality of manufactured products at a desired level depends on the stability of process dispersion and location parameters and detection of perturbations in these parameters as promptly as possible. Shewhart control chart is the most widely used technique in statistical process monitoring to monitor the quality of products and control process mean and variability. In the application of Xbar control charts, sample standard deviation and sample mean are known to be the most efficient conventional estimators in determining process dispersion and location parameters, respectively, based on the assumption of independent and normally distributed datasets. On the other hand, there is no guarantee that the real-world data would be normally distributed. In the cases of estimated process parameters from Phase I data clouded with outliers, efficiency of traditional estimators is significantly reduced, and performance of Xbar charts are undesirably low, e.g. occasional outliers in the rational subgroups in Phase I data set may considerably affect the sample mean and standard deviation, resulting a serious delay in detection of inferior products in Phase II. For more efficient application of control charts, it is required to use robust estimators against contaminations, which may exist in Phase I. In the current study, we present a simple approach to construct robust Xbar control charts using average distance to the median, Qn-estimator of scale, M-estimator of scale with logistic psi-function in the estimation of process dispersion parameter, and Harrell-Davis qth quantile estimator, Hodge-Lehmann estimator and M-estimator of location with Huber psi-function and logistic psi-function in the estimation of process location parameter. Phase I efficiency of proposed estimators and Phase II performance of Xbar charts constructed from these estimators are compared with the conventional mean and standard deviation statistics both under normality and against diffuse-localized and symmetric-asymmetric contaminations using 50,000 Monte Carlo simulations on MATLAB. Consequently, it is found that robust estimators yield parameter estimates with higher efficiency against all types of contaminations, and Xbar charts constructed using robust estimators have higher power in detecting disturbances, compared to conventional methods. Additionally, utilizing individuals charts to screen outlier subgroups and employing different combination of dispersion and location estimators on subgroups and individual observations are found to improve the performance of Xbar charts.Keywords: average run length, M-estimators, quality control, robust estimators
Procedia PDF Downloads 1902261 Evaluation of Earthquake Induced Cost for Mid-Rise Buildings
Authors: Gulsah Olgun, Ozgur Bozdag, Yildirim Ertutar
Abstract:
This paper mainly focuses on performance assessment of buildings by associating the damage level with the damage cost. For this purpose a methodology is explained and applied to the representative mid-rise concrete building residing in Izmir. In order to consider uncertainties in occurrence of earthquakes, the structural analyses are conducted for all possible earthquakes in the region through the hazard curve. By means of the analyses, probability of the structural response being in different limit states are obtained and used to calculate expected damage cost. The expected damage cost comprises diverse cost components related to earthquake such as cost of casualties, replacement or repair cost of building etc. In this study, inter-story drift is used as an effective response variable to associate expected damage cost with different damage levels. The structural analysis methods performed to obtain inter story drifts are response spectrum method as a linear one, accurate push-over and time history methods to demonstrate the nonlinear effects on loss estimation. Comparison of the results indicates that each method provides similar values of expected damage cost. To sum up, this paper explains an approach which enables to minimize the expected damage cost of buildings and relate performance level to damage cost.Keywords: expected damage cost, limit states, loss estimation, performance based design
Procedia PDF Downloads 2692260 Nuclear Terrorism Decision Making: A Comparative Study of South Asian Nuclear Weapons States
Authors: Muhammad Jawad Hashmi
Abstract:
The idea of nuclear terrorism is as old as nuclear weapons but the global concerns of likelihood of nuclear terrorism are uncertain. Post 9/11 trends manifest that terrorists are believers of massive causalities. Innovation in terrorist’s tactics, sophisticated weaponry, vulnerability, theft and smuggling of nuclear/radiological material, connections between terrorists, black market and rough regimes are signaling seriousness of upcoming challenges as well as global trends of “terror-transnationalism.” Furthermore, the International-Atomic-Energy-Agency’s database recorded 2734 incidents regarding misuse, unauthorized possession, trafficking of nuclear material etc. Since, this data also includes incidents from south Asia, so, there is every possibility to claim that such illicit activities may increase in future, mainly due to expansion of nuclear industry in South Asia. Moreover, due to such mishaps the region is vulnerable to threats of nuclear terrorism. This is also a reason that the region is in limelight along with issues such as rapidly growing nuclear arsenals, nuclear safety and security, terrorism and political instability. With this backdrop, this study is aimed to investigate the prevailing threats and challenges in South Asia vis a vis nuclear safety and security. A comparative analysis of the overall capabilities would be done to identify the areas of cooperation to eliminate the probability of nuclear/radiological terrorism in the region.Keywords: nuclear terrorism, safety, security, South Asia, india, Pakistan
Procedia PDF Downloads 3562259 Assessment of the Tectonic Effects on Soil Radon Activity along the Margin of the Arabian Plate Boundary in Northwestern Syria
Authors: Mohamed Al-Hilal
Abstract:
The main purpose of the present study is to assess the role of active tectonics in influencing the emanation level of soil radon across two tectonically active structures of the Northern Dead Sea Fault (NDSF) in northwestern Syria: namely, the Qastoon and Al-Harif fault segments. The radon measurements were basically directed by the results drawn from earlier studies of archaeoseismic and paleoseismic investigation in Al-Harif, besides integrated geophysical and morphotectonic survey at the Qastoon site. In view of that, a total of 80 soil gas radon points were measured in this work with a sampling depth of 75 cm, using the AlphaGUARD PQ 2000Pro radon detector. The background range of normal radon emission from local soil was determined in area located away from the influence of the tectonic disturbances. The obtained radon data were statistically analyzed, and the mean values have been standardized in terms of probability of magnitude, which enhances the comparison process and so facilitating the separation of normal radon variations from other anomalous or geotectonic related values. The overall results revealed remarkable occurrences of fault-associated radon anomalies with maximum peak values of ~6 to 7 times above the background, trending in accordance with the predicted traces of the fault ruptures at the Qastoon and Al-Harif, respectively.Keywords: soil gas radon, active tectonic structure, northern dead sea fault, western Syria
Procedia PDF Downloads 1752258 Software Cloning and Agile Environment
Authors: Ravi Kumar, Dhrubajit Barman, Nomi Baruah
Abstract:
Software Cloning has grown an active area in software engineering research community yielding numerous techniques, various tools and other methods for clone detection and removal. The copying, modifying a block of code is identified as cloning as it is the most basic means of software reuse. Agile Software Development is an approach which is currently being used in various software projects, so that it helps to respond the unpredictability of building software through incremental, iterative, work cadences. Software Cloning has been introduced to Agile Environment and many Agile Software Development approaches are using the concept of Software Cloning. This paper discusses the various Agile Software Development approaches. It also discusses the degree to which the Software Cloning concept is being introduced in the Agile Software Development approaches.Keywords: agile environment, refactoring, reuse, software cloning
Procedia PDF Downloads 5312257 An Empirical Study of the Best Fitting Probability Distributions for Stock Returns Modeling
Authors: Jayanta Pokharel, Gokarna Aryal, Netra Kanaal, Chris Tsokos
Abstract:
Investment in stocks and shares aims to seek potential gains while weighing the risk of future needs, such as retirement, children's education etc. Analysis of the behavior of the stock market returns and making prediction is important for investors to mitigate risk on investment. Historically, the normal variance models have been used to describe the behavior of stock market returns. However, the returns of the financial assets are actually skewed with higher kurtosis, heavier tails, and a higher center than the normal distribution. The Laplace distribution and its family are natural candidates for modeling stock returns. The Variance-Gamma (VG) distribution is the most sought-after distributions for modeling asset returns and has been extensively discussed in financial literatures. In this paper, it explore the other Laplace family, such as Asymmetric Laplace, Skewed Laplace, Kumaraswamy Laplace (KS) together with Variance-Gamma to model the weekly returns of the S&P 500 Index and it's eleven business sector indices. The method of maximum likelihood is employed to estimate the parameters of the distributions and our empirical inquiry shows that the Kumaraswamy Laplace distribution performs much better for stock returns modeling among the choice of distributions used in this study and in practice, KS can be used as a strong alternative to VG distribution.Keywords: stock returns, variance-gamma, kumaraswamy laplace, maximum likelihood
Procedia PDF Downloads 702256 Automatic Segmentation of Lung Pleura Based On Curvature Analysis
Authors: Sasidhar B., Bhaskar Rao N., Ramesh Babu D. R., Ravi Shankar M.
Abstract:
Segmentation of lung pleura is a preprocessing step in Computer-Aided Diagnosis (CAD) which helps in reducing false positives in detection of lung cancer. The existing methods fail in extraction of lung regions with the nodules at the pleura of the lungs. In this paper, a new method is proposed which segments lung regions with nodules at the pleura of the lungs based on curvature analysis and morphological operators. The proposed algorithm is tested on 06 patient’s dataset which consists of 60 images of Lung Image Database Consortium (LIDC) and the results are found to be satisfactory with 98.3% average overlap measure (AΩ).Keywords: curvature analysis, image segmentation, morphological operators, thresholding
Procedia PDF Downloads 5962255 Development of a Pain Detector Using Microwave Radiometry Method
Authors: Nanditha Rajamani, Anirudhaa R. Rao, Divya Sriram
Abstract:
One of the greatest difficulties in treating patients with pain is the highly subjective nature of pain sensation. The measurement of pain intensity is primarily dependent on the patient’s report, often with little physical evidence to provide objective corroboration. This is also complicated by the fact that there are only few and expensive existing technologies (Functional Magnetic Resonance Imaging-fMRI). The need is thus clear and urgent for a reliable, non-invasive, non-painful, objective, readily adoptable, and coefficient diagnostic platform that provides additional diagnostic information to supplement its current regime with more information to assist doctors in diagnosing these patients. Thus, our idea of developing a pain detector was conceived to take a step further the detection and diagnosis of chronic and acute pain.Keywords: pain sensor, microwave radiometery, pain sensation, fMRI
Procedia PDF Downloads 4562254 Exploring the Feasibility of Introducing Particular Polyphenols into Cow Milk Naturally through Animal Feeding
Authors: Steve H. Y. Lee, Jeremy P. E. Spencer
Abstract:
The aim of the present study was to explore the feasibility of enriching polyphenols in cow milk via addition of flavanone-rich citrus pulp to existing animal feed. 8 Holstein lactating cows were enrolled onto the 4 week feeding study. 4 cows were fed the standard farm diet (control group), with another 4 (treatment group) which are fed a standard farm diet mixed with citrus pulp diet. Milk was collected twice a day, 3 times a week. The resulting milk yield and its macronutrient composition as well as lactose content were measured. The milk phenolic compounds were analysed using electrochemical detection (ECD).Keywords: milk, polyphenol, animal feeding, lactating cows
Procedia PDF Downloads 2992253 Analysis of Risks of Adopting Integrated Project Delivery: Application of Bayesian Theory
Abstract:
Integrated project delivery (IPD) is a project delivery method distinguished by a shared risk/rewards mechanism and multiparty agreement. IPD has drawn increasing attention from construction industry due to its reliability to deliver high-performing buildings. However, unavailable IPD specific insurance concerns the industry participants who are interested in IPD implementation. Even though the risk management capability can be enhanced using shared risk mechanism, some risks may occur when the partners do not commit themselves into the integrated practices in a desired manner. This is because the intense collaboration and close integration can not only create added value but bring new opportunistic behaviors and disputes. The study is aimed to investigate the risks of implementing IPD using Bayesian theory. IPD risk taxonomy is presented to identify all potential risks of implementing IPD and a risk network map is developed to capture the interdependencies between IPD risks. The conditional relations between risk occurrences and the impacts of IPD risks on project performances are evaluated and simulated based on Bayesian theory. The probability of project outcomes is predicted by simulation. In addition, it is found that some risks caused by integration are most possible occurred risks. This study can help the IPD project participants identify critical risks of adopting IPD to improve project performances. In addition, it is helpful to develop IPD specific insurance when the pertinent risks can be identified.Keywords: Bayesian theory, integrated project delivery, project risks, project performances
Procedia PDF Downloads 3002252 Semantic Data Schema Recognition
Authors: Aïcha Ben Salem, Faouzi Boufares, Sebastiao Correia
Abstract:
The subject covered in this paper aims at assisting the user in its quality approach. The goal is to better extract, mix, interpret and reuse data. It deals with the semantic schema recognition of a data source. This enables the extraction of data semantics from all the available information, inculding the data and the metadata. Firstly, it consists of categorizing the data by assigning it to a category and possibly a sub-category, and secondly, of establishing relations between columns and possibly discovering the semantics of the manipulated data source. These links detected between columns offer a better understanding of the source and the alternatives for correcting data. This approach allows automatic detection of a large number of syntactic and semantic anomalies.Keywords: schema recognition, semantic data profiling, meta-categorisation, semantic dependencies inter columns
Procedia PDF Downloads 4182251 Breast Cancer Risk Factors: A Big Data Analysis of Black and White Women in the USA
Authors: Tejasvi Parupudi, Mochen Li, Lakshya Mittal, Ignacio G. Camarillo, Raji Sundararajan
Abstract:
With breast cancer becoming a global pandemic, it is very important to assess a woman’s risk profile accurately in a timely manner. Providing an estimate of the risk of developing breast cancer to a woman gives her an opportunity to consider options to decrease this risk. Women at low risk may be suggested yearly screenings whereas women with a high risk of developing breast cancer would be candidates for aggressive surveillance. Fortunately, there is a set of risk factors that are used to predict the probability of a woman being diagnosed with breast cancer in the future. Studying risk factors and understanding how they correlate to cancer is important for early diagnosis, prevention and reducing mortality rates. The effect of crucial risk factors among black and white women was compared in this study. The various risk factors analyzed include breast density, age, cancer in a first-degree relative, menopausal status, body mass index (BMI) and prior breast cancer diagnosis, etc. Breast density, age at first full-term birth and BMI were utilized in this study as important risk factors for the comparison of incidence rates between women of black and white races in the USA. Understanding the differences could lead to the development of solutions to reduce disparity in mortality rates among black women by improving overall access to care.Keywords: big data, breast cancer, risk factors, incidence rates, mortality, race
Procedia PDF Downloads 2752250 Antibiotic and Fungicide Exposure Reveal the Evolution of Soil-Lettuce System Resistome
Authors: Chenyu Huang, Minrong Cui, Hua Fang, Luqing Zhang, Yunlong Yu
Abstract:
The emergence and spread of antibiotic resistance genes (ARGs) have become a pressing issue in global agricultural production. However, understanding how these ARGs spread across different spatial scales, especially when exposed to both pesticides and antibiotics, has remained a challenge. Here, metagenomic assembly and binning methodologies were used to determine the mechanism of ARG propagation within soil-lettuce systems exposed to both fungicides and antibiotics. The results of our study showed that the presence of fungicide and antibiotic stresses had a significant impact on certain bacterial communities. Notably, we observed that ARGs were primarily transferred from the soil to the plant through plasmids. The selective pressure exerted by fungicides and antibiotics contributed to an increase in unique ARGs present on lettuce leaves. Moreover, ARGs located on chromosomes and plasmids followed different transmission patterns. The presence of diverse selective pressures, a result of compound treatments involving antibiotics and fungicides, amplifies this phenomenon. Consequently, there is a higher probability of bacteria developing multi-antibiotic resistance under the combined pressure of fungicides and antibiotics. In summary, our findings highlight that combined fungicide and antibiotic treatments are more likely to drive the acquisition of ARGs within the soil-plant system and may increase the risk of human ingestion.Keywords: soil-lettuce system, fungicide, antibiotic, ARG, transmission
Procedia PDF Downloads 1092249 1-Butyl-2,3-Dimethylimidazolium Bis (Trifluoromethanesulfonyl) Imide and Titanium Oxide Based Voltammetric Sensor for the Quantification of Flunarizine Dihydrochloride in Solubilized Media
Authors: Rajeev Jain, Nimisha Jadon, Kshiti Singh
Abstract:
Titanium oxide nanoparticles and 1-butyl-2,3-dimethylimidazolium bis (trifluoromethane- sulfonyl) imide modified glassy carbon electrode (TiO2/IL/GCE) has been fabricated for electrochemical sensing of flunarizine dihydrochloride (FRH). The electrochemical properties and morphology of the prepared nanocomposite were studied by electrochemical impedance spectroscopy (EIS) and transmission electron microscopy (TEM). The response of the electrochemical sensor was found to be proportional to the concentrations of FRH in the range from 0.5 µg mL-1 to 16 µg mL-1. The detection limit obtained was 0.03 µg mL-1. The proposed method was also applied to the determination of FRH in pharmaceutical formulation and human serum with good recoveries.Keywords: flunarizine dihydrochloride, ionic liquid, nanoparticles, voltammetry, human serum
Procedia PDF Downloads 3292248 The Underestimate of the Annual Maximum Rainfall Depths Due to Coarse Time Resolution Data
Authors: Renato Morbidelli, Carla Saltalippi, Alessia Flammini, Tommaso Picciafuoco, Corrado Corradini
Abstract:
A considerable part of rainfall data to be used in the hydrological practice is available in aggregated form within constant time intervals. This can produce undesirable effects, like the underestimate of the annual maximum rainfall depth, Hd, associated with a given duration, d, that is the basic quantity in the development of rainfall depth-duration-frequency relationships and in determining if climate change is producing effects on extreme event intensities and frequencies. The errors in the evaluation of Hd from data characterized by a coarse temporal aggregation, ta, and a procedure to reduce the non-homogeneity of the Hd series are here investigated. Our results indicate that: 1) in the worst conditions, for d=ta, the estimation of a single Hd value can be affected by an underestimation error up to 50%, while the average underestimation error for a series with at least 15-20 Hd values, is less than or equal to 16.7%; 2) the underestimation error values follow an exponential probability density function; 3) each very long time series of Hd contains many underestimated values; 4) relationships between the non-dimensional ratio ta/d and the average underestimate of Hd, derived from continuous rainfall data observed in many stations of Central Italy, may overcome this issue; 5) these equations should allow to improve the Hd estimates and the associated depth-duration-frequency curves at least in areas with similar climatic conditions.Keywords: central Italy, extreme events, rainfall data, underestimation errors
Procedia PDF Downloads 1912247 Unsupervised Learning of Spatiotemporally Coherent Metrics
Authors: Ross Goroshin, Joan Bruna, Jonathan Tompson, David Eigen, Yann LeCun
Abstract:
Current state-of-the-art classification and detection algorithms rely on supervised training. In this work we study unsupervised feature learning in the context of temporally coherent video data. We focus on feature learning from unlabeled video data, using the assumption that adjacent video frames contain semantically similar information. This assumption is exploited to train a convolutional pooling auto-encoder regularized by slowness and sparsity. We establish a connection between slow feature learning to metric learning and show that the trained encoder can be used to define a more temporally and semantically coherent metric.Keywords: machine learning, pattern clustering, pooling, classification
Procedia PDF Downloads 4562246 Spatial and Temporal Analysis of Forest Cover Change with Special Reference to Anthropogenic Activities in Kullu Valley, North-Western Indian Himalayan Region
Authors: Krisala Joshi, Sayanta Ghosh, Renu Lata, Jagdish C. Kuniyal
Abstract:
Throughout the world, monitoring and estimating the changing pattern of forests across diverse landscapes through remote sensing is instrumental in understanding the interactions of human activities and the ecological environment with the changing climate. Forest change detection using satellite imageries has emerged as an important means to gather information on a regional scale. Kullu valley in Himachal Pradesh, India is situated in a transitional zone between the lesser and the greater Himalayas. Thus, it presents a typical rugged mountainous terrain with moderate to high altitude which varies from 1200 meters to over 6000 meters. Due to changes in agricultural cropping patterns, urbanization, industrialization, hydropower generation, climate change, tourism, and anthropogenic forest fire, it has undergone a tremendous transformation in forest cover in the past three decades. The loss and degradation of forest cover results in soil erosion, loss of biodiversity including damage to wildlife habitats, and degradation of watershed areas, and deterioration of the overall quality of nature and life. The supervised classification of LANDSAT satellite data was performed to assess the changes in forest cover in Kullu valley over the years 2000 to 2020. Normalized Burn Ratio (NBR) was calculated to discriminate between burned and unburned areas of the forest. Our study reveals that in Kullu valley, the increasing number of forest fire incidents specifically, those due to anthropogenic activities has been on a rise, each subsequent year. The main objective of the present study is, therefore, to estimate the change in the forest cover of Kullu valley and to address the various social aspects responsible for the anthropogenic forest fires. Also, to assess its impact on the significant changes in the regional climatic factors, specifically, temperature, humidity, and precipitation over three decades, with the help of satellite imageries and ground data. The main outcome of the paper, we believe, will be helpful for the administration for making a quantitative assessment of the forest cover area changes due to anthropogenic activities and devising long-term measures for creating awareness among the local people of the area.Keywords: Anthropogenic Activities, Forest Change Detection, Normalized Burn Ratio (NBR), Supervised Classification
Procedia PDF Downloads 1732245 Facial Expression Recognition Using Sparse Gaussian Conditional Random Field
Authors: Mohammadamin Abbasnejad
Abstract:
The analysis of expression and facial Action Units (AUs) detection are very important tasks in fields of computer vision and Human Computer Interaction (HCI) due to the wide range of applications in human life. Many works have been done during the past few years which has their own advantages and disadvantages. In this work, we present a new model based on Gaussian Conditional Random Field. We solve our objective problem using ADMM and we show how well the proposed model works. We train and test our work on two facial expression datasets, CK+, and RU-FACS. Experimental evaluation shows that our proposed approach outperform state of the art expression recognition.Keywords: Gaussian Conditional Random Field, ADMM, convergence, gradient descent
Procedia PDF Downloads 3562244 Size Reduction of Images Using Constraint Optimization Approach for Machine Communications
Authors: Chee Sun Won
Abstract:
This paper presents the size reduction of images for machine-to-machine communications. Here, the salient image regions to be preserved include the image patches of the key-points such as corners and blobs. Based on a saliency image map from the key-points and their image patches, an axis-aligned grid-size optimization is proposed for the reduction of image size. To increase the size-reduction efficiency the aspect ratio constraint is relaxed in the constraint optimization framework. The proposed method yields higher matching accuracy after the size reduction than the conventional content-aware image size-reduction methods.Keywords: image compression, image matching, key-point detection and description, machine-to-machine communication
Procedia PDF Downloads 4182243 Risk Assessment of Lead Element in Red Peppers Collected from Marketplaces in Antalya, Southern Turkey
Authors: Serpil Kilic, Ihsan Burak Cam, Murat Kilic, Timur Tongur
Abstract:
Interest in the lead (Pb) has considerably increased due to knowledge about the potential toxic effects of this element, recently. Exposure to heavy metals above the acceptable limit affects human health. Indeed, Pb is accumulated through food chains up to toxic concentrations; therefore, it can pose an adverse potential threat to human health. A sensitive and reliable method for determination of Pb element in red pepper were improved in the present study. Samples (33 red pepper products having different brands) were purchased from different markets in Turkey. The selected method validation criteria (linearity, Limit of Detection, Limit of Quantification, recovery, and trueness) demonstrated. Recovery values close to 100% showed adequate precision and accuracy for analysis. According to the results of red pepper analysis, all of the tested lead element in the samples was determined at various concentrations. A Perkin- Elmer ELAN DRC-e model ICP-MS system was used for detection of Pb. Organic red pepper was used to obtain a matrix for all method validation studies. The certified reference material, Fapas chili powder, was digested and analyzed, together with the different sample batches. Three replicates from each sample were digested and analyzed. The results of the exposure levels of the elements were discussed considering the scientific opinions of the European Food Safety Authority (EFSA), which is the European Union’s (EU) risk assessment source associated with food safety. The Target Hazard Quotient (THQ) was described by the United States Environmental Protection Agency (USEPA) for the calculation of potential health risks associated with long-term exposure to chemical pollutants. THQ value contains intake of elements, exposure frequency and duration, body weight and the oral reference dose (RfD). If the THQ value is lower than one, it means that the exposed population is assumed to be safe and 1 < THQ < 5 means that the exposed population is in a level of concern interval. In this study, the THQ of Pb was obtained as < 1. The results of THQ calculations showed that the values were below one for all the tested, meaning the samples did not pose a health risk to the local population. This work was supported by The Scientific Research Projects Coordination Unit of Akdeniz University. Project Number: FBA-2017-2494.Keywords: lead analyses, red pepper, risk assessment, daily exposure
Procedia PDF Downloads 1672242 Controlling Interactions and Non-Equilibrium Steady State in Spinning Active Matter Monolayers
Authors: Joshua Paul Steimel, Michael Pappas, Ethan Hall
Abstract:
Particle-particle interactions are critical in determining the state of an active matter system. Unique and ubiquitous non-equilibrium behavior like swarming, vortexing, spiraling, and much more is governed by interactions between active units or particles. In hybrid active-passive matter systems, the attraction between spinning active units in a 2D monolayer of passive particles is controlled by the mechanical behavior of the passive monolayer. We demonstrate here that the range and dynamics of this attraction can be controlled by changing the composition of the passive monolayer by adding dopant passive particles. These dopant passive particles effectively pin the movement of dislocation motion in the passive media and reduce the probability of defect motion required to erode the bridge of passive particles between active spinners, thus reducing the range of attraction. Additionally, by adding an out of plane component to the magnetic moment and creating a top-like motion a short range repulsion emerges between the top-like particle. At inter-top distances less than four particle diameters apart, the tops repel but beyond that, distance attract up to 13 particle diameters apart. The tops were also able to locally and transiently anneal the passive monolayer. Thus we demonstrate that by tuning several parameters of the hybrid active matter system, one can observe very different emergent behavior.Keywords: active matter, colloids, ferromagnetic, annealing
Procedia PDF Downloads 1102241 Scalable UI Test Automation for Large-scale Web Applications
Authors: Kuniaki Kudo, Raviraj Solanki, Kaushal Patel, Yash Virani
Abstract:
This research mainly concerns optimizing UI test automation for large-scale web applications. The test target application is the HHAexchange homecare management WEB application that seamlessly connects providers, state Medicaid programs, managed care organizations (MCOs), and caregivers through one platform with large-scale functionalities. This study focuses on user interface automation testing for the WEB application. The quality assurance team must execute many manual users interface test cases in the development process to confirm no regression bugs. The team automated 346 test cases; the UI automation test execution time was over 17 hours. The business requirement was reducing the execution time to release high-quality products quickly, and the quality assurance automation team modernized the test automation framework to optimize the execution time. The base of the WEB UI automation test environment is Selenium, and the test code is written in Python. Adopting a compilation language to write test code leads to an inefficient flow when introducing scalability into a traditional test automation environment. In order to efficiently introduce scalability into Test Automation, a scripting language was adopted. The scalability implementation is mainly implemented with AWS's serverless technology, an elastic container service. The definition of scalability here is the ability to automatically set up computers to test automation and increase or decrease the number of computers running those tests. This means the scalable mechanism can help test cases run parallelly. Then test execution time is dramatically decreased. Also, introducing scalable test automation is for more than just reducing test execution time. There is a possibility that some challenging bugs are detected by introducing scalable test automation, such as race conditions, Etc. since test cases can be executed at same timing. If API and Unit tests are implemented, the test strategies can be adopted more efficiently for this scalability testing. However, in WEB applications, as a practical matter, API and Unit testing cannot cover 100% functional testing since they do not reach front-end codes. This study applied a scalable UI automation testing strategy to the large-scale homecare management system. It confirmed the optimization of the test case execution time and the detection of a challenging bug. This study first describes the detailed architecture of the scalable test automation environment, then describes the actual performance reduction time and an example of challenging issue detection.Keywords: aws, elastic container service, scalability, serverless, ui automation test
Procedia PDF Downloads 1072240 Bayesian Analysis of Change Point Problems Using Conditionally Specified Priors
Authors: Golnaz Shahtahmassebi, Jose Maria Sarabia
Abstract:
In this talk, we introduce a new class of conjugate prior distributions obtained from conditional specification methodology. We illustrate the application of such distribution in Bayesian change point detection in Poisson processes. We obtain the posterior distribution of model parameters using a general bivariate distribution with gamma conditionals. Simulation from the posterior is readily implemented using a Gibbs sampling algorithm. The Gibbs sampling is implemented even when using conditional densities that are incompatible or only compatible with an improper joint density. The application of such methods will be demonstrated using examples of simulated and real data.Keywords: change point, bayesian inference, Gibbs sampler, conditional specification, gamma conditional distributions
Procedia PDF Downloads 1892239 Explicit Numerical Approximations for a Pricing Weather Derivatives Model
Authors: Clarinda V. Nhangumbe, Ercília Sousa
Abstract:
Weather Derivatives are financial instruments used to cover non-catastrophic weather events and can be expressed in the form of standard or plain vanilla products, structured or exotics products. The underlying asset, in this case, is the weather index, such as temperature, rainfall, humidity, wind, and snowfall. The complexity of the Weather Derivatives structure shows the weakness of the Black Scholes framework. Therefore, under the risk-neutral probability measure, the option price of a weather contract can be given as a unique solution of a two-dimensional partial differential equation (parabolic in one direction and hyperbolic in other directions), with an initial condition and subjected to adequate boundary conditions. To calculate the price of the option, one can use numerical methods such as the Monte Carlo simulations and implicit finite difference schemes conjugated with Semi-Lagrangian methods. This paper is proposed two explicit methods, namely, first-order upwind in the hyperbolic direction combined with Lax-Wendroff in the parabolic direction and first-order upwind in the hyperbolic direction combined with second-order upwind in the parabolic direction. One of the advantages of these methods is the fact that they take into consideration the boundary conditions obtained from the financial interpretation and deal efficiently with the different choices of the convection coefficients.Keywords: incomplete markets, numerical methods, partial differential equations, stochastic process, weather derivatives
Procedia PDF Downloads 852238 Comparison of Seismic Response for Two RC Curved Bridges with Different Column Shapes
Authors: Nina N. Serdar, Jelena R. Pejović
Abstract:
This paper presents seismic risk assessment of two bridge structure, based on the probabilistic performance-based seismic assessment methodology. Both investigated bridges are tree span continuous RC curved bridges with the difference in column shapes. First bridge (type A) has a wall-type pier and second (type B) has a two-column bent with circular columns. Bridges are designed according to European standards: EN 1991-2, EN1992-1-1 and EN 1998-2. Aim of the performed analysis is to compare seismic behavior of these two structures and to detect the influence of column shapes on the seismic response. Seismic risk assessment is carried out by obtaining demand fragility curves. Non-linear model was constructed and time-history analysis was performed using thirty five pairs of horizontal ground motions selected to match site specific hazard. In performance based analysis, peak column drift ratio (CDR) was selected as engineering demand parameter (EDP). For seismic intensity measure (IM) spectral displacement was selected. Demand fragility curves that give probability of exceedance of certain value for chosen EDP were constructed and based on them conclusions were made.Keywords: RC curved bridge, demand fragility curve, wall type column, nonlinear time-history analysis, circular column
Procedia PDF Downloads 3412237 The Application of Hellomac Rockfall Alert System in Rockfall Barriers: An Explainer
Authors: Kinjal Parmar, Matteo Lelli
Abstract:
The usage of IoT technology as a rockfall alert system is relatively new. This paper explains the potential of such an alert system called HelloMac from Maccaferri which provides transportation infrastructure asset owners the way to effectively utilize their resources in the detection of boulder impacts on rockfall barriers. This would ensure a faster assessment of the impacted barrier and subsequently facilitates the implementation of remedial works in an effective and timely manner. In addition, the HelloMac can also be integrated with another warning system to alert vehicle users of the unseen dangers ahead. HelloMac is developed to work also in remote areas, where cell coverage is not available. User gets notified when a rockfall even occurs via mobile app, SMS and email. Using such alarming systems effectively, we can reduce the risk of rockfall hazard.Keywords: rockfall, barrier, HelloMac, rockfall alert system
Procedia PDF Downloads 522236 Electrochemical Response Transductions of Graphenated-Polyaniline Nanosensor for Environmental Anthracene
Authors: O. Tovide, N. Jahed, N. Mohammed, C. E. Sunday, H. R. Makelane, R. F. Ajayi, K. M. Molapo, A. Tsegaye, M. Masikini, S. Mailu, A. Baleg, T. Waryo, P. G. Baker, E. I. Iwuoha
Abstract:
A graphenated–polyaniline (GR-PANI) nanocomposite sensor was constructed and used for the determination of anthracene. The direct electro-oxidation behavior of anthracene on the GR-PANI modified glassy carbon electrode (GCE) was used as the sensing principle. The results indicate thatthe response profile of the oxidation of anthracene on GR-PANI-modified GCE provides for the construction of sensor systems based onamperometric and potentiometric signal transductions. A dynamic linear range of 0.12- 100 µM anthracene and a detection limit of 0.044 µM anthracene were established for the sensor system.Keywords: electrochemical sensors, environmental pollutants, graphenated-polymers, polyaromatic hydrocarbon
Procedia PDF Downloads 3562235 Quantification and Detection of Non-Sewer Water Infiltration and Inflow in Urban Sewer Systems
Authors: M. Beheshti, S. Saegrov, T. M. Muthanna
Abstract:
Separated sewer systems are designed to transfer the wastewater from houses and industrial sections to wastewater treatment plants. Unwanted water in the sewer systems is a well-known problem, i.e. storm-water inflow is around 50% of the foul sewer, and groundwater infiltration to the sewer system can exceed 50% of total wastewater volume in deteriorated networks. Infiltration and inflow of non-sewer water (I/I) into sewer systems is unfavorable in separated sewer systems and can trigger overloading the system and reducing the efficiency of wastewater treatment plants. Moreover, I/I has negative economic, environmental, and social impacts on urban areas. Therefore, for having sustainable management of urban sewer systems, I/I of unwanted water into the urban sewer systems should be considered carefully and maintenance and rehabilitation plan should be implemented on these water infrastructural assets. This study presents a methodology to identify and quantify the level of I/I into the sewer system. Amount of I/I is evaluated by accurate flow measurement in separated sewer systems for specified isolated catchments in Trondheim city (Norway). Advanced information about the characteristics of I/I is gained by CCTV inspection of sewer pipelines with high I/I contribution. Achieving enhanced knowledge about the detection and localization of non-sewer water in foul sewer system during the wet and dry weather conditions will enable the possibility for finding the problem of sewer system and prioritizing them and taking decisions for rehabilitation and renewal planning in the long-term. Furthermore, preventive measures and optimization of sewer systems functionality and efficiency can be executed by maintenance of sewer system. In this way, the exploitation of sewer system can be improved by maintenance and rehabilitation of existing pipelines in a sustainable way by more practical cost-effective and environmental friendly way. This study is conducted on specified catchments with different properties in Trondheim city. Risvollan catchment is one of these catchments with a measuring station to investigate hydrological parameters through the year, which also has a good database. For assessing the infiltration in a separated sewer system, applying the flow rate measurement method can be utilized in obtaining a general view of the network condition from infiltration point of view. This study discusses commonly used and advanced methods of localizing and quantifying I/I in sewer systems. A combination of these methods give sewer operators the possibility to compare different techniques and obtain reliable and accurate I/I data which is vital for long-term rehabilitation plans.Keywords: flow rate measurement, infiltration and inflow (I/I), non-sewer water, separated sewer systems, sustainable management
Procedia PDF Downloads 333