Search results for: random testing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4863

Search results for: random testing

4473 Markov Random Field-Based Segmentation Algorithm for Detection of Land Cover Changes Using Uninhabited Aerial Vehicle Synthetic Aperture Radar Polarimetric Images

Authors: Mehrnoosh Omati, Mahmod Reza Sahebi

Abstract:

The information on land use/land cover changing plays an essential role for environmental assessment, planning and management in regional development. Remotely sensed imagery is widely used for providing information in many change detection applications. Polarimetric Synthetic aperture radar (PolSAR) image, with the discrimination capability between different scattering mechanisms, is a powerful tool for environmental monitoring applications. This paper proposes a new boundary-based segmentation algorithm as a fundamental step for land cover change detection. In this method, first, two PolSAR images are segmented using integration of marker-controlled watershed algorithm and coupled Markov random field (MRF). Then, object-based classification is performed to determine changed/no changed image objects. Compared with pixel-based support vector machine (SVM) classifier, this novel segmentation algorithm significantly reduces the speckle effect in PolSAR images and improves the accuracy of binary classification in object-based level. The experimental results on Uninhabited Aerial Vehicle Synthetic Aperture Radar (UAVSAR) polarimetric images show a 3% and 6% improvement in overall accuracy and kappa coefficient, respectively. Also, the proposed method can correctly distinguish homogeneous image parcels.

Keywords: coupled Markov random field (MRF), environment, object-based analysis, polarimetric SAR (PolSAR) images

Procedia PDF Downloads 196
4472 Statistical Analysis with Prediction Models of User Satisfaction in Software Project Factors

Authors: Katawut Kaewbanjong

Abstract:

We analyzed a volume of data and found significant user satisfaction in software project factors. A statistical significance analysis (logistic regression) and collinearity analysis determined the significance factors from a group of 71 pre-defined factors from 191 software projects in ISBSG Release 12. The eight prediction models used for testing the prediction potential of these factors were Neural network, k-NN, Naïve Bayes, Random forest, Decision tree, Gradient boosted tree, linear regression and logistic regression prediction model. Fifteen pre-defined factors were truly significant in predicting user satisfaction, and they provided 82.71% prediction accuracy when used with a neural network prediction model. These factors were client-server, personnel changes, total defects delivered, project inactive time, industry sector, application type, development type, how methodology was acquired, development techniques, decision making process, intended market, size estimate approach, size estimate method, cost recording method, and effort estimate method. These findings may benefit software development managers considerably.

Keywords: prediction model, statistical analysis, software project, user satisfaction factor

Procedia PDF Downloads 91
4471 Structural Reliability Analysis Using Extreme Learning Machine

Authors: Mehul Srivastava, Sharma Tushar Ravikant, Mridul Krishn Mishra

Abstract:

In structural design, the evaluation of safety and probability failure of structure is of significant importance, mainly when the variables are random. On real structures, structural reliability can be evaluated obtaining an implicit limit state function. The structural reliability limit state function is obtained depending upon the statistically independent variables. In the analysis of reliability, we considered the statistically independent random variables to be the load intensity applied and the depth or height of the beam member considered. There are many approaches for structural reliability problems. In this paper Extreme Learning Machine technique and First Order Second Moment Method is used to determine the reliability indices for the same set of variables. The reliability index obtained using ELM is compared with the reliability index obtained using FOSM. Higher the reliability index, more feasible is the method to determine the reliability.

Keywords: reliability, reliability index, statistically independent, extreme learning machine

Procedia PDF Downloads 654
4470 An Artificial Intelligence Framework to Forecast Air Quality

Authors: Richard Ren

Abstract:

Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.

Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms

Procedia PDF Downloads 95
4469 Influence of High Temperature and Humidity on Polymer Composites Used in Relining of Sewage

Authors: Parastou Kharazmi, Folke Björk

Abstract:

Some of the main causes for degradation of polymeric materials are thermal aging, hydrolysis, oxidation or chemical degradation by acids, alkalis or water. The first part of this paper provides a brief summary of advances in technology, methods and specification of composite materials for relining as a rehabilitation technique for sewage systems. The second part summarizes an investigation on frequently used composite materials for relining in Sweden, the rubber filled epoxy composite and reinforced polyester composite when they were immersed in deionized water or in dry conditions, and elevated temperatures up to 80°C in the laboratory. The tests were conducted by visual inspection, microscopy, Dynamic Mechanical Analysis (DMA), Differential Scanning Calorimetry (DSC) as well as mechanical testing, three point bending and tensile testing.

Keywords: composite, epoxy, polyester, relining, sewage

Procedia PDF Downloads 310
4468 Application of ATP7B Gene Mutation Analysis in Prenatal Diagnosis of Wilson’s Disease

Authors: Huong M. T. Nguyen, Hoa A. P. Nguyen, Chi V. Phan, Mai P. T. Nguyen, Ngoc D. Ngo, Van T. Ta, Hai T. Le

Abstract:

Wilson’s disease is an autosomal recessive disorder of copper metabolism, which is caused by mutation in copper- transporting P-type ATPase (ATP7B). The mechanism of this disease is a failure of hepatic excretion of copper to the bile, and it leads to copper deposits in the liver and other organs. Most clinical symptoms of Wilson’s disease can present as liver disease and/or neurologic disease. Objective: The goal of the study is prenatal diagnosis for pregnant women at high risk of Wilson’s disease in Northern Vietnam. Material and method: Three probands with clinically diagnosed liver disease were detected in the mutations of 21 exons and exon-intron boundaries of the ATP7B gene by direct Sanger-sequencing. Prenatal diagnoses were performed by amniotic fluid sampling from pregnant women in the 16th-18th weeks of pregnancy after the genotypes of parents with the probands were identified. Result: A total of three different mutations of the probands, including of S105*, P1052L, P1273G, were detected. Among three fetuses which underwent prenatal genetic testing, one fetus was homozygote; two fetuses were carriers. Conclusion: Genetic testing provided a useful method for prenatal diagnosis, and is a basis for genetic counseling.

Keywords: ATP7B gene, genetic testing, prenatal diagnosis, pedigree, Wilson disease

Procedia PDF Downloads 431
4467 Tree-Based Inference for Regionalization: A Comparative Study of Global Topological Perturbation Methods

Authors: Orhun Aydin, Mark V. Janikas, Rodrigo Alves, Renato Assuncao

Abstract:

In this paper, a tree-based perturbation methodology for regionalization inference is presented. Regionalization is a constrained optimization problem that aims to create groups with similar attributes while satisfying spatial contiguity constraints. Similar to any constrained optimization problem, the spatial constraint may hinder convergence to some global minima, resulting in spatially contiguous members of a group with dissimilar attributes. This paper presents a general methodology for rigorously perturbing spatial constraints through the use of random spanning trees. The general framework presented can be used to quantify the effect of the spatial constraints in the overall regionalization result. We compare several types of stochastic spanning trees used in inference problems such as fuzzy regionalization and determining the number of regions. Performance of stochastic spanning trees is juxtaposed against the traditional permutation-based hypothesis testing frequently used in spatial statistics. Inference results for fuzzy regionalization and determining the number of regions is presented on the Local Area Personal Incomes for Texas Counties provided by the Bureau of Economic Analysis.

Keywords: regionalization, constrained clustering, probabilistic inference, fuzzy clustering

Procedia PDF Downloads 197
4466 Artificial Intelligence Based Comparative Analysis for Supplier Selection in Multi-Echelon Automotive Supply Chains via GEP and ANN Models

Authors: Seyed Esmail Seyedi Bariran, Laysheng Ewe, Amy Ling

Abstract:

Since supplier selection appears as a vital decision, selecting supplier based on the best and most accurate ways has a lot of importance for enterprises. In this study, a new Artificial Intelligence approach is exerted to remove weaknesses of supplier selection. The paper has three parts. First part is choosing the appropriate criteria for assessing the suppliers’ performance. Next one is collecting the data set based on experts. Afterwards, the data set is divided into two parts, the training data set and the testing data set. By the training data set the best structure of GEP and ANN are selected and to evaluate the power of the mentioned methods the testing data set is used. The result obtained shows that the accuracy of GEP is more than ANN. Moreover, unlike ANN, a mathematical equation is presented by GEP for the supplier selection.

Keywords: supplier selection, automotive supply chains, ANN, GEP

Procedia PDF Downloads 601
4465 Breast Cancer Detection Using Machine Learning Algorithms

Authors: Jiwan Kumar, Pooja, Sandeep Negi, Anjum Rouf, Amit Kumar, Naveen Lakra

Abstract:

In modern times where, health issues are increasing day by day, breast cancer is also one of them, which is very crucial and really important to find in the early stages. Doctors can use this model in order to tell their patients whether a cancer is not harmful (benign) or harmful (malignant). We have used the knowledge of machine learning in order to produce the model. we have used algorithms like Logistic Regression, Random forest, support Vector Classifier, Bayesian Network and Radial Basis Function. We tried to use the data of crucial parts and show them the results in pictures in order to make it easier for doctors. By doing this, we're making ML better at finding breast cancer, which can lead to saving more lives and better health care.

Keywords: Bayesian network, radial basis function, ensemble learning, understandable, data making better, random forest, logistic regression, breast cancer

Procedia PDF Downloads 22
4464 Crack Growth Life Prediction of a Fighter Aircraft Wing Splice Joint Under Spectrum Loading Using Random Forest Regression and Artificial Neural Networks with Hyperparameter Optimization

Authors: Zafer Yüce, Paşa Yayla, Alev Taşkın

Abstract:

There are heaps of analytical methods to estimate the crack growth life of a component. Soft computing methods have an increasing trend in predicting fatigue life. Their ability to build complex relationships and capability to handle huge amounts of data are motivating researchers and industry professionals to employ them for challenging problems. This study focuses on soft computing methods, especially random forest regressors and artificial neural networks with hyperparameter optimization algorithms such as grid search and random grid search, to estimate the crack growth life of an aircraft wing splice joint under variable amplitude loading. TensorFlow and Scikit-learn libraries of Python are used to build the machine learning models for this study. The material considered in this work is 7050-T7451 aluminum, which is commonly preferred as a structural element in the aerospace industry, and regarding the crack type; corner crack is used. A finite element model is built for the joint to calculate fastener loads and stresses on the structure. Since finite element model results are validated with analytical calculations, findings of the finite element model are fed to AFGROW software to calculate analytical crack growth lives. Based on Fighter Aircraft Loading Standard for Fatigue (FALSTAFF), 90 unique fatigue loading spectra are developed for various load levels, and then, these spectrums are utilized as inputs to the artificial neural network and random forest regression models for predicting crack growth life. Finally, the crack growth life predictions of the machine learning models are compared with analytical calculations. According to the findings, a good correlation is observed between analytical and predicted crack growth lives.

Keywords: aircraft, fatigue, joint, life, optimization, prediction.

Procedia PDF Downloads 143
4463 Automatic MC/DC Test Data Generation from Software Module Description

Authors: Sekou Kangoye, Alexis Todoskoff, Mihaela Barreau

Abstract:

Modified Condition/Decision Coverage (MC/DC) is a structural coverage criterion that is highly recommended or required for safety-critical software coverage. Therefore, many testing standards include this criterion and require it to be satisfied at a particular level of testing (e.g. validation and unit levels). However, an important amount of time is needed to meet those requirements. In this paper we propose to automate MC/DC test data generation. Thus, we present an approach to automatically generate MC/DC test data, from software module description written over a dedicated language. We introduce a new merging approach that provides high MC/DC coverage for the description, with only a little number of test cases.

Keywords: domain-specific language, MC/DC, test data generation, safety-critical software coverage

Procedia PDF Downloads 413
4462 An Empirical Dynamic Fuel Cell Model Used for Power System Verification in Aerospace

Authors: Giuliano Raimondo, Jörg Wangemann, Peer Drechsel

Abstract:

In systems development involving Fuel Cells generators, it is important to have from an early stage of the project a dynamic model for the electrical behavior of the stack to be shared between involved development parties. It allows independent and early design and tests of fuel cell related power electronic. This paper presents an empirical Fuel Cell system model derived from characterization tests on a real system. Moreover, it is illustrated how the obtained model is used to build and validate a real-time Fuel Cell system emulator which is used for aerospace electrical integration testing activities.

Keywords: fuel cell, modelling, real time emulation, testing

Procedia PDF Downloads 313
4461 Performance and Emission Prediction in a Biodiesel Engine Fuelled with Honge Methyl Ester Using RBF Neural Networks

Authors: Shiva Kumar, G. S. Vijay, Srinivas Pai P., Shrinivasa Rao B. R.

Abstract:

In the present study RBF neural networks were used for predicting the performance and emission parameters of a biodiesel engine. Engine experiments were carried out in a 4 stroke diesel engine using blends of diesel and Honge methyl ester as the fuel. Performance parameters like BTE, BSEC, Tech and emissions from the engine were measured. These experimental results were used for ANN modeling. RBF center initialization was done by random selection and by using Clustered techniques. Network was trained by using fixed and varying widths for the RBF units. It was observed that RBF results were having a good agreement with the experimental results. Networks trained by using clustering technique gave better results than using random selection of centers in terms of reduced MRE and increased prediction accuracy. The average MRE for the performance parameters was 3.25% with the prediction accuracy of 98% and for emissions it was 10.4% with a prediction accuracy of 80%.

Keywords: radial basis function networks, emissions, performance parameters, fuzzy c means

Procedia PDF Downloads 532
4460 Remanufacturing and Integrity Assessment of a 27-Year-Old 25kV Gas Insulated Switchgear: A Comprehensive Study on Dismantling, Inspection, and Testing

Authors: Yechan Kim, Bonhyuk Ku, Minkyung Jung, Hyoungku Kang

Abstract:

This study presents the remanufacturing of a 25kV gas insulated switchgear (GIS) that operated indoors for 27 years before being decommissioned due to aging. The research involved a detailed process of dismantling, visual inspection, component-wise examination, and various testing methodologies to assess the equipment's condition. The focus was on evaluating the GIS's integrity and feasibility for remanufacturing. The results highlight the potential of remanufacturing in extending the life of electrical power equipment, offering insights into the best practices, challenges, and technical considerations of such an undertaking. This contributes to a sustainable approach in the power industry, emphasizing the reuse and restoration of aging equipment.

Keywords: remanufacturing, dismantling, gas insulated switchgear, sustainability, life extension

Procedia PDF Downloads 45
4459 Field-Programmable Gate Array Based Tester for Protective Relay

Authors: H. Bentarzi, A. Zitouni

Abstract:

The reliability of the power grid depends on the successful operation of thousands of protective relays. The failure of one relay to operate as intended may lead the entire power grid to blackout. In fact, major power system failures during transient disturbances may be caused by unnecessary protective relay tripping rather than by the failure of a relay to operate. Adequate relay testing provides a first defense against false trips of the relay and hence improves power grid stability and prevents catastrophic bulk power system failures. The goal of this research project is to design and enhance the relay tester using a technology such as Field Programmable Gate Array (FPGA) card NI 7851. A PC based tester framework has been developed using Simulink power system model for generating signals under different conditions (faults or transient disturbances) and LabVIEW for developing the graphical user interface and configuring the FPGA. Besides, the interface system has been developed for outputting and amplifying the signals without distortion. These signals should be like the generated ones by the real power system and large enough for testing the relay’s functionality. The signals generated that have been displayed on the scope are satisfactory. Furthermore, the proposed testing system can be used for improving the performance of protective relay.

Keywords: amplifier class D, field-programmable gate array (FPGA), protective relay, tester

Procedia PDF Downloads 189
4458 Bi-Criteria Objective Network Design Model for Multi Period Multi Product Green Supply Chain

Authors: Shahul Hamid Khan, S. Santhosh, Abhinav Kumar Sharma

Abstract:

Environmental performance along with social performance is becoming vital factors for industries to achieve global standards. With a good environmental policy global industries are differentiating them from their competitors. This paper concentrates on multi stage, multi product and multi period manufacturing network. Bi-objective mathematical models for total cost and total emission for the entire forward supply chain are considered. Here five different problems are considered by varying the number of suppliers, manufacturers, and environmental levels, for illustrating the taken mathematical model. GA, and Random search are used for finding the optimal solution. The input parameters of the optimal solution are used to find the tradeoff between the initial investment by the industry and the long term benefit of the environment.

Keywords: closed loop supply chain, genetic algorithm, random search, green supply chain

Procedia PDF Downloads 526
4457 Heart Ailment Prediction Using Machine Learning Methods

Authors: Abhigyan Hedau, Priya Shelke, Riddhi Mirajkar, Shreyash Chaple, Mrunali Gadekar, Himanshu Akula

Abstract:

The heart is the coordinating centre of the major endocrine glandular structure of the body, which produces hormones that profoundly affect the operations of the body, and diagnosing cardiovascular disease is a difficult but critical task. By extracting knowledge and information about the disease from patient data, data mining is a more practical technique to help doctors detect disorders. We use a variety of machine learning methods here, including logistic regression and support vector classifiers (SVC), K-nearest neighbours Classifiers (KNN), Decision Tree Classifiers, Random Forest classifiers and Gradient Boosting classifiers. These algorithms are applied to patient data containing 13 different factors to build a system that predicts heart disease in less time with more accuracy.

Keywords: logistic regression, support vector classifier, k-nearest neighbour, decision tree, random forest and gradient boosting

Procedia PDF Downloads 26
4456 Kinetic Study of Municipal Plastic Waste

Authors: Laura Salvia Diaz Silvarrey, Anh Phan

Abstract:

Municipal Plastic Waste (MPW) comprises a mixture of thermoplastics such as high and low density polyethylene (HDPE and LDPE), polypropylene (PP), polystyrene (PS) and polyethylene terephthalate (PET). Recycling rate of these plastics is low, e.g. only 27% in 2013. The remains were incinerated or disposed in landfills. As MPW generation increases approximately 5% per annum, MPW management technologies have to be developed to comply with legislation . Pyrolysis, thermochemical decomposition, provides an excellent alternative to convert MPW into valuable resources like fuels and chemicals. Most studies on waste plastic kinetics only focused on HDPE and LDPE with a simple assumption of first order decomposition, which is not the real reaction mechanism. The aim of this study was to develop a kinetic study for each of the polymers in the MPW mixture using thermogravimetric analysis (TGA) over a range of heating rates (5, 10, 20 and 40°C/min) in N2 atmosphere and sample size of 1 – 4mm. A model-free kinetic method was applied to quantify the activation energy at each level of conversion. Kissinger–Akahira–Sunose (KAS) and Flynn–Wall–Ozawa (FWO) equations jointly with Master Plots confirmed that the activation energy was not constant along all the reaction for all the five plastic studied, showing that MPW decomposed through a complex mechanism and not by first-order kinetics. Master plots confirmed that MPW decomposed following a random scission mechanism at conversions above 40%. According to the random scission mechanism, different radicals are formed along the backbone producing the cleavage of bonds by chain scission into molecules of different lengths. The cleavage of bonds during random scission follows first-order kinetics and it is related with the conversion. When a bond is broken one part of the initial molecule becomes an unsaturated one and the other a terminal free radical. The latter can react with hydrogen from and adjacent carbon releasing another free radical and a saturated molecule or reacting with another free radical and forming an alkane. Not every time a bonds is broken a molecule is evaporated. At early stages of the reaction (conversion and temperature below 40% and 300°C), most products are not short enough to evaporate. Only at higher degrees of conversion most of cleavage of bonds releases molecules small enough to evaporate.

Keywords: kinetic, municipal plastic waste, pyrolysis, random scission

Procedia PDF Downloads 331
4455 Numerical Simulation and Laboratory Tests for Rebar Detection in Reinforced Concrete Structures using Ground Penetrating Radar

Authors: Maha Al-Soudani, Gilles Klysz, Jean-Paul Balayssac

Abstract:

The aim of this paper is to use Ground Penetrating Radar (GPR) as a non-destructive testing (NDT) method to increase its accuracy in recognizing the geometric reinforced concrete structures and in particular, the position of steel bars. This definition will help the managers to assess the state of their structures on the one hand vis-a-vis security constraints and secondly to quantify the need for maintenance and repair. Several configurations of acquisition and processing of the simulated signal were tested to propose and develop an appropriate imaging algorithm in the propagation medium to locate accurately the rebar. A subsequent experimental validation was used by testing the imaging algorithm on real reinforced concrete structures. The results indicate that, this algorithm is capable of estimating the reinforcing steel bar position to within (0-1) mm.

Keywords: GPR, NDT, Reinforced concrete structures, Rebar location.

Procedia PDF Downloads 482
4454 Uncertainty Quantification of Crack Widths and Crack Spacing in Reinforced Concrete

Authors: Marcel Meinhardt, Manfred Keuser, Thomas Braml

Abstract:

Cracking of reinforced concrete is a complex phenomenon induced by direct loads or restraints affecting reinforced concrete structures as soon as the tensile strength of the concrete is exceeded. Hence it is important to predict where cracks will be located and how they will propagate. The bond theory and the crack formulas in the actual design codes, for example, DIN EN 1992-1-1, are all based on the assumption that the reinforcement bars are embedded in homogeneous concrete without taking into account the influence of transverse reinforcement and the real stress situation. However, it can often be observed that real structures such as walls, slabs or beams show a crack spacing that is orientated to the transverse reinforcement bars or to the stirrups. In most Finite Element Analysis studies, the smeared crack approach is used for crack prediction. The disadvantage of this model is that the typical strain localization of a crack on element level can’t be seen. The crack propagation in concrete is a discontinuous process characterized by different factors such as the initial random distribution of defects or the scatter of material properties. Such behavior presupposes the elaboration of adequate models and methods of simulation because traditional mechanical approaches deal mainly with average material parameters. This paper concerned with the modelling of the initiation and the propagation of cracks in reinforced concrete structures considering the influence of transverse reinforcement and the real stress distribution in reinforced concrete (R/C) beams/plates in bending action. Therefore, a parameter study was carried out to investigate: (I) the influence of the transversal reinforcement to the stress distribution in concrete in bending mode and (II) the crack initiation in dependence of the diameter and distance of the transversal reinforcement to each other. The numerical investigations on the crack initiation and propagation were carried out with a 2D reinforced concrete structure subjected to quasi static loading and given boundary conditions. To model the uncertainty in the tensile strength of concrete in the Finite Element Analysis correlated normally and lognormally distributed random filed with different correlation lengths were generated. The paper also presents and discuss different methods to generate random fields, e.g. the Covariance Matrix Decomposition Method. For all computations, a plastic constitutive law with softening was used to model the crack initiation and the damage of the concrete in tension. It was found that the distributions of crack spacing and crack widths are highly dependent of the used random field. These distributions are validated to experimental studies on R/C panels which were carried out at the Laboratory for Structural Engineering at the University of the German Armed Forces in Munich. Also, a recommendation for parameters of the random field for realistic modelling the uncertainty of the tensile strength is given. The aim of this research was to show a method in which the localization of strains and cracks as well as the influence of transverse reinforcement on the crack initiation and propagation in Finite Element Analysis can be seen.

Keywords: crack initiation, crack modelling, crack propagation, cracks, numerical simulation, random fields, reinforced concrete, stochastic

Procedia PDF Downloads 119
4453 Nonlinear Finite Element Modeling of Deep Beam Resting on Linear and Nonlinear Random Soil

Authors: M. Seguini, D. Nedjar

Abstract:

An accuracy nonlinear analysis of a deep beam resting on elastic perfectly plastic soil is carried out in this study. In fact, a nonlinear finite element modeling for large deflection and moderate rotation of Euler-Bernoulli beam resting on linear and nonlinear random soil is investigated. The geometric nonlinear analysis of the beam is based on the theory of von Kàrmàn, where the Newton-Raphson incremental iteration method is implemented in a Matlab code to solve the nonlinear equation of the soil-beam interaction system. However, two analyses (deterministic and probabilistic) are proposed to verify the accuracy and the efficiency of the proposed model where the theory of the local average based on the Monte Carlo approach is used to analyze the effect of the spatial variability of the soil properties on the nonlinear beam response. The effect of six main parameters are investigated: the external load, the length of a beam, the coefficient of subgrade reaction of the soil, the Young’s modulus of the beam, the coefficient of variation and the correlation length of the soil’s coefficient of subgrade reaction. A comparison between the beam resting on linear and nonlinear soil models is presented for different beam’s length and external load. Numerical results have been obtained for the combination of the geometric nonlinearity of beam and material nonlinearity of random soil. This comparison highlighted the need of including the material nonlinearity and spatial variability of the soil in the geometric nonlinear analysis, when the beam undergoes large deflections.

Keywords: finite element method, geometric nonlinearity, material nonlinearity, soil-structure interaction, spatial variability

Procedia PDF Downloads 387
4452 On Four Models of a Three Server Queue with Optional Server Vacations

Authors: Kailash C. Madan

Abstract:

We study four models of a three server queueing system with Bernoulli schedule optional server vacations. Customers arriving at the system one by one in a Poisson process are provided identical exponential service by three parallel servers according to a first-come, first served queue discipline. In model A, all three servers may be allowed a vacation at one time, in Model B at the most two of the three servers may be allowed a vacation at one time, in model C at the most one server is allowed a vacation, and in model D no server is allowed a vacation. We study steady the state behavior of the four models and obtain steady state probability generating functions for the queue size at a random point of time for all states of the system. In model D, a known result for a three server queueing system without server vacations is derived.

Keywords: a three server queue, Bernoulli schedule server vacations, queue size distribution at a random epoch, steady state

Procedia PDF Downloads 280
4451 Regression Model Evaluation on Depth Camera Data for Gaze Estimation

Authors: James Purnama, Riri Fitri Sari

Abstract:

We investigate the machine learning algorithm selection problem in the term of a depth image based eye gaze estimation, with respect to its essential difficulty in reducing the number of required training samples and duration time of training. Statistics based prediction accuracy are increasingly used to assess and evaluate prediction or estimation in gaze estimation. This article evaluates Root Mean Squared Error (RMSE) and R-Squared statistical analysis to assess machine learning methods on depth camera data for gaze estimation. There are 4 machines learning methods have been evaluated: Random Forest Regression, Regression Tree, Support Vector Machine (SVM), and Linear Regression. The experiment results show that the Random Forest Regression has the lowest RMSE and the highest R-Squared, which means that it is the best among other methods.

Keywords: gaze estimation, gaze tracking, eye tracking, kinect, regression model, orange python

Procedia PDF Downloads 512
4450 An Investigation on Ultrasonic Pulse Velocity of Hybrid Fiber Reinforced Concretes

Authors: Soner Guler, Demet Yavuz, Refik Burak Taymuş, Fuat Korkut

Abstract:

Because of the easy applying and not costing too much, ultrasonic pulse velocity (UPV) is one of the most used non-destructive techniques to determine concrete characteristics along with impact-echo, Schmidt rebound hammer (SRH) and pulse-echo. This article investigates the relationship between UPV and compressive strength of hybrid fiber reinforced concretes. Water/cement ratio (w/c) was kept at 0.4 for all concrete mixes. Compressive strength of concrete was targeted at 35 MPa. UPV testing and compressive strength tests were carried out at the curing age of 28 days. The UPV of concrete containing steel fibers has been found to be higher than plain concrete for all the testing groups. It is decided that there is not a certain relationship between fiber addition and strength.

Keywords: ultrasonic pulse velocity, hybrid fiber, compressive strength, fiber

Procedia PDF Downloads 328
4449 Pre-Analytical Laboratory Performance Evaluation Utilizing Quality Indicators between Private and Government-Owned Hospitals Affiliated to University of Santo Tomas

Authors: A. J. Francisco, K. C. Gallosa, R. J. Gasacao, J. R. Ros, B. J. Viado

Abstract:

The study focuses on the use of quality indicators (QI)s based on the standards made by the (IFCC), that could effectively identify and minimize errors occurring throughout the total testing process (TTP), in order to improve patient safety. The study was conducted through a survey questionnaire that was given to a random sample of 19 respondents (eight privately-owned and eleven government-owned hospitals), mainly CMTs, MTs, and Supervisors from UST-affiliated hospitals. The pre-analytical laboratory errors, which include misidentification errors, transcription errors, sample collection errors and sample handling and transportation errors, were considered as variables according to the IFCC WG-LEPS. Data gathered were analyzed using the Mann-Whitney U test, Percentile, Linear Regression, Percentage, and Frequency. The laboratory performance of both hospitals is High level. There is no significant difference between the laboratory performance between the two stated variables. Moreover, among the four QIs, sample handling and transportation errors contributed most to the difference between the two variables. Outcomes indicate satisfactory performance between both variables. However, in order to ensure high-quality and efficient laboratory operation, constant vigilance and improvements in pre-analytical QI are still needed. Expanding the coverage of the study, the inclusion of other phases, utilization of parametric tests are recommended.

Keywords: pre-analytical phase, quality indicators, laboratory performance, pre-analytical error

Procedia PDF Downloads 119
4448 Automated Testing to Detect Instance Data Loss in Android Applications

Authors: Anusha Konduru, Zhiyong Shan, Preethi Santhanam, Vinod Namboodiri, Rajiv Bagai

Abstract:

Mobile applications are increasing in a significant amount, each to address the requirements of many users. However, the quick developments and enhancements are resulting in many underlying defects. Android apps create and handle a large variety of 'instance' data that has to persist across runs, such as the current navigation route, workout results, antivirus settings, or game state. Due to the nature of Android, an app can be paused, sent into the background, or killed at any time. If the instance data is not saved and restored between runs, in addition to data loss, partially-saved or corrupted data can crash the app upon resume or restart. However, it is difficult for the programmer to manually test this issue for all the activities. This results in the issue of data loss that the data entered by the user are not saved when there is any interruption. This issue can degrade user experience because the user needs to reenter the information each time there is an interruption. Automated testing to detect such data loss is important to improve the user experience. This research proposes a tool, DroidDL, a data loss detector for Android, which detects the instance data loss from a given android application. We have tested 395 applications and found 12 applications with the issue of data loss. This approach is proved highly accurate and reliable to find the apps with this defect, which can be used by android developers to avoid such errors.

Keywords: Android, automated testing, activity, data loss

Procedia PDF Downloads 213
4447 Text Based Shuffling Algorithm on Graphics Processing Unit for Digital Watermarking

Authors: Zayar Phyo, Ei Chaw Htoon

Abstract:

In a New-LSB based Steganography method, the Fisher-Yates algorithm is used to permute an existing array randomly. However, that algorithm performance became slower and occurred memory overflow problem while processing the large dimension of images. Therefore, the Text-Based Shuffling algorithm aimed to select only necessary pixels as hiding characters at the specific position of an image according to the length of the input text. In this paper, the enhanced text-based shuffling algorithm is presented with the powered of GPU to improve more excellent performance. The proposed algorithm employs the OpenCL Aparapi framework, along with XORShift Kernel including the Pseudo-Random Number Generator (PRNG) Kernel. PRNG is applied to produce random numbers inside the kernel of OpenCL. The experiment of the proposed algorithm is carried out by practicing GPU that it can perform faster-processing speed and better efficiency without getting the disruption of unnecessary operating system tasks.

Keywords: LSB based steganography, Fisher-Yates algorithm, text-based shuffling algorithm, OpenCL, XORShiftKernel

Procedia PDF Downloads 124
4446 Controlled Shock Response Spectrum Test on Spacecraft Subsystem Using Electrodynamic Shaker

Authors: M. Madheswaran, A. R. Prashant, S. Ramakrishna, V. Ramesh Naidu, P. Govindan, P. Aravindakshan

Abstract:

Shock Response spectrum (SRS) tests are one of the tests that are conducted on some critical systems of spacecraft as part of environmental testing. The SRS tests are conducted to simulate the pyro shocks that occur during launch phases as well as during deployment of spacecraft appendages. Some of the methods to carryout SRS tests are pyro technique method, impact hammer method, drop shock method and using electro dynamic shakers. The pyro technique, impact hammer and drop shock methods are open loop tests, whereas SRS testing using electrodynamic shaker is a controlled closed loop test. SRS testing using electrodynamic shaker offers various advantages such as simple test set up, better controllability and repeatability. However, it is important to devise a a proper test methodology so that safety of the electro dynamic shaker and that of test specimen are not compromised. This paper discusses the challenges that are involved in conducting SRS tests, shaker validation and the necessary precautions to be considered. Approach involved in choosing various test parameters like synthesis waveform, spectrum convergence level, etc., are discussed. A case study of SRS test conducted on an optical payload of Indian Geo stationary spacecraft is presented.

Keywords: maxi-max spectrum, SRS (shock response spectrum), SDOf (single degree of freedom), wavelet synthesis

Procedia PDF Downloads 330
4445 The Quality of Working Life and the Organizational Commitment of Municipal Employee in Samut Sakhon Province

Authors: Mananya Meenakorn

Abstract:

This research aims to investigate: (1) Relationship between the quality of working life and organizational commitment of municipal employee in Samut Sakhon Province. (2) To compare the quality of working life and the organizational commitment of municipal employee in Samut Sakhon Province by the gender, age, education, official experience, position, division, and income. This study is a quantitative research; data was collected by questionnaires distributed to the municipal employee in Samut Sakhon province for 241 sample by stratified random sampling. Data was analyzed by descriptive statistic including percentage, mean, standard deviation and inferential statistic including t-test, F-test and Pearson correlation for hypothesis testing. Finding showed that the quality of working life and the organizational commitment of municipal Employee in Samut Sakhon province in terms of compensation and fair has a positive correlation (r = 0.673) and the comparison of the quality of working life and organizational commitment of municipal employees in Samut Sakhon province by gender. We found that the overall difference was statistically significant at the 0.05 level and we also found stability and progress in career path and the characteristics are beneficial to society has a difference was statistically significant at the 0.01 level, and the participation and social acceptance has a difference was statistically significant at the 0.05 level.

Keywords: quality of working life, organizational commitment, municipal employee, Samut Sakhon province

Procedia PDF Downloads 262
4444 Iron Deficiency and Iron Deficiency Anaemia/Anaemia as a Diagnostic Indicator for Coeliac Disease: A Systematic Review With Meta-Analysis

Authors: Sahar Shams

Abstract:

Coeliac disease (CD) is a widely reported disease particularly in countries with predominant Caucasian populations. It presents with many signs and symptoms including iron deficiency (ID) and iron deficiency anaemia/anaemia (IDA/A). The exact association between ID, IDA/A and CD and how accurate these signs are in diagnosing CD is not fully known. This systematic review was conducted to investigate the accuracy of both ID & IDA/A as a diagnostic indicator for CD and whether it warrants point of care testing. A systematic review was performed looking at studies published in MEDLINE, Embase, Cochrane Library, and Web of Science. QUADAS-2 tool was used to assess risk of bias in each study. ROC curve and forest plots were generated as part of the meta-analysis after data extraction. 16 studies were identified in total, 13 of which were IDA/A studies and 3 ID studies. The prevalence of CD regardless of diagnostic indicator was assumed as 1%. The QUADAS-2 tool indicated most of studies as having high risk of bias. The PPV for CD was higher in those with ID than for those with IDA/A. Meta-analysis showed the overall odds of having CD is 5 times higher in individuals with ID & IDA/A. The ROC curve showed that there is definitely an association between both diagnostic indicators and CD, the association is not a particularly strong one due to great heterogeneity between studies. Whilst an association between IDA/A & ID and coeliac disease was evident, the results were not deemed significant enough to prompt coeliac disease testing in those with IDA/A & ID.

Keywords: anemia, iron deficiency anemia, coeliac disease, point of care testing

Procedia PDF Downloads 108