Search results for: estimation algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3837

Search results for: estimation algorithms

2157 Building Carbon Footprint Comparison between Building Permit, as Built, as Built with Circular Material Usage

Authors: Kadri-Ann Kertsmik, Martin Talvik, Kimmo Lylykangas, Simo Ilomets, Targo Kalamees

Abstract:

This study compares the building carbon footprint (CF) values for a case study of a private house located in a cold climate, using the Level(s) methodology. It provides a framework for measuring the environmental performance of buildings throughout their life cycle, taking into account various factors. The study presents the results of the three scenarios, comparing their carbon emissions and highlighting the benefits of circular material usage. The construction process was thoroughly documented, and all materials and components (including minuscule mechanical fasteners, each meter of cable, a kilogram of mortar, and the component of HVAC systems, among other things) delivered to the construction site were noted. Transportation distances of each delivery, the fuel consumption of construction machines, and electricity consumption for temporary heating and electrical tools were also monitored. Using the detailed data on material and energy resources, the CF was calculated for two scenarios: one where circular material usage was applied and another where virgin materials were used instead of reused ones. The results were compared with the CF calculated based on the building permit design model using the Level(s) methodology. To study the range of possible results in the early stage of CF assessment, the same building permit design was given to several experts. Results showed that embodied carbon values for a built scenario were significantly lower than the values predicted by the building permit stage as a result of more precise material quantities, as the calculation methodology is designed to overestimate the CF. Moreover, designers made an effort to reduce the building's CF by reusing certain materials such as ceramic tiles, lightweight concrete blocks, and timber during the construction process. However, in a cold climate context where operational energy (B6) continues to dominate, the total building CF value changes between the three scenarios were less significant. The calculation for the building permit project was performed by several experts, and CF results were in the same range. It alludes that, for the first estimation of preliminary building CF, using average values proves to be an appropriate method for the Estonian national carbon footprint estimation phase during building permit application. The study also identified several opportunities for reducing the carbon footprint of the building, such as reusing materials from other construction sites, preferring local material producers, and reducing wastage on site. The findings suggest that using circular materials can significantly reduce the carbon footprint of buildings. Overall, the study highlights the importance of using a comprehensive approach to measure the environmental performance of buildings, taking into account both the project and the actually built house. It also emphasises the need for ongoing monitoring for designing the building and construction site waste. The study also gives some examples of how to enable future circularity of building components and materials, e.g., building in layers, using wood as untreated, etc.

Keywords: carbon footprint, circular economy, sustainable construction, level(s) methodology

Procedia PDF Downloads 87
2156 Particle Swarm Optimization Based Method for Minimum Initial Marking in Labeled Petri Nets

Authors: Hichem Kmimech, Achref Jabeur Telmoudi, Lotfi Nabli

Abstract:

The estimation of the initial marking minimum (MIM) is a crucial problem in labeled Petri nets. In the case of multiple choices, the search for the initial marking leads to a problem of optimization of the minimum allocation of resources with two constraints. The first concerns the firing sequence that could be legal on the initial marking with respect to the firing vector. The second deals with the total number of tokens that can be minimal. In this article, the MIM problem is solved by the meta-heuristic particle swarm optimization (PSO). The proposed approach presents the advantages of PSO to satisfy the two previous constraints and find all possible combinations of minimum initial marking with the best computing time. This method, more efficient than conventional ones, has an excellent impact on the resolution of the MIM problem. We prove through a set of definitions, lemmas, and examples, the effectiveness of our approach.

Keywords: marking, production system, labeled Petri nets, particle swarm optimization

Procedia PDF Downloads 179
2155 Estimation of Location and Scale Parameters of Extended Exponential Distribution Based on Record Statistics

Authors: E. Krishna

Abstract:

An Extended form of exponential distribution using Marshall and Olkin method is introduced.The location scale family of these distributions is considered. For location scale free family, exact expressions for single and product moments of upper record statistics are derived. The mean, variance and covariance of record values are computed for various values of the shape parameter. Using these the BLUE's of location and scale parameters are derived.The variances and covariance of estimates are obtained.Through Monte Carlo simulation the con dence intervals for location and scale parameters are constructed.The Best liner unbiased Predictor (BLUP) of future records are also discussed.

Keywords: BLUE, BLUP, con dence interval, Marshall-Olkin distribution, Monte Carlo simulation, prediction of future records, record statistics

Procedia PDF Downloads 417
2154 Model Predictive Control (MPC) and Proportional-Integral-Derivative (PID) Control of Quadcopters: A Comparative Analysis

Authors: Anel Hasić, Naser Prljača

Abstract:

In the domain of autonomous or piloted flights, the accurate control of quadrotor trajectories is of paramount significance for large numbers of tasks. These adaptable aerial platforms find applications that span from high-precision aerial photography and surveillance to demanding search and rescue missions. Among the fundamental challenges confronting quadrotor operation is the demand for accurate following of desired flight paths. To address this control challenge, among others, two celebrated well-established control strategies have emerged as noteworthy contenders: Model Predictive Control (MPC) and Proportional-Integral-Derivative (PID) control. In this work, we focus on the extensive examination of MPC and PID control techniques by using comprehensive simulation studies in MATLAB/Simulink. Intensive simulation results demonstrate the performance of the studied control algorithms.

Keywords: MATLAB, MPC, PID, quadcopter, simulink

Procedia PDF Downloads 70
2153 Evaluating Classification with Efficacy Metrics

Authors: Guofan Shao, Lina Tang, Hao Zhang

Abstract:

The values of image classification accuracy are affected by class size distributions and classification schemes, making it difficult to compare the performance of classification algorithms across different remote sensing data sources and classification systems. Based on the term efficacy from medicine and pharmacology, we have developed the metrics of image classification efficacy at the map and class levels. The novelty of this approach is that a baseline classification is involved in computing image classification efficacies so that the effects of class statistics are reduced. Furthermore, the image classification efficacies are interpretable and comparable, and thus, strengthen the assessment of image data classification methods. We use real-world and hypothetical examples to explain the use of image classification efficacies. The metrics of image classification efficacy meet the critical need to rectify the strategy for the assessment of image classification performance as image classification methods are becoming more diversified.

Keywords: accuracy assessment, efficacy, image classification, machine learning, uncertainty

Procedia PDF Downloads 211
2152 Optimized Cluster Head Selection Algorithm Based on LEACH Protocol for Wireless Sensor Networks

Authors: Wided Abidi, Tahar Ezzedine

Abstract:

Low-Energy Adaptive Clustering Hierarchy (LEACH) has been considered as one of the effective hierarchical routing algorithms that optimize energy and prolong the lifetime of network. Since the selection of Cluster Head (CH) in LEACH is carried out randomly, in this paper, we propose an approach of electing CH based on LEACH protocol. In other words, we present a formula for calculating the threshold responsible for CH election. In fact, we adopt three principle criteria: the remaining energy of node, the number of neighbors within cluster range and the distance between node and CH. Simulation results show that our proposed approach beats LEACH protocol in regards of prolonging the lifetime of network and saving residual energy.

Keywords: wireless sensors networks, LEACH protocol, cluster head election, energy efficiency

Procedia PDF Downloads 330
2151 Fuzzy Decision Support System for Human-Realistic Overtaking in Railway Traffic Simulations

Authors: Tomáš Vyčítal

Abstract:

In a simulation model of a railway system it is important, besides other crucial algorithms, to have correct behaviour of train overtaking in stochastic conditions. This problem is being addressed in many simulation tools focused on railway traffic, however these are not very human-realistic. The goal of this paper is to create a more human-realistic overtaking decision support system for the use in railway traffic simulations. A fuzzy system has been chosen for this task as fuzzy systems are well-suited for human-like decision making. The fuzzy system designed takes into account timetables, train positions, delays and buffer times as inputs and provides an instruction to overtake or not overtake.

Keywords: decision-making support, fuzzy systems, simulation, railway, transport

Procedia PDF Downloads 140
2150 Poisoning Admission in Pediatrics Benghazi Hospital in Libya: Three Years Review of Medical Record

Authors: Mudafara Bengleil

Abstract:

Estimation of the magnitude and causes of poisoning was the objective of the current study. A retrospective study of medical records of all poisoning children admitted to Benghazi Children Hospital in Libya from January 2008 up to December 2010. Number of children admitted was 244; the age ranged from less than one to 13 years old. Most of cases were admitted with mild symptom and the majority of them were boys. Only few cases admitted to intensive care unit and there was no mortality recorded through the period of study. Age group 1 to 3 years (50.8%) had the highest frequency of admission and the peak of admission was during summer. The most common cause of admission was due to ingestion of medication (53.69%), House hold product exposure (26.64%) was the second causes of admission while, 19.67% of admissions were due to Food poisoning. Almost all admitted cases were accidental and medicines were the most consumed substances in addition, improper storage of toxic agents were the first risk factor of poisoning. Present results indicated that, children poisoning seems to be a common pediatric care problem which need to control and prevent.

Keywords: poisoning, children, hospital, medical

Procedia PDF Downloads 421
2149 Elimination of Low Order Harmonics in Multilevel Inverter Using Nature-Inspired Metaheuristic Algorithm

Authors: N. Ould Cherchali, A. Tlemçani, M. S. Boucherit, A. Morsli

Abstract:

Nature-inspired metaheuristic algorithms, particularly those founded on swarm intelligence, have attracted much attention over the past decade. Firefly algorithm has appeared in approximately seven years ago, its literature has enlarged considerably with different applications. It is inspired by the behavior of fireflies. The aim of this paper is the application of firefly algorithm for solving a nonlinear algebraic system. This resolution is needed to study the Selective Harmonic Eliminated Pulse Width Modulation strategy (SHEPWM) to eliminate the low order harmonics; results have been applied on multilevel inverters. The final results from simulations indicate the elimination of the low order harmonics as desired. Finally, experimental results are presented to confirm the simulation results and validate the efficaciousness of the proposed approach.

Keywords: firefly algorithm, metaheuristic algorithm, multilevel inverter, SHEPWM

Procedia PDF Downloads 148
2148 Mixed Treatment (Physical-Chemical and Biological) of Ouled Fayet Landfill Leachates

Authors: O. Balamane-Zizi, L. M. Rouidi, A. Boukhrissa, N. Daas, H. Ait-amar

Abstract:

The objective of this study was to test the possibility of a mixed treatment (physical-chemical and biological) of Ouled Fayet leachates which date of 10 years and has a large fraction of hard COD that can be reduced by coagulation-flocculation. Previous batch tests showed the possibility of applying the physical-chemical and biological treatments separately; the removal efficiencies obtained in this case were not interesting. We propose, therefore, to test the possibility of a combined treatment, in order to improve the quality of the leachates. Estimation of the treatment’s effectiveness was done by analysis of some pollution parameters such as COD, suspended solids, and heavy metals (particularly iron and nickel). The main results obtained after the combination of treatments, show reduction rate of about 63% for COD, 73% for suspended solids and 80% for iron and nickel. We also noted an improvement in the turbidity of treated leachates.

Keywords: landfill leachates, COD, physical-chemical treatment, biological treatment

Procedia PDF Downloads 474
2147 Measuring Banking Risk

Authors: Mike Tsionas

Abstract:

The paper develops new indices of financial stability based on an explicit model of expected utility maximization by financial institutions subject to the classical technology restrictions of neoclassical production theory. The model can be estimated using standard econometric techniques, like GMM for dynamic panel data and latent factor analysis for the estimation of co-variance matrices. An explicit functional form for the utility function is not needed and we show how measures of risk aversion and prudence (downside risk aversion) can be derived and estimated from the model. The model is estimated using data for Eurozone countries and we focus particularly on (i) the use of the modeling approach as an “early warning mechanism”, (ii) the bank- and country-specific estimates of risk aversion and prudence (downside risk aversion), and (iii) the derivation of a generalized measure of risk that relies on loan-price uncertainty.

Keywords: financial stability, banking, expected utility maximization, sub-prime crisis, financial crisis, eurozone, PIIGS

Procedia PDF Downloads 349
2146 Localization Mobile Beacon Using RSSI

Authors: Sallama Resen, Celal Öztürk

Abstract:

Distance estimation between tow nodes has wide scope of surveillance and tracking applications. This paper suggests a Bluetooth Low Energy (BLE) technology as a media for transceiver and receiver signal in small indoor areas. As an example, BLE communication technologies used in child safety domains. Local network is designed to detect child position in indoor school area consisting Mobile Beacons (MB), Access Points (AP) and Smart Phones (SP) where MBs stuck in children’s shoes as wearable sensors. This paper presents a technique that can detect mobile beacons’ position and help finding children’s location within dynamic environment. By means of bluetooth beacons that are attached to child’s shoes, the distance between the MB and teachers SP is estimated with an accuracy of less than one meter. From the simulation results, it is shown that high accuracy of position coordinates are achieved for multi-mobile beacons in different environments.

Keywords: bluetooth low energy, child safety, mobile beacons, received signal strength

Procedia PDF Downloads 346
2145 Plant Disease Detection Using Image Processing and Machine Learning

Authors: Sanskar, Abhinav Pal, Aryush Gupta, Sushil Kumar Mishra

Abstract:

One of the critical and tedious assignments in agricultural practices is the detection of diseases on vegetation. Agricultural production is very important in today’s economy because plant diseases are common, and early detection of plant diseases is important in agriculture. Automatic detection of such early diseases is useful because it reduces control efforts in large productive farms. Using digital image processing and machine learning algorithms, this paper presents a method for plant disease detection. Detection of the disease occurs on different leaves of the plant. The proposed system for plant disease detection is simple and computationally efficient, requiring less time than learning-based approaches. The accuracy of various plant and foliar diseases is calculated and presented in this paper.

Keywords: plant diseases, machine learning, image processing, deep learning

Procedia PDF Downloads 10
2144 Approach on Conceptual Design and Dimensional Synthesis of the Linear Delta Robot for Additive Manufacturing

Authors: Efrain Rodriguez, Cristhian Riano, Alberto Alvares

Abstract:

In recent years, robots manipulators with parallel architectures are used in additive manufacturing processes – 3D printing. These robots have advantages such as speed and lightness that make them suitable to help with the efficiency and productivity of these processes. Consequently, the interest for the development of parallel robots for additive manufacturing applications has increased. This article deals with the conceptual design and dimensional synthesis of the linear delta robot for additive manufacturing. Firstly, a methodology based on structured processes for the development of products through the phases of informational design, conceptual design and detailed design is adopted: a) In the informational design phase the Mudge diagram and the QFD matrix are used to aid a set of technical requirements, to define the form, functions and features of the robot. b) In the conceptual design phase, the functional modeling of the system through of an IDEF0 diagram is performed, and the solution principles for the requirements are formulated using a morphological matrix. This phase includes the description of the mechanical, electro-electronic and computational subsystems that constitute the general architecture of the robot. c) In the detailed design phase, a digital model of the robot is drawn on CAD software. A list of commercial and manufactured parts is detailed. Tolerances and adjustments are defined for some parts of the robot structure. The necessary manufacturing processes and tools are also listed, including: milling, turning and 3D printing. Secondly, a dimensional synthesis method applied on design of the linear delta robot is presented. One of the most important key factors in the design of a parallel robot is the useful workspace, which strongly depends on the joint space, the dimensions of the mechanism bodies and the possible interferences between these bodies. The objective function is based on the verification of the kinematic model for a prescribed cylindrical workspace, considering geometric constraints that possibly lead to singularities of the mechanism. The aim is to determine the minimum dimensional parameters of the mechanism bodies for the proposed workspace. A method based on genetic algorithms was used to solve this problem. The method uses a cloud of points with the cylindrical shape of the workspace and checks the kinematic model for each of the points within the cloud. The evolution of the population (point cloud) provides the optimal parameters for the design of the delta robot. The development process of the linear delta robot with optimal dimensions for additive manufacture is presented. The dimensional synthesis enabled to design the mechanism of the delta robot in function of the prescribed workspace. Finally, the implementation of the robotic platform developed based on a linear delta robot in an additive manufacturing application using the Fused Deposition Modeling (FDM) technique is presented.

Keywords: additive manufacturing, delta parallel robot, dimensional synthesis, genetic algorithms

Procedia PDF Downloads 190
2143 Efficient Subgoal Discovery for Hierarchical Reinforcement Learning Using Local Computations

Authors: Adrian Millea

Abstract:

In hierarchical reinforcement learning, one of the main issues encountered is the discovery of subgoal states or options (which are policies reaching subgoal states) by partitioning the environment in a meaningful way. This partitioning usually requires an expensive global clustering operation or eigendecomposition of the Laplacian of the states graph. We propose a local solution to this issue, much more efficient than algorithms using global information, which successfully discovers subgoal states by computing a simple function, which we call heterogeneity for each state as a function of its neighbors. Moreover, we construct a value function using the difference in heterogeneity from one step to the next, as reward, such that we are able to explore the state space much more efficiently than say epsilon-greedy. The same principle can then be applied to higher level of the hierarchy, where now states are subgoals discovered at the level below.

Keywords: exploration, hierarchical reinforcement learning, locality, options, value functions

Procedia PDF Downloads 171
2142 Nonparametric Specification Testing for the Drift of the Short Rate Diffusion Process Using a Panel of Yields

Authors: John Knight, Fuchun Li, Yan Xu

Abstract:

Based on a new method of the nonparametric estimator of the drift function, we propose a consistent test for the parametric specification of the drift function in the short rate diffusion process using observations from a panel of yields. The test statistic is shown to follow an asymptotic normal distribution under the null hypothesis that the parametric drift function is correctly specified, and converges to infinity under the alternative. Taking the daily 7-day European rates as a proxy of the short rate, we use our test to examine whether the drift of the short rate diffusion process is linear or nonlinear, which is an unresolved important issue in the short rate modeling literature. The testing results indicate that none of the drift functions in this literature adequately captures the dynamics of the drift, but nonlinear specification performs better than the linear specification.

Keywords: diffusion process, nonparametric estimation, derivative security price, drift function and volatility function

Procedia PDF Downloads 368
2141 Special Case of Trip Distribution Model and Its Use for Estimation of Detailed Transport Demand in the Czech Republic

Authors: Jiri Dufek

Abstract:

The national model of the Czech Republic has been modified in a detailed way to get detailed travel demand in the municipality level (cities, villages over 300 inhabitants). As a technique for this detailed modelling, three-dimensional procedure for calibrating gravity models, was used. Besides of zone production and attraction, which is usual in gravity models, the next additional parameter for trip distribution was introduced. Usually it is called by “third dimension”. In the model, this parameter is a demand between regions. The distribution procedure involved calculation of appropriate skim matrices and its multiplication by three coefficients obtained by iterative balancing of production, attraction and third dimension. This type of trip distribution was processed in R-project and the results were used in the Czech Republic transport model, created in PTV Vision. This process generated more precise results in local level od the model (towns, villages)

Keywords: trip distribution, three dimension, transport model, municipalities

Procedia PDF Downloads 131
2140 A Multi-Agent Urban Traffic Simulator for Generating Autonomous Driving Training Data

Authors: Florin Leon

Abstract:

This paper describes a simulator of traffic scenarios tailored to facilitate autonomous driving model training for urban environments. With the rising prominence of self-driving vehicles, the need for diverse datasets is very important. The proposed simulator provides a flexible framework that allows the generation of custom scenarios needed for the validation and enhancement of trajectory prediction algorithms. Its controlled yet dynamic environment addresses the challenges associated with real-world data acquisition and ensures adaptability to diverse driving scenarios. By providing an adaptable solution for scenario creation and algorithm testing, this tool proves to be a valuable resource for advancing autonomous driving technology that aims to ensure safe and efficient self-driving vehicles.

Keywords: autonomous driving, car simulator, machine learning, model training, urban simulation environment

Procedia PDF Downloads 62
2139 Simulation Data Summarization Based on Spatial Histograms

Authors: Jing Zhao, Yoshiharu Ishikawa, Chuan Xiao, Kento Sugiura

Abstract:

In order to analyze large-scale scientific data, research on data exploration and visualization has gained popularity. In this paper, we focus on the exploration and visualization of scientific simulation data, and define a spatial V-Optimal histogram for data summarization. We propose histogram construction algorithms based on a general binary hierarchical partitioning as well as a more specific one, the l-grid partitioning. For effective data summarization and efficient data visualization in scientific data analysis, we propose an optimal algorithm as well as a heuristic algorithm for histogram construction. To verify the effectiveness and efficiency of the proposed methods, we conduct experiments on the massive evacuation simulation data.

Keywords: simulation data, data summarization, spatial histograms, exploration, visualization

Procedia PDF Downloads 177
2138 Estimation of PM2.5 Emissions and Source Apportionment Using Receptor and Dispersion Models

Authors: Swetha Priya Darshini Thammadi, Sateesh Kumar Pisini, Sanjay Kumar Shukla

Abstract:

Source apportionment using Dispersion model depends primarily on the quality of Emission Inventory. In the present study, a CMB receptor model has been used to identify the sources of PM2.5, while the AERMOD dispersion model has been used to account for missing sources of PM2.5 in the Emission Inventory. A statistical approach has been developed to quantify the missing sources not considered in the Emission Inventory. The inventory of each grid was improved by adjusting emissions based on road lengths and deficit in measured and modelled concentrations. The results showed that in CMB analyses, fugitive sources - soil and road dust - contribute significantly to ambient PM2.5 pollution. As a result, AERMOD significantly underestimated the ambient air concentration at most locations. The revised Emission Inventory showed a significant improvement in AERMOD performance which is evident through statistical tests.

Keywords: CMB, GIS, AERMOD, PM₂.₅, fugitive, emission inventory

Procedia PDF Downloads 340
2137 Simultaneous Optimization of Design and Maintenance through a Hybrid Process Using Genetic Algorithms

Authors: O. Adjoul, A. Feugier, K. Benfriha, A. Aoussat

Abstract:

In general, issues related to design and maintenance are considered in an independent manner. However, the decisions made in these two sets influence each other. The design for maintenance is considered an opportunity to optimize the life cycle cost of a product, particularly in the nuclear or aeronautical field, where maintenance expenses represent more than 60% of life cycle costs. The design of large-scale systems starts with product architecture, a choice of components in terms of cost, reliability, weight and other attributes, corresponding to the specifications. On the other hand, the design must take into account maintenance by improving, in particular, real-time monitoring of equipment through the integration of new technologies such as connected sensors and intelligent actuators. We noticed that different approaches used in the Design For Maintenance (DFM) methods are limited to the simultaneous characterization of the reliability and maintainability of a multi-component system. This article proposes a method of DFM that assists designers to propose dynamic maintenance for multi-component industrial systems. The term "dynamic" refers to the ability to integrate available monitoring data to adapt the maintenance decision in real time. The goal is to maximize the availability of the system at a given life cycle cost. This paper presents an approach for simultaneous optimization of the design and maintenance of multi-component systems. Here the design is characterized by four decision variables for each component (reliability level, maintainability level, redundancy level, and level of monitoring data). The maintenance is characterized by two decision variables (the dates of the maintenance stops and the maintenance operations to be performed on the system during these stops). The DFM model helps the designers choose technical solutions for the large-scale industrial products. Large-scale refers to the complex multi-component industrial systems and long life-cycle, such as trains, aircraft, etc. The method is based on a two-level hybrid algorithm for simultaneous optimization of design and maintenance, using genetic algorithms. The first level is to select a design solution for a given system that considers the life cycle cost and the reliability. The second level consists of determining a dynamic and optimal maintenance plan to be deployed for a design solution. This level is based on the Maintenance Free Operating Period (MFOP) concept, which takes into account the decision criteria such as, total reliability, maintenance cost and maintenance time. Depending on the life cycle duration, the desired availability, and the desired business model (sales or rental), this tool provides visibility of overall costs and optimal product architecture.

Keywords: availability, design for maintenance (DFM), dynamic maintenance, life cycle cost (LCC), maintenance free operating period (MFOP), simultaneous optimization

Procedia PDF Downloads 118
2136 Short and Long Crack Growth Behavior in Ferrite Bainite Dual Phase Steels

Authors: Ashok Kumar, Shiv Brat Singh, Kalyan Kumar Ray

Abstract:

There is growing awareness to design steels against fatigue damage Ferrite martensite dual-phase steels are known to exhibit favourable mechanical properties like good strength, ductility, toughness, continuous yielding, and high work hardening rate. However, dual-phase steels containing bainite as second phase are potential alternatives for ferrite martensite steels for certain applications where good fatigue property is required. Fatigue properties of dual phase steels are popularly assessed by the nature of variation of crack growth rate (da/dN) with stress intensity factor range (∆K), and the magnitude of fatigue threshold (∆Kth) for long cracks. There exists an increased emphasis to understand not only the long crack fatigue behavior but also short crack growth behavior of ferrite bainite dual phase steels. The major objective of this report is to examine the influence of microstructures on the short and long crack growth behavior of a series of developed dual-phase steels with varying amounts of bainite and. Three low carbon steels containing Nb, Cr and Mo as microalloying elements steels were selected for making ferrite-bainite dual-phase microstructures by suitable heat treatments. The heat treatment consisted of austenitizing the steel at 1100°C for 20 min, cooling at different rates in air prior to soaking these in a salt bath at 500°C for one hour, and finally quenching in water. Tensile tests were carried out on 25 mm gauge length specimens with 5 mm diameter using nominal strain rate 0.6x10⁻³ s⁻¹ at room temperature. Fatigue crack growth studies were made on a recently developed specimen configuration using a rotating bending machine. The crack growth was monitored by interrupting the test and observing the specimens under an optical microscope connected to an Image analyzer. The estimated crack lengths (a) at varying number of cycles (N) in different fatigue experiments were analyzed to obtain log da/dN vs. log °∆K curves for determining ∆Kthsc. The microstructural features of these steels have been characterized and their influence on the near threshold crack growth has been examined. This investigation, in brief, involves (i) the estimation of ∆Kthsc and (ii) the examination of the influence of microstructure on short and long crack fatigue threshold. The maximum fatigue threshold values obtained from short crack growth experiments on various specimens of dual-phase steels containing different amounts of bainite are found to increase with increasing bainite content in all the investigated steels. The variations of fatigue behavior of the selected steel samples have been explained with the consideration of varying amounts of the constituent phases and their interactions with the generated microstructures during cyclic loading. Quantitative estimation of the different types of fatigue crack paths indicates that the propensity of a crack to pass through the interfaces depends on the relative amount of the microstructural constituents. The fatigue crack path is found to be predominantly intra-granular except for the ones containing > 70% bainite in which it is predominantly inter-granular.

Keywords: bainite, dual phase steel, fatigue crack growth rate, long crack fatigue threshold, short crack fatigue threshold

Procedia PDF Downloads 204
2135 Computerized Adaptive Testing for Ipsative Tests with Multidimensional Pairwise-Comparison Items

Authors: Wen-Chung Wang, Xue-Lan Qiu

Abstract:

Ipsative tests have been widely used in vocational and career counseling (e.g., the Jackson Vocational Interest Survey). Pairwise-comparison items are a typical item format of ipsative tests. When the two statements in a pairwise-comparison item measure two different constructs, the item is referred to as a multidimensional pairwise-comparison (MPC) item. A typical MPC item would be: Which activity do you prefer? (A) playing with young children, or (B) working with tools and machines. These two statements aim at the constructs of social interest and investigative interest, respectively. Recently, new item response theory (IRT) models for ipsative tests with MPC items have been developed. Among them, the Rasch ipsative model (RIM) deserves special attention because it has good measurement properties, in which the log-odds of preferring statement A to statement B are defined as a competition between two parts: the sum of a person’s latent trait to which statement A is measuring and statement A’s utility, and the sum of a person’s latent trait to which statement B is measuring and statement B’s utility. The RIM has been extended to polytomous responses, such as preferring statement A strongly, preferring statement A, preferring statement B, and preferring statement B strongly. To promote the new initiatives, in this study we developed computerized adaptive testing algorithms for MFC items and evaluated their performance using simulations and two real tests. Both the RIM and its polytomous extension are multidimensional, which calls for multidimensional computerized adaptive testing (MCAT). A particular issue in MCAT for MPC items is the within-person statement exposure (WPSE); that is, a respondent may keep seeing the same statement (e.g., my life is empty) for many times, which is certainly annoying. In this study, we implemented two methods to control the WPSE rate. In the first control method, items would be frozen when their statements had been administered more than a prespecified times. In the second control method, a random component was added to control the contribution of the information at different stages of MCAT. The second control method was found to outperform the first control method in our simulation studies. In addition, we investigated four item selection methods: (a) random selection (as a baseline), (b) maximum Fisher information method without WPSE control, (c) maximum Fisher information method with the first control method, and (d) maximum Fisher information method with the second control method. These four methods were applied to two real tests: one was a work survey with dichotomous MPC items and the other is a career interests survey with polytomous MPC items. There were three dependent variables: the bias and root mean square error across person measures, and measurement efficiency which was defined as the number of items needed to achieve the same degree of test reliability. Both applications indicated that the proposed MCAT algorithms were successful and there was no loss in measurement proficiency when the control methods were implemented, and among the four methods, the last method performed the best.

Keywords: computerized adaptive testing, ipsative tests, item response theory, pairwise comparison

Procedia PDF Downloads 246
2134 Designing State Feedback Multi-Target Controllers by the Use of Particle Swarm Optimization Algorithm

Authors: Seyedmahdi Mousavihashemi

Abstract:

One of the most important subjects of interest in researches is 'improving' which result in various algorithms. In so many geometrical problems we are faced with target functions which should be optimized. In group practices, all the functions’ cooperation lead to convergence. In the study, the optimization algorithm of dense particles is used. Usage of the algorithm improves the given performance norms. The results reveal that usage of swarm algorithm for reinforced particles in designing state feedback improves the given performance norm and in optimized designing of multi-target state feedback controlling, the network will maintain its bearing structure. The results also show that PSO is usable for optimization of state feedback controllers.

Keywords: multi-objective, enhanced, feedback, optimization, algorithm, particle, design

Procedia PDF Downloads 500
2133 Simulation-Based Unmanned Surface Vehicle Design Using PX4 and Robot Operating System With Kubernetes and Cloud-Native Tooling

Authors: Norbert Szulc, Jakub Wilk, Franciszek Górski

Abstract:

This paper presents an approach for simulating and testing robotic systems based on PX4, using a local Kubernetes cluster. The approach leverages modern cloud-native tools and runs on single-board computers. Additionally, this solution enables the creation of datasets for computer vision and the evaluation of control system algorithms in an end-to-end manner. This paper compares this approach to method commonly used Docker based approach. This approach was used to develop simulation environment for an unmanned surface vehicle (USV) for RoboBoat 2023 by running a containerized configuration of the PX4 Open-source Autopilot connected to ROS and the Gazebo simulation environment.

Keywords: cloud computing, Kubernetes, single board computers, simulation, ROS

Procedia PDF Downloads 77
2132 Estimating Tree Height and Forest Classification from Multi Temporal Risat-1 HH and HV Polarized Satellite Aperture Radar Interferometric Phase Data

Authors: Saurav Kumar Suman, P. Karthigayani

Abstract:

In this paper the height of the tree is estimated and forest types is classified from the multi temporal RISAT-1 Horizontal-Horizontal (HH) and Horizontal-Vertical (HV) Polarised Satellite Aperture Radar (SAR) data. The novelty of the proposed project is combined use of the Back-scattering Coefficients (Sigma Naught) and the Coherence. It uses Water Cloud Model (WCM). The approaches use two main steps. (a) Extraction of the different forest parameter data from the Product.xml, BAND-META file and from Grid-xxx.txt file come with the HH & HV polarized data from the ISRO (Indian Space Research Centre). These file contains the required parameter during height estimation. (b) Calculation of the Vegetation and Ground Backscattering, Coherence and other Forest Parameters. (c) Classification of Forest Types using the ENVI 5.0 Tool and ROI (Region of Interest) calculation.

Keywords: RISAT-1, classification, forest, SAR data

Procedia PDF Downloads 407
2131 Axle Load Estimation of Moving Vehicles Using BWIM Technique

Authors: Changgil Lee, Seunghee Park

Abstract:

Although vehicle driving test for the development of BWIM system is necessary, but it needs much cost and time in addition application of various driving condition. Thus, we need the numerical-simulation method resolving the cost and time problems of vehicle driving test and the way of measuring response of bridge according to the various driving condition. Using the precision analysis model reflecting the dynamic characteristic is contributed to increase accuracy in numerical simulation. In this paper, we conduct a numerical simulation to apply precision analysis model, which reflects the dynamic characteristic of bridge using Bridge Weigh-in-Motion technique and suggest overload vehicle enforcement technology using precision analysis model.

Keywords: bridge weigh-in-motion(BWIM) system, precision analysis model, dynamic characteristic of bridge, numerical simulation

Procedia PDF Downloads 292
2130 Optical Flow Localisation and Appearance Mapping (OFLAAM) for Long-Term Navigation

Authors: Daniel Pastor, Hyo-Sang Shin

Abstract:

This paper presents a novel method to use optical flow navigation for long-term navigation. Unlike standard SLAM approaches for augmented reality, OFLAAM is designed for Micro Air Vehicles (MAV). It uses an optical flow camera pointing downwards, an IMU and a monocular camera pointing frontwards. That configuration avoids the expensive mapping and tracking of the 3D features. It only maps these features in a vocabulary list by a localization module to tackle the loss of the navigation estimation. That module, based on the well-established algorithm DBoW2, will be also used to close the loop and allow long-term navigation in confined areas. That combination of high-speed optical flow navigation with a low rate localization algorithm allows fully autonomous navigation for MAV, at the same time it reduces the overall computational load. This framework is implemented in ROS (Robot Operating System) and tested attached to a laptop. A representative scenarios is used to analyse the performance of the system.

Keywords: vision, UAV, navigation, SLAM

Procedia PDF Downloads 606
2129 Analysis of Various Copy Move Image Forgery Techniques for Better Detection Accuracy

Authors: Grishma D. Solanki, Karshan Kandoriya

Abstract:

In modern era of information age, digitalization has revolutionized like never before. Powerful computers, advanced photo editing software packages and high resolution capturing devices have made manipulation of digital images incredibly easy. As per as image forensics concerns, one of the most actively researched area are detection of copy move forgeries. Higher computational complexity is one of the major component of existing techniques to detect such tampering. Moreover, copy move forgery is usually performed in three steps. First, copying of a region in an image then pasting the same one in the same respective image and finally doing some post-processing like rotation, scaling, shift, noise, etc. Consequently, pseudo Zernike moment is used as a features extraction method for matching image blocks and as a primary factor on which performance of detection algorithms depends.

Keywords: copy-move image forgery, digital forensics, image forensics, image forgery

Procedia PDF Downloads 288
2128 Stock Price Prediction Using Time Series Algorithms

Authors: Sumit Sen, Sohan Khedekar, Umang Shinde, Shivam Bhargava

Abstract:

This study has been undertaken to investigate whether the deep learning models are able to predict the future stock prices by training the model with the historical stock price data. Since this work required time series analysis, various models are present today to perform time series analysis such as Recurrent Neural Network LSTM, ARIMA and Facebook Prophet. Applying these models the movement of stock price of stocks are predicted and also tried to provide the future prediction of the stock price of a stock. Final product will be a stock price prediction web application that is developed for providing the user the ease of analysis of the stocks and will also provide the predicted stock price for the next seven days.

Keywords: Autoregressive Integrated Moving Average, Deep Learning, Long Short Term Memory, Time-series

Procedia PDF Downloads 142