Search results for: noise estimation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2942

Search results for: noise estimation

1442 A Comprehensive Procedure of Spatial Panel Modelling with R, A Study of Agricultural Productivity Growth of the 38 East Java’s Regencies/Municipalities

Authors: Rahma Fitriani, Zerlita Fahdha Pusdiktasari, Herman Cahyo Diartho

Abstract:

Spatial panel model is commonly used to specify more complicated behavior of economic agent distributed in space at an individual-spatial unit level. There are several spatial panel models which can be adapted based on certain assumptions. A package called splm in R has several functions, ranging from the estimation procedure, specification tests, and model selection tests. In the absence of prior assumptions, a comprehensive procedure which utilizes the available functions in splm must be formed, which is the objective of this study. In this way, the best specification and model can be fitted based on data. The implementation of the procedure works well. It specifies SARAR-FE as the best model for agricultural productivity growth of the 38 East Java’s Regencies/Municipalities.

Keywords: spatial panel, specification, splm, agricultural productivity growth

Procedia PDF Downloads 167
1441 Performance Analysis of MIMO-OFDM Using Convolution Codes with QAM Modulation

Authors: I Gede Puja Astawa, Yoedy Moegiharto, Ahmad Zainudin, Imam Dui Agus Salim, Nur Annisa Anggraeni

Abstract:

Performance of Orthogonal Frequency Division Multiplexing (OFDM) system can be improved by adding channel coding (error correction code) to detect and correct the errors that occur during data transmission. One can use the convolution code. This paper presents performance of OFDM using Space Time Block Codes (STBC) diversity technique use QAM modulation with code rate 1/2. The evaluation is done by analyzing the value of Bit Error Rate (BER) vs. Energy per Bit to Noise Power Spectral Density Ratio (Eb/No). This scheme is conducted 256 sub-carrier which transmits Rayleigh multipath channel in OFDM system. To achieve a BER of 10-3 is required 30 dB SNR in SISO-OFDM scheme. For 2x2 MIMO-OFDM scheme requires 10 dB to achieve a BER of 10-3. For 4x4 MIMO-OFDM scheme requires 5 dB while adding convolution in a 4x4 MIMO-OFDM can improve performance up to 0 dB to achieve the same BER. This proves the existence of saving power by 3 dB of 4x4 MIMO-OFDM system without coding, power saving 7 dB of 2x2 MIMO-OFDM system without coding and significant power savings from SISO-OFDM system.

Keywords: convolution code, OFDM, MIMO, QAM, BER

Procedia PDF Downloads 386
1440 Enhancing Transfer Path Analysis with In-Situ Component Transfer Path Analysis for Interface Forces Identification

Authors: Raef Cherif, Houssine Bakkali, Wafaa El Khatiri, Yacine Yaddaden

Abstract:

The analysis of how vibrations are transmitted between components is required in many engineering applications. Transfer path analysis (TPA) has been a valuable engineering tool for solving Noise, Vibration, and Harshness (NVH problems using sub-structuring applications. The most challenging part of a TPA analysis is estimating the equivalent forces at the contact points between the active and the passive side. Component TPA in situ Method calculates these forces by inverting the frequency response functions (FRFs) measured at the passive subsystem, relating the motion at indicator points to forces at the interface. However, matrix inversion could pose problems due to the ill-conditioning of the matrices leading to inaccurate results. This paper establishes a TPA model for an academic system consisting of two plates linked by four springs. A numerical study has been performed to improve the interface forces identification. Several parameters are studied and discussed, such as the singular value rejection and the number and position of indicator points chosen and used in the inversion matrix.

Keywords: transfer path analysis, matrix inverse method, indicator points, SVD decomposition

Procedia PDF Downloads 79
1439 Overconfidence and Self-Attribution Bias: The Difference among Economic Students at Different Stage of the Study and Non-Economic Students

Authors: Vera Jancurova

Abstract:

People are, in general, exposed to behavioral biases, however, the degree and impact are affected by experience, knowledge, and other characteristics. The purpose of this article is to study two of defined behavioral biases, the overconfidence and self-attribution bias, and its impact on economic and non-economic students at different stage of the study. The research method used for the purpose of this study is a controlled field study that contains questions on perception of own confidence and self-attribution and estimation of limits to analyse actual abilities. The results of the research show that economic students seem to be more overconfident than their non–economic colleagues, which seems to be caused by the fact the questionnaire was asking for predicting economic indexes and own knowledge and abilities in financial environment. Surprisingly, the most overconfidence was detected by the students at the beginning of their study (1st-semester students). However, the estimations of real numbers do not point out, that economic students have better results by the prediction itself. The study confirmed the presence of self-attribution bias at all of the respondents.

Keywords: behavioral finance, overconfidence, self-attribution, heuristics and biases

Procedia PDF Downloads 255
1438 Meta Mask Correction for Nuclei Segmentation in Histopathological Image

Authors: Jiangbo Shi, Zeyu Gao, Chen Li

Abstract:

Nuclei segmentation is a fundamental task in digital pathology analysis and can be automated by deep learning-based methods. However, the development of such an automated method requires a large amount of data with precisely annotated masks which is hard to obtain. Training with weakly labeled data is a popular solution for reducing the workload of annotation. In this paper, we propose a novel meta-learning-based nuclei segmentation method which follows the label correction paradigm to leverage data with noisy masks. Specifically, we design a fully conventional meta-model that can correct noisy masks by using a small amount of clean meta-data. Then the corrected masks are used to supervise the training of the segmentation model. Meanwhile, a bi-level optimization method is adopted to alternately update the parameters of the main segmentation model and the meta-model. Extensive experimental results on two nuclear segmentation datasets show that our method achieves the state-of-the-art result. In particular, in some noise scenarios, it even exceeds the performance of training on supervised data.

Keywords: deep learning, histopathological image, meta-learning, nuclei segmentation, weak annotations

Procedia PDF Downloads 136
1437 A Digital Pulse-Width Modulation Controller for High-Temperature DC-DC Power Conversion Application

Authors: Jingjing Lan, Jun Yu, Muthukumaraswamy Annamalai Arasu

Abstract:

This paper presents a digital non-linear pulse-width modulation (PWM) controller in a high-voltage (HV) buck-boost DC-DC converter for the piezoelectric transducer of the down-hole acoustic telemetry system. The proposed design controls the generation of output signal with voltage higher than the supply voltage and is targeted to work under high temperature. To minimize the power consumption and silicon area, a simple and efficient design scheme is employed to develop the PWM controller. The proposed PWM controller consists of serial to parallel (S2P) converter, data assign block, a mode and duty cycle controller (MDC), linearly PWM (LPWM) and noise shaper, pulse generator and clock generator. To improve the reliability of circuit operation at higher temperature, this design is fabricated with the 1.0-μm silicon-on-insulator (SOI) CMOS process. The implementation results validated that the proposed design has the advantages of smaller size, lower power consumption and robust thermal stability.

Keywords: DC-DC power conversion, digital control, high temperatures, pulse-width modulation

Procedia PDF Downloads 391
1436 Building Carbon Footprint Comparison between Building Permit, as Built, as Built with Circular Material Usage

Authors: Kadri-Ann Kertsmik, Martin Talvik, Kimmo Lylykangas, Simo Ilomets, Targo Kalamees

Abstract:

This study compares the building carbon footprint (CF) values for a case study of a private house located in a cold climate, using the Level(s) methodology. It provides a framework for measuring the environmental performance of buildings throughout their life cycle, taking into account various factors. The study presents the results of the three scenarios, comparing their carbon emissions and highlighting the benefits of circular material usage. The construction process was thoroughly documented, and all materials and components (including minuscule mechanical fasteners, each meter of cable, a kilogram of mortar, and the component of HVAC systems, among other things) delivered to the construction site were noted. Transportation distances of each delivery, the fuel consumption of construction machines, and electricity consumption for temporary heating and electrical tools were also monitored. Using the detailed data on material and energy resources, the CF was calculated for two scenarios: one where circular material usage was applied and another where virgin materials were used instead of reused ones. The results were compared with the CF calculated based on the building permit design model using the Level(s) methodology. To study the range of possible results in the early stage of CF assessment, the same building permit design was given to several experts. Results showed that embodied carbon values for a built scenario were significantly lower than the values predicted by the building permit stage as a result of more precise material quantities, as the calculation methodology is designed to overestimate the CF. Moreover, designers made an effort to reduce the building's CF by reusing certain materials such as ceramic tiles, lightweight concrete blocks, and timber during the construction process. However, in a cold climate context where operational energy (B6) continues to dominate, the total building CF value changes between the three scenarios were less significant. The calculation for the building permit project was performed by several experts, and CF results were in the same range. It alludes that, for the first estimation of preliminary building CF, using average values proves to be an appropriate method for the Estonian national carbon footprint estimation phase during building permit application. The study also identified several opportunities for reducing the carbon footprint of the building, such as reusing materials from other construction sites, preferring local material producers, and reducing wastage on site. The findings suggest that using circular materials can significantly reduce the carbon footprint of buildings. Overall, the study highlights the importance of using a comprehensive approach to measure the environmental performance of buildings, taking into account both the project and the actually built house. It also emphasises the need for ongoing monitoring for designing the building and construction site waste. The study also gives some examples of how to enable future circularity of building components and materials, e.g., building in layers, using wood as untreated, etc.

Keywords: carbon footprint, circular economy, sustainable construction, level(s) methodology

Procedia PDF Downloads 80
1435 Influence of the Line Parameters in Transmission Line Fault Location

Authors: Marian Dragomir, Alin Dragomir

Abstract:

In the paper, two fault location algorithms are presented for transmission lines which use the line parameters to estimate the distance to the fault. The first algorithm uses only the measurements from one end of the line and the positive and zero sequence parameters of the line, while the second one uses the measurements from both ends of the line and only the positive sequence parameters of the line. The algorithms were tested using a transmission grid transposed in MATLAB. In a first stage it was established a fault location base line, where the algorithms mentioned above estimate the fault locations using the exact line parameters. After that, the positive and zero sequence resistance and reactance of the line were calculated again for different ground resistivity values and then the fault locations were estimated again in order to compare the results with the base line results. The results show that the algorithm which uses the zero sequence impedance of the line is the most sensitive to the line parameters modifications. The other algorithm is less sensitive to the line parameters modification.

Keywords: estimation algorithms, fault location, line parameters, simulation tool

Procedia PDF Downloads 350
1434 Particle Swarm Optimization Based Method for Minimum Initial Marking in Labeled Petri Nets

Authors: Hichem Kmimech, Achref Jabeur Telmoudi, Lotfi Nabli

Abstract:

The estimation of the initial marking minimum (MIM) is a crucial problem in labeled Petri nets. In the case of multiple choices, the search for the initial marking leads to a problem of optimization of the minimum allocation of resources with two constraints. The first concerns the firing sequence that could be legal on the initial marking with respect to the firing vector. The second deals with the total number of tokens that can be minimal. In this article, the MIM problem is solved by the meta-heuristic particle swarm optimization (PSO). The proposed approach presents the advantages of PSO to satisfy the two previous constraints and find all possible combinations of minimum initial marking with the best computing time. This method, more efficient than conventional ones, has an excellent impact on the resolution of the MIM problem. We prove through a set of definitions, lemmas, and examples, the effectiveness of our approach.

Keywords: marking, production system, labeled Petri nets, particle swarm optimization

Procedia PDF Downloads 174
1433 Spread Spectrum with Notch Frequency Using Pulse Coding Method for Switching Converter of Communication Equipment

Authors: Yasunori Kobori, Futoshi Fukaya, Takuya Arafune, Nobukazu Tsukiji, Nobukazu Takai, Haruo Kobayashi

Abstract:

This paper proposes an EMI spread spectrum technique to enable to set notch frequencies using pulse coding method for DC-DC switching converters of communication equipment. The notches in the spectrum of the switching pulses appear at the frequencies obtained from empirically derived equations with the proposed spread spectrum technique using the pulse coding methods, the PWM (Pulse Width Modulation) coding or the PCM (Pulse Cycle Modulation) coding. This technique would be useful for the switching converters in the communication equipment which receives standard radio waves, without being affected by noise from the switching converters. In our proposed technique, the notch frequencies in the spectrum depend on the pulse coding method. We have investigated this technique to apply to the switching converters and found that there is good relationship agreement between the notch frequencies and the empirical equations. The notch frequencies with the PWM coding is equal to the equation F=k/(WL-WS). With the PCM coding, that is equal to the equation F=k/(TL-TS).

Keywords: notch frequency, pulse coding, spread spectrum, switching converter

Procedia PDF Downloads 366
1432 Estimation of Location and Scale Parameters of Extended Exponential Distribution Based on Record Statistics

Authors: E. Krishna

Abstract:

An Extended form of exponential distribution using Marshall and Olkin method is introduced.The location scale family of these distributions is considered. For location scale free family, exact expressions for single and product moments of upper record statistics are derived. The mean, variance and covariance of record values are computed for various values of the shape parameter. Using these the BLUE's of location and scale parameters are derived.The variances and covariance of estimates are obtained.Through Monte Carlo simulation the con dence intervals for location and scale parameters are constructed.The Best liner unbiased Predictor (BLUP) of future records are also discussed.

Keywords: BLUE, BLUP, con dence interval, Marshall-Olkin distribution, Monte Carlo simulation, prediction of future records, record statistics

Procedia PDF Downloads 415
1431 Effect of Composite Material on Damping Capacity Improvement of Cutting Tool in Machining Operation Using Taguchi Approach

Authors: Siamak Ghorbani, Nikolay Ivanovich Polushin

Abstract:

Chatter vibrations, occurring during cutting process, cause vibration between the cutting tool and workpiece, which deteriorates surface roughness and reduces tool life. The purpose of this study is to investigate the influence of cutting parameters and tool construction on surface roughness and vibration in turning of aluminum alloy AA2024. A new design of cutting tool is proposed, which is filled up with epoxy granite in order to improve damping capacity of the tool. Experiments were performed at the lathe using carbide cutting insert coated with TiC and two different cutting tools made of AISI 5140 steel. Taguchi L9 orthogonal array was applied to design of experiment and to optimize cutting conditions. By the help of signal-to-noise ratio and analysis of variance the optimal cutting condition and the effect of the cutting parameters on surface roughness and vibration were determined. Effectiveness of Taguchi method was verified by confirmation test. It was revealed that new cutting tool with epoxy granite has reduced vibration and surface roughness due to high damping properties of epoxy granite in toolholder.

Keywords: ANOVA, damping capacity, surface roughness, Taguchi method, vibration

Procedia PDF Downloads 307
1430 Study on the Transition to Pacemaker of Two Coupled Neurons

Authors: Sun Zhe, Ruggero Micheletto

Abstract:

The research of neural network is very important for the development of advanced next generation intelligent devices and the medical treatment. The most important part of the neural network research is the learning. The process of learning in our brain is essentially several adjustment processes of connection strength between neurons. It is very difficult to figure out how this mechanism works in the complex network and how the connection strength influences brain functions. For this reason, we made a model with only two coupled neurons and studied the influence of connection strength between them. To emulate the neuronal activity of realistic neurons, we prefer to use the Izhikevich neuron model. This model can simulate the neuron variables accurately and it’s simplicity is very suitable to implement on computers. In this research, the parameter ρ is used to estimate the correlation coefficient between spike train of two coupling neurons.We think the results is very important for figuring out the mechanism between synchronization of coupling neurons and synaptic plasticity. The result also presented the importance of the spike frequency adaptation in complex systems.

Keywords: neural networks, noise, stochastic processes, coupled neurons, correlation coefficient, synchronization, pacemaker, synaptic plasticity

Procedia PDF Downloads 278
1429 Poisoning Admission in Pediatrics Benghazi Hospital in Libya: Three Years Review of Medical Record

Authors: Mudafara Bengleil

Abstract:

Estimation of the magnitude and causes of poisoning was the objective of the current study. A retrospective study of medical records of all poisoning children admitted to Benghazi Children Hospital in Libya from January 2008 up to December 2010. Number of children admitted was 244; the age ranged from less than one to 13 years old. Most of cases were admitted with mild symptom and the majority of them were boys. Only few cases admitted to intensive care unit and there was no mortality recorded through the period of study. Age group 1 to 3 years (50.8%) had the highest frequency of admission and the peak of admission was during summer. The most common cause of admission was due to ingestion of medication (53.69%), House hold product exposure (26.64%) was the second causes of admission while, 19.67% of admissions were due to Food poisoning. Almost all admitted cases were accidental and medicines were the most consumed substances in addition, improper storage of toxic agents were the first risk factor of poisoning. Present results indicated that, children poisoning seems to be a common pediatric care problem which need to control and prevent.

Keywords: poisoning, children, hospital, medical

Procedia PDF Downloads 411
1428 Analysis of Vertical Hall Effect Device Using Current-Mode

Authors: Kim Jin Sup

Abstract:

This paper presents a vertical hall effect device using current-mode. Among different geometries that have been studied and simulated using COMSOL Multiphysics, optimized cross-shaped model displayed the best sensitivity. The cross-shaped model emerged as the optimum plate to fit the lowest noise and residual offset and the best sensitivity. The symmetrical cross-shaped hall plate is widely used because of its high sensitivity and immunity to alignment tolerances resulting from the fabrication process. The hall effect device has been designed using a 0.18-μm CMOS technology. The simulation uses the nominal bias current of 12μA. The applied magnetic field is from 0 mT to 20 mT. Simulation results achieved in COMSOL and validated with respect to the electrical behavior of equivalent circuit for Cadence. Simulation results of the one structure over the 13 available samples shows for the best geometry a current-mode sensitivity of 6.6 %/T at 20mT. Acknowledgment: This work was supported by Institute for Information & communications Technology Promotion (IITP) grant funded by the Korea government (MSIP) (No. R7117-16-0165, Development of Hall Effect Semiconductor for Smart Car and Device).

Keywords: vertical hall device, current-mode, crossed-shaped model, CMOS technology

Procedia PDF Downloads 286
1427 Mixed Treatment (Physical-Chemical and Biological) of Ouled Fayet Landfill Leachates

Authors: O. Balamane-Zizi, L. M. Rouidi, A. Boukhrissa, N. Daas, H. Ait-amar

Abstract:

The objective of this study was to test the possibility of a mixed treatment (physical-chemical and biological) of Ouled Fayet leachates which date of 10 years and has a large fraction of hard COD that can be reduced by coagulation-flocculation. Previous batch tests showed the possibility of applying the physical-chemical and biological treatments separately; the removal efficiencies obtained in this case were not interesting. We propose, therefore, to test the possibility of a combined treatment, in order to improve the quality of the leachates. Estimation of the treatment’s effectiveness was done by analysis of some pollution parameters such as COD, suspended solids, and heavy metals (particularly iron and nickel). The main results obtained after the combination of treatments, show reduction rate of about 63% for COD, 73% for suspended solids and 80% for iron and nickel. We also noted an improvement in the turbidity of treated leachates.

Keywords: landfill leachates, COD, physical-chemical treatment, biological treatment

Procedia PDF Downloads 466
1426 Measuring Banking Risk

Authors: Mike Tsionas

Abstract:

The paper develops new indices of financial stability based on an explicit model of expected utility maximization by financial institutions subject to the classical technology restrictions of neoclassical production theory. The model can be estimated using standard econometric techniques, like GMM for dynamic panel data and latent factor analysis for the estimation of co-variance matrices. An explicit functional form for the utility function is not needed and we show how measures of risk aversion and prudence (downside risk aversion) can be derived and estimated from the model. The model is estimated using data for Eurozone countries and we focus particularly on (i) the use of the modeling approach as an “early warning mechanism”, (ii) the bank- and country-specific estimates of risk aversion and prudence (downside risk aversion), and (iii) the derivation of a generalized measure of risk that relies on loan-price uncertainty.

Keywords: financial stability, banking, expected utility maximization, sub-prime crisis, financial crisis, eurozone, PIIGS

Procedia PDF Downloads 344
1425 Localization Mobile Beacon Using RSSI

Authors: Sallama Resen, Celal Öztürk

Abstract:

Distance estimation between tow nodes has wide scope of surveillance and tracking applications. This paper suggests a Bluetooth Low Energy (BLE) technology as a media for transceiver and receiver signal in small indoor areas. As an example, BLE communication technologies used in child safety domains. Local network is designed to detect child position in indoor school area consisting Mobile Beacons (MB), Access Points (AP) and Smart Phones (SP) where MBs stuck in children’s shoes as wearable sensors. This paper presents a technique that can detect mobile beacons’ position and help finding children’s location within dynamic environment. By means of bluetooth beacons that are attached to child’s shoes, the distance between the MB and teachers SP is estimated with an accuracy of less than one meter. From the simulation results, it is shown that high accuracy of position coordinates are achieved for multi-mobile beacons in different environments.

Keywords: bluetooth low energy, child safety, mobile beacons, received signal strength

Procedia PDF Downloads 341
1424 LGG Architecture for Brain Tumor Segmentation Using Convolutional Neural Network

Authors: Sajeeha Ansar, Asad Ali Safi, Sheikh Ziauddin, Ahmad R. Shahid, Faraz Ahsan

Abstract:

The most aggressive form of brain tumor is called glioma. Glioma is kind of tumor that arises from glial tissue of the brain and occurs quite often. A fully automatic 2D-CNN model for brain tumor segmentation is presented in this paper. We performed pre-processing steps to remove noise and intensity variances using N4ITK and standard intensity correction, respectively. We used Keras open-source library with Theano as backend for fast implementation of CNN model. In addition, we used BRATS 2015 MRI dataset to evaluate our proposed model. Furthermore, we have used SimpleITK open-source library in our proposed model to analyze images. Moreover, we have extracted random 2D patches for proposed 2D-CNN model for efficient brain segmentation. Extracting 2D patched instead of 3D due to less dimensional information present in 2D which helps us in reducing computational time. Dice Similarity Coefficient (DSC) is used as performance measure for the evaluation of the proposed method. Our method achieved DSC score of 0.77 for complete, 0.76 for core, 0.77 for enhanced tumor regions. However, these results are comparable with methods already implemented 2D CNN architecture.

Keywords: brain tumor segmentation, convolutional neural networks, deep learning, LGG

Procedia PDF Downloads 176
1423 Robot Movement Using the Trust Region Policy Optimization

Authors: Romisaa Ali

Abstract:

The Policy Gradient approach is one of the deep reinforcement learning families that combines deep neural networks (DNN) with reinforcement learning RL to discover the optimum of the control problem through experience gained from the interaction between the robot and its surroundings. In contrast to earlier policy gradient algorithms, which were unable to handle these two types of error because of over-or under-estimation introduced by the deep neural network model, this article will discuss the state-of-the-art SOTA policy gradient technique, trust region policy optimization (TRPO), by applying this method in various environments compared to another policy gradient method, the Proximal Policy Optimization (PPO), to explain their robust optimization, using this SOTA to gather experience data during various training phases after observing the impact of hyper-parameters on neural network performance.

Keywords: deep neural networks, deep reinforcement learning, proximal policy optimization, state-of-the-art, trust region policy optimization

Procedia PDF Downloads 164
1422 Nonparametric Specification Testing for the Drift of the Short Rate Diffusion Process Using a Panel of Yields

Authors: John Knight, Fuchun Li, Yan Xu

Abstract:

Based on a new method of the nonparametric estimator of the drift function, we propose a consistent test for the parametric specification of the drift function in the short rate diffusion process using observations from a panel of yields. The test statistic is shown to follow an asymptotic normal distribution under the null hypothesis that the parametric drift function is correctly specified, and converges to infinity under the alternative. Taking the daily 7-day European rates as a proxy of the short rate, we use our test to examine whether the drift of the short rate diffusion process is linear or nonlinear, which is an unresolved important issue in the short rate modeling literature. The testing results indicate that none of the drift functions in this literature adequately captures the dynamics of the drift, but nonlinear specification performs better than the linear specification.

Keywords: diffusion process, nonparametric estimation, derivative security price, drift function and volatility function

Procedia PDF Downloads 359
1421 Special Case of Trip Distribution Model and Its Use for Estimation of Detailed Transport Demand in the Czech Republic

Authors: Jiri Dufek

Abstract:

The national model of the Czech Republic has been modified in a detailed way to get detailed travel demand in the municipality level (cities, villages over 300 inhabitants). As a technique for this detailed modelling, three-dimensional procedure for calibrating gravity models, was used. Besides of zone production and attraction, which is usual in gravity models, the next additional parameter for trip distribution was introduced. Usually it is called by “third dimension”. In the model, this parameter is a demand between regions. The distribution procedure involved calculation of appropriate skim matrices and its multiplication by three coefficients obtained by iterative balancing of production, attraction and third dimension. This type of trip distribution was processed in R-project and the results were used in the Czech Republic transport model, created in PTV Vision. This process generated more precise results in local level od the model (towns, villages)

Keywords: trip distribution, three dimension, transport model, municipalities

Procedia PDF Downloads 124
1420 Estimation of PM2.5 Emissions and Source Apportionment Using Receptor and Dispersion Models

Authors: Swetha Priya Darshini Thammadi, Sateesh Kumar Pisini, Sanjay Kumar Shukla

Abstract:

Source apportionment using Dispersion model depends primarily on the quality of Emission Inventory. In the present study, a CMB receptor model has been used to identify the sources of PM2.5, while the AERMOD dispersion model has been used to account for missing sources of PM2.5 in the Emission Inventory. A statistical approach has been developed to quantify the missing sources not considered in the Emission Inventory. The inventory of each grid was improved by adjusting emissions based on road lengths and deficit in measured and modelled concentrations. The results showed that in CMB analyses, fugitive sources - soil and road dust - contribute significantly to ambient PM2.5 pollution. As a result, AERMOD significantly underestimated the ambient air concentration at most locations. The revised Emission Inventory showed a significant improvement in AERMOD performance which is evident through statistical tests.

Keywords: CMB, GIS, AERMOD, PM₂.₅, fugitive, emission inventory

Procedia PDF Downloads 334
1419 Short and Long Crack Growth Behavior in Ferrite Bainite Dual Phase Steels

Authors: Ashok Kumar, Shiv Brat Singh, Kalyan Kumar Ray

Abstract:

There is growing awareness to design steels against fatigue damage Ferrite martensite dual-phase steels are known to exhibit favourable mechanical properties like good strength, ductility, toughness, continuous yielding, and high work hardening rate. However, dual-phase steels containing bainite as second phase are potential alternatives for ferrite martensite steels for certain applications where good fatigue property is required. Fatigue properties of dual phase steels are popularly assessed by the nature of variation of crack growth rate (da/dN) with stress intensity factor range (∆K), and the magnitude of fatigue threshold (∆Kth) for long cracks. There exists an increased emphasis to understand not only the long crack fatigue behavior but also short crack growth behavior of ferrite bainite dual phase steels. The major objective of this report is to examine the influence of microstructures on the short and long crack growth behavior of a series of developed dual-phase steels with varying amounts of bainite and. Three low carbon steels containing Nb, Cr and Mo as microalloying elements steels were selected for making ferrite-bainite dual-phase microstructures by suitable heat treatments. The heat treatment consisted of austenitizing the steel at 1100°C for 20 min, cooling at different rates in air prior to soaking these in a salt bath at 500°C for one hour, and finally quenching in water. Tensile tests were carried out on 25 mm gauge length specimens with 5 mm diameter using nominal strain rate 0.6x10⁻³ s⁻¹ at room temperature. Fatigue crack growth studies were made on a recently developed specimen configuration using a rotating bending machine. The crack growth was monitored by interrupting the test and observing the specimens under an optical microscope connected to an Image analyzer. The estimated crack lengths (a) at varying number of cycles (N) in different fatigue experiments were analyzed to obtain log da/dN vs. log °∆K curves for determining ∆Kthsc. The microstructural features of these steels have been characterized and their influence on the near threshold crack growth has been examined. This investigation, in brief, involves (i) the estimation of ∆Kthsc and (ii) the examination of the influence of microstructure on short and long crack fatigue threshold. The maximum fatigue threshold values obtained from short crack growth experiments on various specimens of dual-phase steels containing different amounts of bainite are found to increase with increasing bainite content in all the investigated steels. The variations of fatigue behavior of the selected steel samples have been explained with the consideration of varying amounts of the constituent phases and their interactions with the generated microstructures during cyclic loading. Quantitative estimation of the different types of fatigue crack paths indicates that the propensity of a crack to pass through the interfaces depends on the relative amount of the microstructural constituents. The fatigue crack path is found to be predominantly intra-granular except for the ones containing > 70% bainite in which it is predominantly inter-granular.

Keywords: bainite, dual phase steel, fatigue crack growth rate, long crack fatigue threshold, short crack fatigue threshold

Procedia PDF Downloads 201
1418 Times2D: A Time-Frequency Method for Time Series Forecasting

Authors: Reza Nematirad, Anil Pahwa, Balasubramaniam Natarajan

Abstract:

Time series data consist of successive data points collected over a period of time. Accurate prediction of future values is essential for informed decision-making in several real-world applications, including electricity load demand forecasting, lifetime estimation of industrial machinery, traffic planning, weather prediction, and the stock market. Due to their critical relevance and wide application, there has been considerable interest in time series forecasting in recent years. However, the proliferation of sensors and IoT devices, real-time monitoring systems, and high-frequency trading data introduce significant intricate temporal variations, rapid changes, noise, and non-linearities, making time series forecasting more challenging. Classical methods such as Autoregressive integrated moving average (ARIMA) and Exponential Smoothing aim to extract pre-defined temporal variations, such as trends and seasonality. While these methods are effective for capturing well-defined seasonal patterns and trends, they often struggle with more complex, non-linear patterns present in real-world time series data. In recent years, deep learning has made significant contributions to time series forecasting. Recurrent Neural Networks (RNNs) and their variants, such as Long short-term memory (LSTMs) and Gated Recurrent Units (GRUs), have been widely adopted for modeling sequential data. However, they often suffer from the locality, making it difficult to capture local trends and rapid fluctuations. Convolutional Neural Networks (CNNs), particularly Temporal Convolutional Networks (TCNs), leverage convolutional layers to capture temporal dependencies by applying convolutional filters along the temporal dimension. Despite their advantages, TCNs struggle with capturing relationships between distant time points due to the locality of one-dimensional convolution kernels. Transformers have revolutionized time series forecasting with their powerful attention mechanisms, effectively capturing long-term dependencies and relationships between distant time points. However, the attention mechanism may struggle to discern dependencies directly from scattered time points due to intricate temporal patterns. Lastly, Multi-Layer Perceptrons (MLPs) have also been employed, with models like N-BEATS and LightTS demonstrating success. Despite this, MLPs often face high volatility and computational complexity challenges in long-horizon forecasting. To address intricate temporal variations in time series data, this study introduces Times2D, a novel framework that parallelly integrates 2D spectrogram and derivative heatmap techniques. The spectrogram focuses on the frequency domain, capturing periodicity, while the derivative patterns emphasize the time domain, highlighting sharp fluctuations and turning points. This 2D transformation enables the utilization of powerful computer vision techniques to capture various intricate temporal variations. To evaluate the performance of Times2D, extensive experiments were conducted on standard time series datasets and compared with various state-of-the-art algorithms, including DLinear (2023), TimesNet (2023), Non-stationary Transformer (2022), PatchTST (2023), N-HiTS (2023), Crossformer (2023), MICN (2023), LightTS (2022), FEDformer (2022), FiLM (2022), SCINet (2022a), Autoformer (2021), and Informer (2021) under the same modeling conditions. The initial results demonstrated that Times2D achieves consistent state-of-the-art performance in both short-term and long-term forecasting tasks. Furthermore, the generality of the Times2D framework allows it to be applied to various tasks such as time series imputation, clustering, classification, and anomaly detection, offering potential benefits in any domain that involves sequential data analysis.

Keywords: derivative patterns, spectrogram, time series forecasting, times2D, 2D representation

Procedia PDF Downloads 36
1417 Estimating Tree Height and Forest Classification from Multi Temporal Risat-1 HH and HV Polarized Satellite Aperture Radar Interferometric Phase Data

Authors: Saurav Kumar Suman, P. Karthigayani

Abstract:

In this paper the height of the tree is estimated and forest types is classified from the multi temporal RISAT-1 Horizontal-Horizontal (HH) and Horizontal-Vertical (HV) Polarised Satellite Aperture Radar (SAR) data. The novelty of the proposed project is combined use of the Back-scattering Coefficients (Sigma Naught) and the Coherence. It uses Water Cloud Model (WCM). The approaches use two main steps. (a) Extraction of the different forest parameter data from the Product.xml, BAND-META file and from Grid-xxx.txt file come with the HH & HV polarized data from the ISRO (Indian Space Research Centre). These file contains the required parameter during height estimation. (b) Calculation of the Vegetation and Ground Backscattering, Coherence and other Forest Parameters. (c) Classification of Forest Types using the ENVI 5.0 Tool and ROI (Region of Interest) calculation.

Keywords: RISAT-1, classification, forest, SAR data

Procedia PDF Downloads 401
1416 Axle Load Estimation of Moving Vehicles Using BWIM Technique

Authors: Changgil Lee, Seunghee Park

Abstract:

Although vehicle driving test for the development of BWIM system is necessary, but it needs much cost and time in addition application of various driving condition. Thus, we need the numerical-simulation method resolving the cost and time problems of vehicle driving test and the way of measuring response of bridge according to the various driving condition. Using the precision analysis model reflecting the dynamic characteristic is contributed to increase accuracy in numerical simulation. In this paper, we conduct a numerical simulation to apply precision analysis model, which reflects the dynamic characteristic of bridge using Bridge Weigh-in-Motion technique and suggest overload vehicle enforcement technology using precision analysis model.

Keywords: bridge weigh-in-motion(BWIM) system, precision analysis model, dynamic characteristic of bridge, numerical simulation

Procedia PDF Downloads 284
1415 Optical Flow Localisation and Appearance Mapping (OFLAAM) for Long-Term Navigation

Authors: Daniel Pastor, Hyo-Sang Shin

Abstract:

This paper presents a novel method to use optical flow navigation for long-term navigation. Unlike standard SLAM approaches for augmented reality, OFLAAM is designed for Micro Air Vehicles (MAV). It uses an optical flow camera pointing downwards, an IMU and a monocular camera pointing frontwards. That configuration avoids the expensive mapping and tracking of the 3D features. It only maps these features in a vocabulary list by a localization module to tackle the loss of the navigation estimation. That module, based on the well-established algorithm DBoW2, will be also used to close the loop and allow long-term navigation in confined areas. That combination of high-speed optical flow navigation with a low rate localization algorithm allows fully autonomous navigation for MAV, at the same time it reduces the overall computational load. This framework is implemented in ROS (Robot Operating System) and tested attached to a laptop. A representative scenarios is used to analyse the performance of the system.

Keywords: vision, UAV, navigation, SLAM

Procedia PDF Downloads 602
1414 The Competitive Newsvendor Game with Overestimated Demand

Authors: Chengli Liu, C. K. M. Lee

Abstract:

The tradition competitive newsvendor game assumes decision makers are rational. However, there are behavioral biases when people make decisions, such as loss aversion, mental accounting and overconfidence. Overestimation of a subject’s own performance is one type of overconfidence. The objective of this research is to analyze the impact of the overestimated demand in the newsvendor competitive game with two players. This study builds a competitive newsvendor game model where newsvendors have private information of their demands, which is overestimated. At the same time, demands of each newsvendor forecasted by a third party institution are available. This research shows that the overestimation leads to demand steal effect, which reduces the competitor’s order quantity. However, the overall supply of the product increases due to overestimation. This study illustrates the boundary condition for the overestimated newsvendor to have the equilibrium order drop due to the demand steal effect from the other newsvendor. A newsvendor who has higher critical fractile will see its equilibrium order decrease with the drop of estimation level from the other newsvendor.

Keywords: bias, competing newsvendor, Nash equilibrium, overestimation

Procedia PDF Downloads 257
1413 Diffusion Adaptation Strategies for Distributed Estimation Based on the Family of Affine Projection Algorithms

Authors: Mohammad Shams Esfand Abadi, Mohammad Ranjbar, Reza Ebrahimpour

Abstract:

This work presents the distributed processing solution problem in a diffusion network based on the adapt then combine (ATC) and combine then adapt (CTA)selective partial update normalized least mean squares (SPU-NLMS) algorithms. Also, we extend this approach to dynamic selection affine projection algorithm (DS-APA) and ATC-DS-APA and CTA-DS-APA are established. The purpose of ATC-SPU-NLMS and CTA-SPU-NLMS algorithm is to reduce the computational complexity by updating the selected blocks of weight coefficients at every iteration. In CTA-DS-APA and ATC-DS-APA, the number of the input vectors is selected dynamically. Diffusion cooperation strategies have been shown to provide good performance based on these algorithms. The good performance of introduced algorithm is illustrated with various experimental results.

Keywords: selective partial update, affine projection, dynamic selection, diffusion, adaptive distributed networks

Procedia PDF Downloads 700