Search results for: interval estimation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2621

Search results for: interval estimation

2501 Deep Learning-Based Channel Estimation for Reconfigurable Intelligent Surface-Assisted Unmanned Aerial Vehicle-Enabled Wireless Communication System

Authors: Getaneh Berie Tarekegn

Abstract:

Wireless communication via unmanned aerial vehicles (UAVs) has drawn a great deal of attention due to its flexibility in establishing line-of-sight (LoS) communications. However, in complex urban and dynamic environments, the movement of UAVs can be blocked by trees and high-rise buildings that obstruct directional paths. With reconfigurable intelligent surfaces (RIS), this problem can be effectively addressed. To achieve this goal, accurate channel estimation in RIS-assisted UAV-enabled wireless communications is crucial. This paper proposes an accurate channel estimation model using long short-term memory (LSTM) for a multi-user RIS-assisted UAV-enabled wireless communication system. According to simulation results, LSTM can improve the channel estimation performance of RIS-assisted UAV-enabled wireless communication.

Keywords: channel estimation, reconfigurable intelligent surfaces, long short-term memory, unmanned aerial vehicles

Procedia PDF Downloads 55
2500 Effect of Soil Corrosion in Failures of Buried Gas Pipelines

Authors: Saima Ali, Pathamanathan Rajeev, Imteaz A. Monzur

Abstract:

In this paper, a brief review of the corrosion mechanism in buried pipe and modes of failure is provided together with the available corrosion models. Moreover, the sensitivity analysis is performed to understand the influence of corrosion model parameters on the remaining life estimation. Further, the probabilistic analysis is performed to propagate the uncertainty in the corrosion model on the estimation of the renaming life of the pipe. Finally, the comparison among the corrosion models on the basis of the remaining life estimation will be provided to improve the renewal plan.

Keywords: corrosion, pit depth, sensitivity analysis, exposure period

Procedia PDF Downloads 494
2499 Software Engineering Inspired Cost Estimation for Process Modelling

Authors: Felix Baumann, Aleksandar Milutinovic, Dieter Roller

Abstract:

Up to this point business process management projects in general and business process modelling projects in particular could not rely on a practical and scientifically validated method to estimate cost and effort. Especially the model development phase is not covered by a cost estimation method or model. Further phases of business process modelling starting with implementation are covered by initial solutions which are discussed in the literature. This article proposes a method of filling this gap by deriving a cost estimation method from available methods in similar domains namely software development or software engineering. Software development is regarded as closely similar to process modelling as we show. After the proposition of this method different ideas for further analysis and validation of the method are proposed. We derive this method from COCOMO II and Function Point which are established methods of effort estimation in the domain of software development. For this we lay out similarities of the software development rocess and the process of process modelling which is a phase of the Business Process Management life-cycle.

Keywords: COCOMO II, busines process modeling, cost estimation method, BPM COCOMO

Procedia PDF Downloads 402
2498 Reliability Estimation of Bridge Structures with Updated Finite Element Models

Authors: Ekin Ozer

Abstract:

Assessment of structural reliability is essential for efficient use of civil infrastructure which is subjected hazardous events. Dynamic analysis of finite element models is a commonly used tool to simulate structural behavior and estimate its performance accordingly. However, theoretical models purely based on preliminary assumptions and design drawings may deviate from the actual behavior of the structure. This study proposes up-to-date reliability estimation procedures which engages actual bridge vibration data modifying finite element models for finite element model updating and performing reliability estimation, accordingly. The proposed method utilizes vibration response measurements of bridge structures to identify modal parameters, then uses these parameters to calibrate finite element models which are originally based on design drawings. The proposed method does not only show that reliability estimation based on updated models differs from the original models, but also infer that non-updated models may overestimate the structural capacity.

Keywords: earthquake engineering, engineering vibrations, reliability estimation, structural health monitoring

Procedia PDF Downloads 178
2497 Estimation of Break Points of Housing Price Growth Rate for Top MSAs in Texas Area

Authors: Hui Wu, Ye Li

Abstract:

Applying the structural break estimation method proposed by Perron and Bai (1998) to the housing price growth rate of top 5 MSAs in the Texas area, this paper estimated the structural break date for the growth rate of housing prices index. As shown in the estimation results, the break dates for each region are quite different, which indicates the heterogeneity of the housing market in response to macroeconomic conditions.

Keywords: structural break, housing prices index, ADF test, linear model

Procedia PDF Downloads 117
2496 Point Estimation for the Type II Generalized Logistic Distribution Based on Progressively Censored Data

Authors: Rana Rimawi, Ayman Baklizi

Abstract:

Skewed distributions are important models that are frequently used in applications. Generalized distributions form a class of skewed distributions and gain widespread use in applications because of their flexibility in data analysis. More specifically, the Generalized Logistic Distribution with its different types has received considerable attention recently. In this study, based on progressively type-II censored data, we will consider point estimation in type II Generalized Logistic Distribution (Type II GLD). We will develop several estimators for its unknown parameters, including maximum likelihood estimators (MLE), Bayes estimators and linear estimators (BLUE). The estimators will be compared using simulation based on the criteria of bias and Mean square error (MSE). An illustrative example of a real data set will be given.

Keywords: point estimation, type II generalized logistic distribution, progressive censoring, maximum likelihood estimation

Procedia PDF Downloads 167
2495 Estimation and Restoration of Ill-Posed Parameters for Underwater Motion Blurred Images

Authors: M. Vimal Raj, S. Sakthivel Murugan

Abstract:

Underwater images degrade their quality due to atmospheric conditions. One of the major problems in an underwater image is motion blur caused by the imaging device or the movement of the object. In order to rectify that in post-imaging, parameters of the blurred image are to be estimated. So, the point spread function is estimated by the properties, using the spectrum of the image. To improve the estimation accuracy of the parameters, Optimized Polynomial Lagrange Interpolation (OPLI) method is implemented after the angle and length measurement of motion-blurred images. Initially, the data were collected from real-time environments in Chennai and processed. The proposed OPLI method shows better accuracy than the existing classical Cepstral, Hough, and Radon transform estimation methods for underwater images.

Keywords: image restoration, motion blur, parameter estimation, radon transform, underwater

Procedia PDF Downloads 152
2494 On Parameter Estimation of Simultaneous Linear Functional Relationship Model for Circular Variables

Authors: N. A. Mokhtar, A. G. Hussin, Y. Z. Zubairi

Abstract:

This paper proposes a new simultaneous simple linear functional relationship model by assuming equal error variances. We derive the maximum likelihood estimate of the parameters in the simultaneous model and the covariance. We show by simulation study the small bias values of the parameters suggest the suitability of the estimation method. As an illustration, the proposed simultaneous model is applied to real data of the wind direction and wave direction measured by two different instruments.

Keywords: simultaneous linear functional relationship model, Fisher information matrix, parameter estimation, circular variables

Procedia PDF Downloads 329
2493 Effect of Adjacent Footings on Elastic Settlement of Shallow Foundations

Authors: Mustafa Aytekin

Abstract:

In this study, impact of adjacent footings is considered on the estimation of elastic settlement of shallow foundations. In the estimation of elastic settlement, the Schmertmann’s method that is a very popular method in the elastic settlement estimation of shallow foundations is employed. In order to consider affect of neighboring footings on elastic settlement of main footing in different configurations, a MATLAB script has been generated. Elastic settlements of the various configurations are estimated by the script and several conclusions have been reached.

Keywords: elastic (immediate) settlement, Schmertman Method, adjacent footings, shallow foundations

Procedia PDF Downloads 443
2492 Kalman Filter Gain Elimination in Linear Estimation

Authors: Nicholas D. Assimakis

Abstract:

In linear estimation, the traditional Kalman filter uses the Kalman filter gain in order to produce estimation and prediction of the n-dimensional state vector using the m-dimensional measurement vector. The computation of the Kalman filter gain requires the inversion of an m x m matrix in every iteration. In this paper, a variation of the Kalman filter eliminating the Kalman filter gain is proposed. In the time varying case, the elimination of the Kalman filter gain requires the inversion of an n x n matrix and the inversion of an m x m matrix in every iteration. In the time invariant case, the elimination of the Kalman filter gain requires the inversion of an n x n matrix in every iteration. The proposed Kalman filter gain elimination algorithm may be faster than the conventional Kalman filter, depending on the model dimensions.

Keywords: discrete time, estimation, Kalman filter, Kalman filter gain

Procedia PDF Downloads 160
2491 Brain Age Prediction Based on Brain Magnetic Resonance Imaging by 3D Convolutional Neural Network

Authors: Leila Keshavarz Afshar, Hedieh Sajedi

Abstract:

Estimation of biological brain age from MR images is a topic that has been much addressed in recent years due to the importance it attaches to early diagnosis of diseases such as Alzheimer's. In this paper, we use a 3D Convolutional Neural Network (CNN) to provide a method for estimating the biological age of the brain. The 3D-CNN model is trained by MRI data that has been normalized. In addition, to reduce computation while saving overall performance, some effectual slices are selected for age estimation. By this method, the biological age of individuals using selected normalized data was estimated with Mean Absolute Error (MAE) of 4.82 years.

Keywords: brain age estimation, biological age, 3D-CNN, deep learning, T1-weighted image, SPM, preprocessing, MRI, canny, gray matter

Procedia PDF Downloads 117
2490 Age Estimation Using Atlas Method with Orthopantomogram and Digital Tracing on Lateral Cephalogram

Authors: Astika Swastirani

Abstract:

Chronological age estimation can be done by looking at the stage of growth and development of teeth from orthopantomogram and mandibular remodeling from lateral cephalogram. Mandibular morphological changes associated with the size and remodeling during growth is a strong indicator of age estimation. These changes can be observed with lateral cephalogram. Objective: To prove the difference between chronological age and age estimation using orthopantomogram (dental age) and lateral cephalogram (skeletal age). Methods: Sample consisted of 100 medical records, 100 orthopantomograms digital and 100 lateral cephalograms digital belongs to 50 male and 50 female of Airlangga University hospital of dentistry. Orthopantomogram were matched with London atlas and lateral cephalograms were observed by digital tracing. The difference of dental age and skeletal age was analyzed by pair t –test. Result: Result of the pair t-test between chronological age and dental age in male (p-value 0.002, p<0.05), in female (p-value 0.605, p>0.05). Result of pair t-test between the chronological age and skeletal age (variable length Condylion-Gonion, Gonion-Gnathion, Condylion-Gnathion in male (p-value 0.000, p<0.05) in female (variable Condylion-Gonion length (p-value 0.000, Condylion-Gnathion length (p-value 0,040) and Gonion-Gnathion length (p-value 0.493). Conclusion: Orthopantomogram with London atlas and lateral cephalograms with Gonion- Gnathion variable can be used for age estimation in female. Orthopantomogram with London atlas and lateral cephalograms with Condylion-Gonion variable, Gonion-Gnathion variable and Condylion-Gnathion can not be used for age estimation in male.

Keywords: age estimation, chronological age, dental age, skeletal age

Procedia PDF Downloads 141
2489 The Linear Combination of Kernels in the Estimation of the Cumulative Distribution Functions

Authors: Abdel-Razzaq Mugdadi, Ruqayyah Sani

Abstract:

The Kernel Distribution Function Estimator (KDFE) method is the most popular method for nonparametric estimation of the cumulative distribution function. The kernel and the bandwidth are the most important components of this estimator. In this investigation, we replace the kernel in the KDFE with a linear combination of kernels to obtain a new estimator based on the linear combination of kernels, the mean integrated squared error (MISE), asymptotic mean integrated squared error (AMISE) and the asymptotically optimal bandwidth for the new estimator are derived. We propose a new data-based method to select the bandwidth for the new estimator. The new technique is based on the Plug-in technique in density estimation. We evaluate the new estimator and the new technique using simulations and real-life data.

Keywords: estimation, bandwidth, mean square error, cumulative distribution function

Procedia PDF Downloads 544
2488 Estimation and Comparison of Delay at Signalized Intersections Based on Existing Methods

Authors: Arpita Saha, Satish Chandra, Indrajit Ghosh

Abstract:

Delay implicates the time loss of a traveler while crossing an intersection. Efficiency of traffic operation at signalized intersections is assessed in terms of delay caused to an individual vehicle. Highway Capacity Manual (HCM) method and Webster’s method are the most widely used in India for delay estimation purpose. However, in India, traffic is highly heterogeneous in nature with extremely poor lane discipline. Therefore, to explore best delay estimation technique for Indian condition, a comparison was made. In this study, seven signalized intersections from three different cities where chosen. Data was collected for both during morning and evening peak hours. Only under saturated cycles were considered for this study. Delay was estimated based on the field data. With the help of Simpson’s 1/3 rd rule, delay of under saturated cycles was estimated by measuring the area under the curve of queue length and cycle time. Moreover, the field observed delay was compared with the delay estimated using HCM, Webster, Probabilistic, Taylor’s expansion and Regression methods. The drawbacks of the existing delay estimation methods to be use in Indian heterogeneous traffic conditions were figured out, and best method was proposed. It was observed that direct estimation of delay using field measured data is more accurate than existing conventional and modified methods.

Keywords: delay estimation technique, field delay, heterogeneous traffic, signalised intersection

Procedia PDF Downloads 270
2487 Study of Harmonics Estimation on Analog kWh Meter Using Fast Fourier Transform Method

Authors: Amien Rahardjo, Faiz Husnayain, Iwa Garniwa

Abstract:

PLN used the kWh meter to determine the amount of energy consumed by the household customers. High precision of kWh meter is needed in order to give accuracy results as the accuracy can be decreased due to the presence of harmonic. In this study, an estimation of active power consumed was developed. Based on the first year study results, the largest deviation due to harmonics can reach up to 9.8% in 2200VA and 12.29% in 3500VA with kWh meter analog. In the second year of study, deviation of digital customer meter reaches 2.01% and analog meter up to 9.45% for 3500VA household customers. The aim of this research is to produce an estimation system to calculate the total energy consumed by household customer using analog meter so the losses due to irregularities PLN recording of energy consumption based on the measurement used Analog kWh-meter installed is avoided.

Keywords: harmonics estimation, harmonic distortion, kWh meters analog and digital, THD, household customers

Procedia PDF Downloads 459
2486 On Coverage Probability of Confidence Intervals for the Normal Mean with Known Coefficient of Variation

Authors: Suparat Niwitpong, Sa-aat Niwitpong

Abstract:

Statistical inference of normal mean with known coefficient of variation has been investigated recently. This phenomenon occurs normally in environment and agriculture experiments when the scientist knows the coefficient of variation of their experiments. In this paper, we constructed new confidence intervals for the normal population mean with known coefficient of variation. We also derived analytic expressions for the coverage probability of each confidence interval. To confirm our theoretical results, Monte Carlo simulation will be used to assess the performance of these intervals based on their coverage probabilities.

Keywords: confidence interval, coverage probability, expected length, known coefficient of variation

Procedia PDF Downloads 357
2485 Parametric Estimation of U-Turn Vehicles

Authors: Yonas Masresha Aymeku

Abstract:

The purpose of capacity modelling at U-turns is to develop a relationship between capacity and its geometric characteristics. In fact, the few models available for the estimation of capacity at different transportation facilities do not provide specific guidelines for median openings. For this reason, an effort is made to estimate the capacity by collecting the data sets from median openings at different lane roads in Hyderabad City, India. Wide difference (43% -59%) among the capacity values estimated by the existing models shows the limitation to consider for mixed traffic situations. Thus, a distinct model is proposed for the estimation of the capacity of U-turn vehicles at median openings considering mixed traffic conditions, which would further prompt to investigate the effect of different factors that might affect the capacity.

Keywords: geometric, guiddelines, median, vehicles

Procedia PDF Downloads 24
2484 Power Quality Improvement Using Interval Type-2 Fuzzy Logic Controller for Five-Level Shunt Active Power Filter

Authors: Yousfi Abdelkader, Chaker Abdelkader, Bot Youcef

Abstract:

This article proposes a five-level shunt active power filter for power quality improvement using a interval type-2 fuzzy logic controller (IT2 FLC). The reference compensating current is extracted using the P-Q theory. The majority of works previously reported are based on two-level inverters with a conventional Proportional integral (PI) controller, which requires rigorous mathematical modeling of the system. In this paper, a IT2 FLC controlled five-level active power filter is proposed to overcome the problem associated with PI controller. The IT2 FLC algorithm is applied for controlling the DC-side capacitor voltage as well as the harmonic currents of the five-level active power filter. The active power filter with a IT2 FLC is simulated in MATLAB Simulink environment. The simulated response shows that the proposed shunt active power filter controller has produced a sinusoidal supply current with low harmonic distortion and in phase with the source voltage.

Keywords: power quality, shunt active power filter, interval type-2 fuzzy logic controller (T2FL), multilevel inverter

Procedia PDF Downloads 136
2483 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization

Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman

Abstract:

In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.

Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization

Procedia PDF Downloads 208
2482 Channel Estimation for Orthogonal Frequency Division Multiplexing Systems over Doubly Selective Channels Base on DCS-DCSOMP Algorithm

Authors: Linyu Wang, Furui Huo, Jianhong Xiang

Abstract:

The Doppler shift generated by high-speed movement and multipath effects in the channel are the main reasons for the generation of a time-frequency doubly-selective (DS) channel. There is severe inter-carrier interference (ICI) in the DS channel. Channel estimation for an orthogonal frequency division multiplexing (OFDM) system over a DS channel is very difficult. The simultaneous orthogonal matching pursuit algorithm under distributed compressive sensing theory (DCS-SOMP) has been used in channel estimation for OFDM systems over DS channels. However, the reconstruction accuracy of the DCS-SOMP algorithm is not high enough in the low SNR stage. To solve this problem, in this paper, we propose an improved DCS-SOMP algorithm based on the inner product difference comparison operation (DCS-DCSOMP). The reconstruction accuracy is improved by increasing the number of candidate indexes and designing the comparison conditions of inner product difference. We combine the DCS-DCSOMP algorithm with the basis expansion model (BEM) to reduce the complexity of channel estimation. Simulation results show the effectiveness of the proposed algorithm and its advantages over other algorithms.

Keywords: OFDM, doubly selective, channel estimation, compressed sensing

Procedia PDF Downloads 66
2481 The Prevalence and Associated Factors of Frailty and Its Relationship with Falls in Patients with Schizophrenia

Authors: Bo-Jian Wu, Si-Heng Wu

Abstract:

Objectives: Frailty is a condition of a person who has chronic health problems complicated by a loss of physiological reserve and deteriorating functional abilities. The frailty syndrome was defined by Fried and colleagues, i.e., weight loss, fatigue, decreased grip strength, slow gait speed, and low physical activity. However, to our best knowledge, there have been rare studies exploring the prevalence of frailty and its association with falls in patients with schizophrenia. Methods: A total of 559 hospitalized patients were recruited from a public psychiatric hospital in 2013. The majority of the subjects were males (361, 64.6%). The average age was 53.5 years. All patients received the assessment of frailty status defined by Fried and colleagues. The status of a fall within one year after the assessment of frailty, clinical and demographic data was collected from medical records. Logistic regression was used to calculate the odds ratio of associated factors. Results : A total of 9.2% of the participants met the criteria of frailty. The percentage of patients having a fall was 7.2%. Age were significantly associated with frailty (odds ratio = 1.057, 95% confidence interval = 1.025-1.091); however, sex was not associated with frailty (p = 0.17). After adjustment for age and sex, frailty status was associated with a fall (odds ratio = 3.62, 95% confidence interval = 1.58-8.28). Concerning the components of frailty, decreased grip strength (odds ratio = 2.44, 95% confidence interval = 1.16-5.14), slow gait speed (odds ratio = 2.82, 95% confidence interval = 1.21-6.53), and low physical activity (odds ratio = 2.64, 95% confidence interval = 1.21-5.78) were found to be associated with a fall. Conclusions: Our findings suggest the prevalence of frailty was about 10% in hospitalized patients with chronic patients with schizophrenia, and frailty status was significant with a fall in this group. By using the status of frailty, it may be beneficial to potential target candidates having fallen in the future as early as possible. The effective intervention of prevention of further falls may be given in advance. Our results bridge this gap and open a potential avenue for the prevention of falls in patients with schizophrenia. Frailty is certainly an important factor for maintaining wellbeing among these patients.

Keywords: fall, frailty, schizophrenia, Taiwan

Procedia PDF Downloads 131
2480 Mobile Platform’s Attitude Determination Based on Smoothed GPS Code Data and Carrier-Phase Measurements

Authors: Mohamed Ramdani, Hassen Abdellaoui, Abdenour Boudrassen

Abstract:

Mobile platform’s attitude estimation approaches mainly based on combined positioning techniques and developed algorithms; which aim to reach a fast and accurate solution. In this work, we describe the design and the implementation of an attitude determination (AD) process, using only measurements from GPS sensors. The major issue is based on smoothed GPS code data using Hatch filter and raw carrier-phase measurements integrated into attitude algorithm based on vectors measurement using least squares (LSQ) estimation method. GPS dataset from a static experiment is used to investigate the effectiveness of the presented approach and consequently to check the accuracy of the attitude estimation algorithm. Attitude results from GPS multi-antenna over short baselines are introduced and analyzed. The 3D accuracy of estimated attitude parameters using smoothed measurements is over 0.27°.

Keywords: attitude determination, GPS code data smoothing, hatch filter, carrier-phase measurements, least-squares attitude estimation

Procedia PDF Downloads 135
2479 Price Effect Estimation of Tobacco on Low-wage Male Smokers: A Causal Mediation Analysis

Authors: Kawsar Ahmed, Hong Wang

Abstract:

The study's goal was to estimate the causal mediation impact of tobacco tax before and after price hikes among low-income male smokers, with a particular emphasis on the effect estimating pathways framework for continuous and dichotomous variables. From July to December 2021, a cross-sectional investigation of observational data (n=739) was collected from Bangladeshi low-wage smokers. The Quasi-Bayesian technique, binomial probit model, and sensitivity analysis using a simulation of the computational tools R mediation package had been used to estimate the effect. After a price rise for tobacco products, the average number of cigarettes or bidis sticks taken decreased from 6.7 to 4.56. Tobacco product rising prices have a direct effect on low-income people's decisions to quit or lessen their daily smoking habits of Average Causal Mediation Effect (ACME) [effect=2.31, 95 % confidence interval (C.I.) = (4.71-0.00), p<0.01], Average Direct Effect (ADE) [effect=8.6, 95 percent (C.I.) = (6.8-0.11), p<0.001], and overall significant effects (p<0.001). Tobacco smoking choice is described by the mediated proportion of income effect, which is 26.1% less of following price rise. The curve of ACME and ADE is based on observational figures of the coefficients of determination that asses the model of hypothesis as the substantial consequence after price rises in the sensitivity analysis. To reduce smoking product behaviors, price increases through taxation have a positive causal mediation with income that affects the decision to limit tobacco use and promote low-income men's healthcare policy.

Keywords: causal mediation analysis, directed acyclic graphs, tobacco price policy, sensitivity analysis, pathway estimation

Procedia PDF Downloads 84
2478 Estimation of Chronic Kidney Disease Using Artificial Neural Network

Authors: Ilker Ali Ozkan

Abstract:

In this study, an artificial neural network model has been developed to estimate chronic kidney failure which is a common disease. The patients’ age, their blood and biochemical values, and 24 input data which consists of various chronic diseases are used for the estimation process. The input data have been subjected to preprocessing because they contain both missing values and nominal values. 147 patient data which was obtained from the preprocessing have been divided into as 70% training and 30% testing data. As a result of the study, artificial neural network model with 25 neurons in the hidden layer has been found as the model with the lowest error value. Chronic kidney failure disease has been able to be estimated accurately at the rate of 99.3% using this artificial neural network model. The developed artificial neural network has been found successful for the estimation of chronic kidney failure disease using clinical data.

Keywords: estimation, artificial neural network, chronic kidney failure disease, disease diagnosis

Procedia PDF Downloads 412
2477 On the Cluster of the Families of Hybrid Polynomial Kernels in Kernel Density Estimation

Authors: Benson Ade Eniola Afere

Abstract:

Over the years, kernel density estimation has been extensively studied within the context of nonparametric density estimation. The fundamental components of kernel density estimation are the kernel function and the bandwidth. While the mathematical exploration of the kernel component has been relatively limited, its selection and development remain crucial. The Mean Integrated Squared Error (MISE), serving as a measure of discrepancy, provides a robust framework for assessing the effectiveness of any kernel function. A kernel function with a lower MISE is generally considered to perform better than one with a higher MISE. Hence, the primary aim of this article is to create kernels that exhibit significantly reduced MISE when compared to existing classical kernels. Consequently, this article introduces a cluster of hybrid polynomial kernel families. The construction of these proposed kernel functions is carried out heuristically by combining two kernels from the classical polynomial kernel family using probability axioms. We delve into the analysis of error propagation within these kernels. To assess their performance, simulation experiments, and real-life datasets are employed. The obtained results demonstrate that the proposed hybrid kernels surpass their classical kernel counterparts in terms of performance.

Keywords: classical polynomial kernels, cluster of families, global error, hybrid Kernels, Kernel density estimation, Monte Carlo simulation

Procedia PDF Downloads 62
2476 Anisotropic Approach for Discontinuity Preserving in Optical Flow Estimation

Authors: Pushpendra Kumar, Sanjeev Kumar, R. Balasubramanian

Abstract:

Estimation of optical flow from a sequence of images using variational methods is one of the most successful approach. Discontinuity between different motions is one of the challenging problem in flow estimation. In this paper, we design a new anisotropic diffusion operator, which is able to provide smooth flow over a region and efficiently preserve discontinuity in optical flow. This operator is designed on the basis of intensity differences of the pixels and isotropic operator using exponential function. The combination of these are used to control the propagation of flow. Experimental results on the different datasets verify the robustness and accuracy of the algorithm and also validate the effect of anisotropic operator in the discontinuity preserving.

Keywords: optical flow, variational methods, computer vision, anisotropic operator

Procedia PDF Downloads 840
2475 Functional Decomposition Based Effort Estimation Model for Software-Intensive Systems

Authors: Nermin Sökmen

Abstract:

An effort estimation model is needed for software-intensive projects that consist of hardware, embedded software or some combination of the two, as well as high level software solutions. This paper first focuses on functional decomposition techniques to measure functional complexity of a computer system and investigates its impact on system development effort. Later, it examines effects of technical difficulty and design team capability factors in order to construct the best effort estimation model. With using traditional regression analysis technique, the study develops a system development effort estimation model which takes functional complexity, technical difficulty and design team capability factors as input parameters. Finally, the assumptions of the model are tested.

Keywords: functional complexity, functional decomposition, development effort, technical difficulty, design team capability, regression analysis

Procedia PDF Downloads 260
2474 Optical Flow Direction Determination for Railway Crossing Occupancy Monitoring

Authors: Zdenek Silar, Martin Dobrovolny

Abstract:

This article deals with the obstacle detection on a railway crossing (clearance detection). Detection is based on the optical flow estimation and classification of the flow vectors by K-means clustering algorithm. For classification of passing vehicles is used optical flow direction determination. The optical flow estimation is based on a modified Lucas-Kanade method.

Keywords: background estimation, direction of optical flow, K-means clustering, objects detection, railway crossing monitoring, velocity vectors

Procedia PDF Downloads 485
2473 The Effects of Three Months of HIIT on Plasma Adiponectin on Overweight College Men

Authors: M. J. Pourvaghar, M. E. Bahram, M. Sayyah, Sh. Khoshemehry

Abstract:

Adiponectin is a cytokine secreted by the adipose tissue that functions as an anti-inflammatory, antiathrogenic and anti-diabetic substance. Its density is inversely correlated with body mass index. The purpose of this research was to examine the effect of 12 weeks of high intensity interval training (HIIT) with the level of serum adiponectin and some selected adiposity markers in overweight and fat college students. This was a clinical research in which 24 students with BMI between 25 kg/m2 to 30 kg/m2. The sample was purposefully selected and then randomly assigned into two groups of experimental (age =22.7±1.5 yr.; weight = 85.8±3.18 kg and height =178.7±3.29 cm) and control (age =23.1±1.1 yr.; weight = 79.1±2.4 kg and height =181.3±4.6 cm), respectively. The experimental group participated in an aerobic exercise program for 12 weeks, three sessions per weeks at a high intensity between 85% to 95% of maximum heart rate (considering the over load principle). Prior and after the termination of exercise protocol, the level of serum adiponectin, BMI, waist to hip ratio, and body fat percentages were calculated. The data were analyzed by using SPSS: PC 16.0 and statistical procedure such as ANCOVA, was used. The results indicated that 12 weeks of intensive interval training led to the increase of serum adiponectin level and decrease of body weight, body fat percent, body mass index and waist to hip ratio (P < 0.05). Based on the results of this research, it may be concluded that participation in intensive interval training for 12 weeks is a non-invasive treatment to increase the adiponectin level while decreasing some of the anthropometric indices associated with obesity or being overweight.

Keywords: adiponectin, cardiovascular, interval, overweight, training

Procedia PDF Downloads 292
2472 Indoor Temperature Estimation with FIR Filter Using R-C Network Model

Authors: Sung Hyun You, Jeong Hoon Kim, Dae Ki Kim, Choon Ki Ahn

Abstract:

In this paper, we proposed a new strategy for estimating indoor temperature based on the modified resistance capacitance (R–C) network thermal dynamic model. Using minimum variance finite impulse response (FIR) filter, accurate indoor temperature estimation can be achieved. Our study is clarified by the experimental validation of the proposed indoor temperature estimation method. This experiment scenario environment is composed of a demand response (DR) server and home energy management system (HEMS) in a test bed.

Keywords: energy consumption, resistance-capacitance network model, demand response, finite impulse response filter

Procedia PDF Downloads 423