Search results for: Principle Component Regression
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2033

Search results for: Principle Component Regression

1673 A Renovated Cook's Distance Based On The Buckley-James Estimate In Censored Regression

Authors: Nazrina Aziz, Dong Q. Wang

Abstract:

There have been various methods created based on the regression ideas to resolve the problem of data set containing censored observations, i.e. the Buckley-James method, Miller-s method, Cox method, and Koul-Susarla-Van Ryzin estimators. Even though comparison studies show the Buckley-James method performs better than some other methods, it is still rarely used by researchers mainly because of the limited diagnostics analysis developed for the Buckley-James method thus far. Therefore, a diagnostic tool for the Buckley-James method is proposed in this paper. It is called the renovated Cook-s Distance, (RD* i ) and has been developed based on the Cook-s idea. The renovated Cook-s Distance (RD* i ) has advantages (depending on the analyst demand) over (i) the change in the fitted value for a single case, DFIT* i as it measures the influence of case i on all n fitted values Yˆ∗ (not just the fitted value for case i as DFIT* i) (ii) the change in the estimate of the coefficient when the ith case is deleted, DBETA* i since DBETA* i corresponds to the number of variables p so it is usually easier to look at a diagnostic measure such as RD* i since information from p variables can be considered simultaneously. Finally, an example using Stanford Heart Transplant data is provided to illustrate the proposed diagnostic tool.

Keywords: Buckley-James estimators, censored regression, censored data, diagnostic analysis, product-limit estimator, renovated Cook's Distance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1411
1672 Suitability of Black Box Approaches for the Reliability Assessment of Component-Based Software

Authors: Anjushi Verma, Tirthankar Gayen

Abstract:

Although, reliability is an important attribute of quality, especially for mission critical systems, yet, there does not exist any versatile model even today for the reliability assessment of component-based software. The existing Black Box models are found to make various assumptions which may not always be realistic and may be quite contrary to the actual behaviour of software. They focus on observing the manner in which the system behaves without considering the structure of the system, the components composing the system, their interconnections, dependencies, usage frequencies, etc.As a result, the entropy (uncertainty) in assessment using these models is much high.Though, there are some models based on operation profile yet sometimes it becomes extremely difficult to obtain the exact operation profile concerned with a given operation. This paper discusses the drawbacks, deficiencies and limitations of Black Box approaches from the perspective of various authors and finally proposes a conceptual model for the reliability assessment of software.

Keywords: Black Box, faults, failure, software reliability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1355
1671 6D Posture Estimation of Road Vehicles from Color Images

Authors: Yoshimoto Kurihara, Tad Gonsalves

Abstract:

Currently, in the field of object posture estimation, there is research on estimating the position and angle of an object by storing a 3D model of the object to be estimated in advance in a computer and matching it with the model. However, in this research, we have succeeded in creating a module that is much simpler, smaller in scale, and faster in operation. Our 6D pose estimation model consists of two different networks – a classification network and a regression network. From a single RGB image, the trained model estimates the class of the object in the image, the coordinates of the object, and its rotation angle in 3D space. In addition, we compared the estimation accuracy of each camera position, i.e., the angle from which the object was captured. The highest accuracy was recorded when the camera position was 75°, the accuracy of the classification was about 87.3%, and that of regression was about 98.9%.

Keywords: AlexNet, Deep learning, image recognition, 6D posture estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 546
1670 Lane Changing and Merging Maneuvers of Carlike Robots

Authors: Bibhya Sharma, Jito Vanualailai, Ravindra Rai

Abstract:

This research paper designs a unique motion planner of multiple platoons of nonholonomic car-like robots as a feasible solution to the lane changing/merging maneuvers. The decentralized planner with a leaderless approach and a path-guidance principle derived from the Lyapunov-based control scheme generates collision free avoidance and safe merging maneuvers from multiple lanes to a single lane by deploying a split/merge strategy. The fixed obstacles are the markings and boundaries of the road lanes, while the moving obstacles are the robots themselves. Real and virtual road lane markings and the boundaries of road lanes are incorporated into a workspace to achieve the desired formation and configuration of the robots. Convergence of the robots to goal configurations and the repulsion of the robots from specified obstacles are achieved by suitable attractive and repulsive potential field functions, respectively. The results can be viewed as a significant contribution to the avoidance algorithm of the intelligent vehicle systems (IVS). Computer simulations highlight the effectiveness of the split/merge strategy and the acceleration-based controllers.

Keywords: Lane merging, Lyapunov-based control scheme, path-guidance principle, split/merge strategy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1620
1669 How Do Crisis Affect Economic Policy?

Authors: Eva Kotlánová

Abstract:

After recession that began in 2007 in the United States and subsequently spilled over the Europe we could expect recovery of economic growth. According to the last estimation of economic progress of European countries, this recovery is not strong enough. Among others, it will depend on economic policy, where and in which way, the economic indicators will proceed. Economic theories postulate that the economic subjects prefer stably, continual economic policy without repeated and strong fluctuations. This policy is perceived as support of economic growth. Mostly in crises period, when the government must cope with consequences of recession, the economic policy becomes unpredictable for many subjects and economic policy uncertainty grows, which have negative influence on economic growth. The aim of this paper is to use panel regression to prove or disprove this hypothesis on the example of five largest European economies in the period 2008–2012.

Keywords: Economic Crises in Europe, Economic Policy, Uncertainty, Panel Analysis Regression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1503
1668 Quantitative Analysis of Weld Defect Images in Industrial Radiography Based Invariant Attributes

Authors: N. Nacereddine, M. Tridi, S. S. Belaïfa, M. Zelmat

Abstract:

For the characterization of the weld defect region in the radiographic image, looking for features which are invariant regarding the geometrical transformations (rotation, translation and scaling) proves to be necessary because the same defect can be seen from several angles according to the orientation and the distance from the welded framework to the radiation source. Thus, panoply of geometrical attributes satisfying the above conditions is proposed and which result from the calculation of the geometrical parameters (surface, perimeter, etc.) on the one hand and the calculation of the different order moments, on the other hand. Because the large range in values of the raw features and taking into account other considerations imposed by some classifiers, the scaling of these values to lie between 0 and 1 is indispensable. The principal component analysis technique is used in order to reduce the number of the attribute variables in the aim to give better performance to the further defect classification.

Keywords: Geometric parameters, invariant attributes, principal component analysis, weld defect image.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2157
1667 Multi-Objective Optimization in End Milling of Al-6061 Using Taguchi Based G-PCA

Authors: M. K. Pradhan, Mayank Meena, Shubham Sen, Arvind Singh

Abstract:

In this study, a multi objective optimization for end milling of Al 6061 alloy has been presented to provide better surface quality and higher Material Removal Rate (MRR). The input parameters considered for the analysis are spindle speed, depth of cut and feed. The experiments were planned as per Taguchis design of experiment, with L27 orthogonal array. The Grey Relational Analysis (GRA) has been used for transforming multiple quality responses into a single response and the weights of the each performance characteristics are determined by employing the Principal Component Analysis (PCA), so that their relative importance can be properly and objectively described. The results reveal that Taguchi based G-PCA can effectively acquire the optimal combination of cutting parameters.

Keywords: Material Removal Rate, Surface Roughness, Taguchi Method, Grey Relational Analysis, Principal Component Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2203
1666 Modelling Dengue Fever (DF) and Dengue Haemorrhagic Fever (DHF) Outbreak Using Poisson and Negative Binomial Model

Authors: W. Y. Wan Fairos, W. H. Wan Azaki, L. Mohamad Alias, Y. Bee Wah

Abstract:

Dengue fever has become a major concern for health authorities all over the world particularly in the tropical countries. These countries, in particular are experiencing the most worrying outbreak of dengue fever (DF) and dengue haemorrhagic fever (DHF). The DF and DHF epidemics, thus, have become the main causes of hospital admissions and deaths in Malaysia. This paper, therefore, attempts to examine the environmental factors that may influence the recent dengue outbreak. The aim of this study is twofold, firstly is to establish a statistical model to describe the relationship between the number of dengue cases and a range of explanatory variables and secondly, to identify the lag operator for explanatory variables which affect the dengue incidence the most. The explanatory variables involved include the level of cloud cover, percentage of relative humidity, amount of rainfall, maximum temperature, minimum temperature and wind speed. The Poisson and Negative Binomial regression analyses were used in this study. The results of the analyses on the 915 observations (daily data taken from July 2006 to Dec 2008), reveal that the climatic factors comprising of daily temperature and wind speed were found to significantly influence the incidence of dengue fever after 2 and 3 weeks of their occurrences. The effect of humidity, on the other hand, appears to be significant only after 2 weeks.

Keywords: Dengue Fever, Dengue Hemorrhagic Fever, Negative Binomial Regression model, Poisson Regression model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2780
1665 Application of Residual Correction Method on Hyperbolic Thermoelastic Response of Hollow Spherical Medium in Rapid Transient Heat Conduction

Authors: Po-Jen Su, Huann-Ming Chou

Abstract:

In this article, we used the residual correction method to deal with transient thermoelastic problems with a hollow spherical region when the continuum medium possesses spherically isotropic thermoelastic properties. Based on linear thermoelastic theory, the equations of hyperbolic heat conduction and thermoelastic motion were combined to establish the thermoelastic dynamic model with consideration of the deformation acceleration effect and non-Fourier effect under the condition of transient thermal shock. The approximate solutions of temperature and displacement distributions are obtained using the residual correction method based on the maximum principle in combination with the finite difference method, making it easier and faster to obtain upper and lower approximations of exact solutions. The proposed method is found to be an effective numerical method with satisfactory accuracy. Moreover, the result shows that the effect of transient thermal shock induced by deformation acceleration is enhanced by non-Fourier heat conduction with increased peak stress. The influence on the stress increases with the thermal relaxation time.

Keywords: Maximum principle, non-Fourier heat conduction, residual correction method, thermo-elastic response.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1727
1664 Clustering of Variables Based On a Probabilistic Approach Defined on the Hypersphere

Authors: Paulo Gomes, Adelaide Figueiredo

Abstract:

We consider n individuals described by p standardized variables, represented by points of the surface of the unit hypersphere Sn-1. For a previous choice of n individuals we suppose that the set of observables variables comes from a mixture of bipolar Watson distribution defined on the hypersphere. EM and Dynamic Clusters algorithms are used for identification of such mixture. We obtain estimates of parameters for each Watson component and then a partition of the set of variables into homogeneous groups of variables. Additionally we will present a factor analysis model where unobservable factors are just the maximum likelihood estimators of Watson directional parameters, exactly the first principal component of data matrix associated to each group previously identified. Such alternative model it will yield us to directly interpretable solutions (simple structure), avoiding factors rotations.

Keywords: Dynamic Clusters algorithm, EM algorithm, Factor analysis model, Hierarchical Clustering, Watson distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1594
1663 Free Flapping Vibration of Rotating Inclined Euler Beams

Authors: Chih-Ling Huang, Wen-Yi Lin, Kuo-Mo Hsiao

Abstract:

A method based on the power series solution is proposed to solve the natural frequency of flapping vibration for the rotating inclined Euler beam with constant angular velocity. The vibration of the rotating beam is measured from the position of the corresponding steady state axial deformation. In this paper the governing equations for linear vibration of a rotating Euler beam are derived by the d'Alembert principle, the virtual work principle and the consistent linearization of the fully geometrically nonlinear beam theory in a rotating coordinate system. The governing equation for flapping vibration of the rotating inclined Euler beam is linear ordinary differential equation with variable coefficients and is solved by a power series with four independent coefficients. Substituting the power series solution into the corresponding boundary conditions at two end nodes of the rotating beam, a set of homogeneous equations can be obtained. The natural frequencies may be determined by solving the homogeneous equations using the bisection method. Numerical examples are studied to investigate the effect of inclination angle on the natural frequency of flapping vibration for rotating inclined Euler beams with different angular velocity and slenderness ratio.

Keywords: Flapping vibration, Inclination angle, Natural frequency, Rotating beam.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2157
1662 Alumina Supported Copper-Manganese Catalysts for Combustion of Exhaust Gases: Effect of Preparation Method

Authors: Krasimir I. Ivanov, Elitsa N. Kolentsova, Dimitar Y. Dimitrov

Abstract:

The development of active and stable catalysts without noble metals for low temperature oxidation of exhaust gases remains a significant challenge. The purpose of this study is to determine the influence of the preparation method on the catalytic activity of the supported copper-manganese mixed oxides in terms of VOCs oxidation. The catalysts were prepared by impregnation of γ- Al2O3 with copper and manganese nitrates and acetates and the possibilities for CO, CH3OH and dimethyl ether (DME) oxidation were evaluated using continuous flow equipment with a four-channel isothermal stainless steel reactor. Effect of the support, Cu/Mn mole ratio, heat treatment of the precursor and active component loading were investigated. Highly active alumina supported Cu-Mn catalysts for CO and VOCs oxidation were synthesized. The effect of preparation conditions on the activity behavior of the catalysts was discussed. The synergetic interaction between copper and manganese species increases the activity for complete oxidation over mixed catalysts. Type of support, calcination temperature and active component loading along with catalyst composition are important factors, determining catalytic activity. Cu/Mn molar ratio of 1:5, heat treatment at 450oC and 20 % active component loading are the best compromise for production of active catalyst for simultaneous combustion of CO, CH3OH and DME.

Keywords: Copper-manganese catalysts, Preparation methods, Exhaust gases oxidation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2311
1661 A Taguchi Approach to Investigate Impact of Factors for Reusability of Software Components

Authors: Parvinder S. Sandhu, Pavel Blecharz, Hardeep Singh

Abstract:

Quantitative Investigation of impact of the factors' contribution towards measuring the reusability of software components could be helpful in evaluating the quality of developed or developing reusable software components and in identification of reusable component from existing legacy systems; that can save cost of developing the software from scratch. But the issue of the relative significance of contributing factors has remained relatively unexplored. In this paper, we have use the Taguchi's approach in analyzing the significance of different structural attributes or factors in deciding the reusability level of a particular component. The results obtained shows that the complexity is the most important factor in deciding the better Reusability of a function oriented Software. In case of Object Oriented Software, Coupling and Complexity collectively play significant role in high reusability.

Keywords: Taguchi Approach, Reusability, SoftwareComponents, Structural Attributes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1691
1660 The Effect of Biochar, Inoculated Biochar and Compost Biological Component of the Soil

Authors: H. Dvořáčková, I. Mikajlo, J. Záhora, J. Elbl

Abstract:

Biochar can be produced from the waste matter and its application has been associated with returning of carbon in large amounts into the soil. The impacts of this material on physical and chemical properties of soil have been described. The biggest part of the research work is dedicated to the hypothesis of this material’s toxic effects on the soil life regarding its effect on the soil biological component. At present, it has been worked on methods which could eliminate these undesirable properties of biochar. One of the possibilities is to mix biochar with organic material, such as compost, or focusing on the natural processes acceleration in the soil. In the experiment has been used as the addition of compost as well as the elimination of toxic substances by promoting microbial activity in aerated water environment. Biochar was aerated for 7 days in a container with a volume of 20 l. This way modified biochar had six times higher biomass production and reduce mineral nitrogen leaching. Better results have been achieved by mixing biochar with compost.

Keywords: Leaching of nitrogen, soil, biochar, compost.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2998
1659 A Comprehensive Evaluation of Supervised Machine Learning for the Phase Identification Problem

Authors: Brandon Foggo, Nanpeng Yu

Abstract:

Power distribution circuits undergo frequent network topology changes that are often left undocumented. As a result, the documentation of a circuit’s connectivity becomes inaccurate with time. The lack of reliable circuit connectivity information is one of the biggest obstacles to model, monitor, and control modern distribution systems. To enhance the reliability and efficiency of electric power distribution systems, the circuit’s connectivity information must be updated periodically. This paper focuses on one critical component of a distribution circuit’s topology - the secondary transformer to phase association. This topology component describes the set of phase lines that feed power to a given secondary transformer (and therefore a given group of power consumers). Finding the documentation of this component is call Phase Identification, and is typically performed with physical measurements. These measurements can take time lengths on the order of several months, but with supervised learning, the time length can be reduced significantly. This paper compares several such methods applied to Phase Identification for a large range of real distribution circuits, describes a method of training data selection, describes preprocessing steps unique to the Phase Identification problem, and ultimately describes a method which obtains high accuracy (> 96% in most cases, > 92% in the worst case) using only 5% of the measurements typically used for Phase Identification.

Keywords: Distribution network, machine learning, network topology, phase identification, smart grid.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1032
1658 The Mechanical Response of a Composite Propellant under Harsh Conditions

Authors: Xin Tong, Jin-sheng Xu, Xiong Chen, Ya Zheng

Abstract:

The aim of this paper is to study the mechanical properties of HTPB (Hydroxyl-terminated polybutadiene) composite propellant under harsh conditions. It describes two tests involving uniaxial tensile tests of various strain rates (ranging from 0.0005 s-1 to 1.5 s-1), temperatures (ranging from 223 K to 343 K) and high-cycle fatigue tests under low-temperature (223 K, frequencies were set at 50, 100, 150 Hz) using DMA (Dynamic Mechanical Analyzer). To highlight the effect of small pre-strain on fatigue properties of HTPB propellant, quasi-static stretching was carried out before fatigue loading, and uniaxial tensile tests at constant strain rates were successively applied. The results reveal that flow stress of propellant increases with reduction in temperature and rise in strain rate, and the strain rate-temperature equivalence relationship could be described by TTSP (time-temperature superposition principle) incorporating a modified WLF equation. Moreover, the rate of performance degradations and damage accumulation of propellant during fatigue tests increased with increasing strain amplitude and loading frequencies, while initial quasi-static loading has a negative effect on fatigue properties by comparing stress-strain relations after fatigue tests.

Keywords: Fatigue, HTPB propellant, tensile properties, time-temperature superposition principle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1040
1657 Exploring the Physical Environment and Building Features in Earthquake Disaster Areas

Authors: Chang Hsueh-Sheng, Chen Tzu-Ling

Abstract:

Earthquake is an unpredictable natural disaster and intensive earthquakes have caused serious impacts on social-economic system, environmental and social resilience. Conventional ways to mitigate earthquake disaster are to enhance building codes and advance structural engineering measures. However, earthquake-induced ground damage such as liquefaction, land subsidence, landslide happen on places nearby earthquake prone or poor soil condition areas. Therefore, this study uses spatial statistical analysis to explore the spatial pattern of damaged buildings. Afterwards, principle components analysis (PCA) is applied to categorize the similar features in different kinds of clustered patterns. The results show that serious landslide prone area, close to fault, vegetated ground surface and mudslide prone area are common in those highly damaged buildings. In addition, the oldest building might not be directly referred to the most vulnerable one. In fact, it seems that buildings built between 1974 and 1989 become more fragile during the earthquake. The incorporation of both spatial statistical analyses and PCA can provide more accurate information to subsidize retrofit programs to enhance earthquake resistance in particular areas.

Keywords: Earthquake disaster, spatial statistical analysis, principle components analysis, clustered patterns.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1347
1656 Choosing between the Regression Correlation, the Rank Correlation, and the Correlation Curve

Authors: Roger L Goodwin

Abstract:

This paper presents a rank correlation curve. The traditional correlation coefficient is valid for both continuous variables and for integer variables using rank statistics. Since the correlation coefficient has already been established in rank statistics by Spearman, such a calculation can be extended to the correlation curve. This paper presents two survey questions. The survey collected non-continuous variables. We will show weak to moderate correlation. Obviously, one question has a negative effect on the other. A review of the qualitative literature can answer which question and why. The rank correlation curve shows which collection of responses has a positive slope and which collection of responses has a negative slope. Such information is unavailable from the flat, ”first-glance” correlation statistics.

Keywords: Bayesian estimation, regression model, rank statistics, correlation, correlation curve.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1587
1655 Spatial Distribution and Risk Assessment of As, Hg, Co and Cr in Kaveh Industrial City, using Geostatistic and GIS

Authors: Abbas Hani

Abstract:

The concentrations of As, Hg, Co, Cr and Cd were tested for each soil sample, and their spatial patterns were analyzed by the semivariogram approach of geostatistics and geographical information system technology. Multivariate statistic approaches (principal component analysis and cluster analysis) were used to identify heavy metal sources and their spatial pattern. Principal component analysis coupled with correlation between heavy metals showed that primary inputs of As, Hg and Cd were due to anthropogenic while, Co, and Cr were associated with pedogenic factors. Ordinary kriging was carried out to map the spatial patters of heavy metals. The high pollution sources evaluated was related with usage of urban and industrial wastewater. The results of this study helpful for risk assessment of environmental pollution for decision making for industrial adjustment and remedy soil pollution.

Keywords: Geographic Information system, Geostatistics, Kaveh, Multivariate Statistical Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1940
1654 The Use of Degradation Measures to Design Reliability Test Plans

Authors: Stephen V. Crowder, Jonathan W. Lane

Abstract:

With short production development times, there is an increased need to demonstrate product reliability relatively quickly with minimal testing. In such cases there may be few if any observed failures. Thus it may be difficult to assess reliability using the traditional reliability test plans that measure only time (or cycles) to failure. For many components, degradation measures will contain important information about performance and reliability. These measures can be used to design a minimal test plan, in terms of number of units placed on test and duration of the test, necessary to demonstrate a reliability goal. In this work we present a case study involving an electronic component subject to degradation. The data, consisting of 42 degradation paths of cycles to failure, are first used to estimate a reliability function. Bootstrapping techniques are then used to perform power studies and develop a minimal reliability test plan for future production of this component. 

Keywords: Degradation Measure, Time to Failure Distribution, Bootstrap.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1849
1653 Automatic Sleep Stage Scoring with Wavelet Packets Based on Single EEG Recording

Authors: Luay A. Fraiwan, Natheer Y. Khaswaneh, Khaldon Y. Lweesy

Abstract:

Sleep stage scoring is the process of classifying the stage of the sleep in which the subject is in. Sleep is classified into two states based on the constellation of physiological parameters. The two states are the non-rapid eye movement (NREM) and the rapid eye movement (REM). The NREM sleep is also classified into four stages (1-4). These states and the state wakefulness are distinguished from each other based on the brain activity. In this work, a classification method for automated sleep stage scoring based on a single EEG recording using wavelet packet decomposition was implemented. Thirty two ploysomnographic recording from the MIT-BIH database were used for training and validation of the proposed method. A single EEG recording was extracted and smoothed using Savitzky-Golay filter. Wavelet packets decomposition up to the fourth level based on 20th order Daubechies filter was used to extract features from the EEG signal. A features vector of 54 features was formed. It was reduced to a size of 25 using the gain ratio method and fed into a classifier of regression trees. The regression trees were trained using 67% of the records available. The records for training were selected based on cross validation of the records. The remaining of the records was used for testing the classifier. The overall correct rate of the proposed method was found to be around 75%, which is acceptable compared to the techniques in the literature.

Keywords: Features selection, regression trees, sleep stagescoring, wavelet packets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2302
1652 Fault Detection and Identification of COSMED K4b2 Based On PCA and Neural Network

Authors: Jing Zhou, Steven Su, Aihuang Guo

Abstract:

COSMED K4b2 is a portable electrical device designed to test pulmonary functions. It is ideal for many applications that need the measurement of the cardio-respiratory response either in the field or in the lab is capable with the capability to delivery real time data to a sink node or a PC base station with storing data in the memory at the same time. But the actual sensor outputs and data received may contain some errors, such as impulsive noise which can be related to sensors, low batteries, environment or disturbance in data acquisition process. These abnormal outputs might cause misinterpretations of exercise or living activities to persons being monitored. In our paper we propose an effective and feasible method to detect and identify errors in applications by principal component analysis (PCA) and a back propagation (BP) neural network.

Keywords: BP Neural Network, Exercising Testing, Fault Detection and Identification, Principal Component Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3046
1651 Video Quality Control Using a ROI and Two- Component Weighted Metrics

Authors: Petra Heribanová, Jaroslav Polec, Michal Martinovič

Abstract:

In this paper we propose a new content-weighted method for full reference (FR) video quality control using a region of interest (ROI) and wherein two-component weighted metrics for Deaf People Video Communication. In our approach, an image is partitioned into region of interest and into region "dry-as-dust", then region of interest is partitioned into two parts: edges and background (smooth regions), while the another methods (metrics) combined and weighted three or more parts as edges, edges errors, texture, smooth regions, blur, block distance etc. as we proposed. Using another idea that different image regions from deaf people video communication have different perceptual significance relative to quality. Intensity edges certainly contain considerable image information and are perceptually significant.

Keywords: Video quality assessment, weighted MSE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951
1650 A Video Watermarking Algorithm Based on Chaotic and Wavelet Neural Network

Authors: Jiadong Liang

Abstract:

This paper presented a video watermarking algorithm based on wavelet chaotic neural network. First, to enhance binary image’s security, the algorithm encrypted it with double chaotic based on Arnold and Logistic map, Then, the host video was divided into some equal frames and distilled the key frame through chaotic sequence which generated by Logistic. Meanwhile, we distilled the low frequency coefficients of luminance component and self-adaptively embedded the processed image watermark into the low frequency coefficients of the wavelet transformed luminance component with the wavelet neural network. The experimental result suggested that the presented algorithm has better invisibility and robustness against noise, Gaussian filter, rotation, frame loss and other attacks.

Keywords: Video watermark, double chaotic encryption, wavelet neural network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1026
1649 Walsh-Hadamard Transform for Facial Feature Extraction in Face Recognition

Authors: M. Hassan, I. Osman, M. Yahia

Abstract:

This Paper proposes a new facial feature extraction approach, Wash-Hadamard Transform (WHT). This approach is based on correlation between local pixels of the face image. Its primary advantage is the simplicity of its computation. The paper compares the proposed approach, WHT, which was traditionally used in data compression with two other known approaches: the Principal Component Analysis (PCA) and the Discrete Cosine Transform (DCT) using the face database of Olivetti Research Laboratory (ORL). In spite of its simple computation, the proposed algorithm (WHT) gave very close results to those obtained by the PCA and DCT. This paper initiates the research into WHT and the family of frequency transforms and examines their suitability for feature extraction in face recognition applications.

Keywords: Face Recognition, Facial Feature Extraction, Principal Component Analysis, and Discrete Cosine Transform, Wash-Hadamard Transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2539
1648 Comparative Study - Three Artificial Intelligence Techniques for Rain Domain in Precipitation Forecast

Authors: Nabilah Filzah Mohd Radzuan, Andi Putra, Zalinda Othman, Azuraliza Abu Bakar, Abdul Razak Hamdan

Abstract:

Precipitation forecast is important in avoid incident of natural disaster which can cause loss in involved area. This review paper involves three techniques from artificial intelligence namely logistic regression, decisions tree, and random forest which used in making precipitation forecast. These combination techniques through VAR model in finding advantages and strength for every technique in forecast process. Data contains variables from rain domain. Adaptation of artificial intelligence techniques involved on rain domain enables the process to be easier and systematic for precipitation forecast.

Keywords: Logistic regression, decisions tree, random forest, VAR model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2008
1647 Determinants of the U.S. Current Account

Authors: Shuh Liang

Abstract:

This article provides empirical evidence on the effect of domestic and international factors on the U.S. current account deficit. Linear dynamic regression and vector autoregression models are employed to estimate the relationships during the period from 1986 to 2011. The findings of this study suggest that the current and lagged private saving rate and foreign current account for East Asian economies have played a vital role in affecting the U.S. current account. Additionally, using Granger causality tests and variance decompositions, the change of the productivity growth and foreign domestic demand are determined to influence significantly the change of the U.S. current account. To summarize, the empirical relationship between the U.S. current account deficit and its determinants is sensitive to alternative regression models and specifications.

Keywords: Current account deficit, productivity growth, foreign demand, vector autoregression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1689
1646 Wear and Friction Analysis of Sintered Metal Powder Self Lubricating Bush Bearing

Authors: J. K. Khare, Abhay Kumar Sharma, Ajay Tiwari, Amol A. Talankar

Abstract:

Powder metallurgy (P/M) is the only economic way to produce porous parts/products. P/M can produce near net shape parts hence reduces wastage of raw material and energy, avoids various machining operations. The most vital use of P/M is in production of metallic filters and self lubricating bush bearings and siding surfaces. The porosity of the part can be controlled by varying compaction pressure, sintering temperature and composition of metal powder mix. The present work is aimed for experimental analysis of friction and wear properties of self lubricating copper and tin bush bearing. Experimental results confirm that wear rate of sintered component is lesser for components having 10% tin by weight percentage. Wear rate increases for high tin percentage (experimented for 20% tin and 30% tin) at same sintering temperature. Experimental results also confirms that wear rate of sintered component is also dependent on sintering temperature, soaking period, composition of the preform, compacting pressure, powder particle shape and size. Interfacial friction between die and punch, between inter powder particles, between die face and powder particle depends on compaction pressure, powder particle size and shape, size and shape of component which decides size & shape of die & punch, material of die & punch and material of powder particles.

Keywords: Interfacial friction, porous bronze bearing, sintering temperature, wear rate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3946
1645 Template-Based Object Detection through Partial Shape Matching and Boundary Verification

Authors: Feng Ge, Tiecheng Liu, Song Wang, Joachim Stahl

Abstract:

This paper presents a novel template-based method to detect objects of interest from real images by shape matching. To locate a target object that has a similar shape to a given template boundary, the proposed method integrates three components: contour grouping, partial shape matching, and boundary verification. In the first component, low-level image features, including edges and corners, are grouped into a set of perceptually salient closed contours using an extended ratio-contour algorithm. In the second component, we develop a partial shape matching algorithm to identify the fractions of detected contours that partly match given template boundaries. Specifically, we represent template boundaries and detected contours using landmarks, and apply a greedy algorithm to search the matched landmark subsequences. For each matched fraction between a template and a detected contour, we estimate an affine transform that transforms the whole template into a hypothetic boundary. In the third component, we provide an efficient algorithm based on oriented edge lists to determine the target boundary from the hypothetic boundaries by checking each of them against image edges. We evaluate the proposed method on recognizing and localizing 12 template leaves in a data set of real images with clutter back-grounds, illumination variations, occlusions, and image noises. The experiments demonstrate the high performance of our proposed method1.

Keywords: Object detection, shape matching, contour grouping.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2264
1644 Statistical Wavelet Features, PCA, and SVM Based Approach for EEG Signals Classification

Authors: R. K. Chaurasiya, N. D. Londhe, S. Ghosh

Abstract:

The study of the electrical signals produced by neural activities of human brain is called Electroencephalography. In this paper, we propose an automatic and efficient EEG signal classification approach. The proposed approach is used to classify the EEG signal into two classes: epileptic seizure or not. In the proposed approach, we start with extracting the features by applying Discrete Wavelet Transform (DWT) in order to decompose the EEG signals into sub-bands. These features, extracted from details and approximation coefficients of DWT sub-bands, are used as input to Principal Component Analysis (PCA). The classification is based on reducing the feature dimension using PCA and deriving the supportvectors using Support Vector Machine (SVM). The experimental are performed on real and standard dataset. A very high level of classification accuracy is obtained in the result of classification.

Keywords: Discrete Wavelet Transform, Electroencephalogram, Pattern Recognition, Principal Component Analysis, Support Vector Machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3081