Search results for: forecasting accuracy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4001

Search results for: forecasting accuracy

2171 Accurate Algorithm for Selecting Ground Motions Satisfying Code Criteria

Authors: S. J. Ha, S. J. Baik, T. O. Kim, S. W. Han

Abstract:

For computing the seismic responses of structures, current seismic design provisions permit response history analyses (RHA) that can be used without limitations in height, seismic design category, and building irregularity. In order to obtain accurate seismic responses using RHA, it is important to use adequate input ground motions. Current seismic design provisions provide criteria for selecting ground motions. In this study, the accurate and computationally efficient algorithm is proposed for accurately selecting ground motions that satisfy the requirements specified in current seismic design provisions. The accuracy of the proposed algorithm is verified using single-degree-of-freedom systems with various natural periods and yield strengths. This study shows that the mean seismic responses obtained from RHA with seven and ten ground motions selected using the proposed algorithm produce errors within 20% and 13%, respectively.

Keywords: algorithm, ground motion, response history analysis, selection

Procedia PDF Downloads 277
2170 Comparisons of Surveying with Terrestrial Laser Scanner and Total Station for Volume Determination of Overburden and Coal Excavations in Large Open-Pit Mine

Authors: B. Keawaram, P. Dumrongchai

Abstract:

The volume of overburden and coal excavations in open-pit mine is generally determined by conventional survey such as total station. This study aimed to evaluate the accuracy of terrestrial laser scanner (TLS) used to measure overburden and coal excavations, and to compare TLS survey data sets with the data of the total station. Results revealed that, the reference points measured with the total station showed 0.2 mm precision for both horizontal and vertical coordinates. When using TLS on the same points, the standard deviations of 4.93 cm and 0.53 cm for horizontal and vertical coordinates, respectively, were achieved. For volume measurements covering the mining areas of 79,844 m2, TLS yielded the mean difference of about 1% and the surface error margin of 6 cm at the 95% confidence level when compared to the volume obtained by total station.

Keywords: mine, survey, terrestrial laser scanner, total station

Procedia PDF Downloads 368
2169 The Idea of Reputation in a Post-Truth Era

Authors: Karen Armstrong

Abstract:

This paper considers the importance of acquiring, cultivating, and protecting one’s personal online reputation in a post-truth era. Although the idea of the individual is essential psychological construct, the concept necessarily now includes our online reputation. The idea of this online reputation has expanded to become almost more important than any other factor in terms of our professional, social and psychological development. The discussion will first consider philosophical ideas of the self, followed by an examination of underlying concepts of perception and interpretation in a post-truth world. Then, the idea of the recent shift to a consideration of posted images, through words and photos, in the construction of self, will be discussed. Next, the relation between private personal life and exterior social life, including our reputation in a variety of realms will be addressed. This will include the adoption of specific strategies and behaviors, which facilitate accuracy, currency and necessary modifications with regard to our online reputation. Finally, specific ways in which we can negotiate the fluid dynamic between reputation, and inner and outer selves to optimum effect will conclude the discussion.

Keywords: image, post-truth, privacy, reputation, surveillance

Procedia PDF Downloads 243
2168 Identification of Impact Load and Partial System Parameters Using 1D-CNN

Authors: Xuewen Yu, Danhui Dan

Abstract:

The identification of impact load and some hard-to-obtain system parameters is crucial for the activities of analysis, validation, and evaluation in the engineering field. This paper proposes a method that utilizes neural networks based on 1D-CNN to identify the impact load and partial system parameters from measured responses. To this end, forward computations are conducted to provide datasets consisting of the triples (parameter θ, input u, output y). Then neural networks are trained to learn the mapping from input to output, fu|{θ} : y → u, as well as from input and output to parameter, fθ : (u, y) → θ. Afterward, feeding the trained neural networks the measured output response, the input impact load and system parameter can be calculated, respectively. The method is tested on two simulated examples and shows sound accuracy in estimating the impact load (waveform and location) and system parameters.

Keywords: convolutional neural network, impact load identification, system parameter identification, inverse problem

Procedia PDF Downloads 93
2167 The Effect of Choke on the Efficiency of Coaxial Antenna for Percutaneous Microwave Coagulation Therapy for Hepatic Tumor

Authors: Surita Maini

Abstract:

There are many perceived advantages of microwave ablation have driven researchers to develop innovative antennas to effectively treat deep-seated, non-resectable hepatic tumors. In this paper a coaxial antenna with a miniaturized sleeve choke has been discussed for microwave interstitial ablation therapy, in order to reduce backward heating effects irrespective of the insertion depth into the tissue. Two dimensional Finite Element Method (FEM) is used to simulate and measure the results of miniaturized sleeve choke antenna. This paper emphasizes the importance of factors that can affect simulation accuracy, which include mesh resolution, surface heating and reflection coefficient. Quarter wavelength choke effectiveness has been discussed by comparing it with the unchoked antenna with same dimensions.

Keywords: microwave ablation, tumor, finite element method, coaxial slot antenna, coaxial dipole antenna

Procedia PDF Downloads 346
2166 A Similarity Measure for Classification and Clustering in Image Based Medical and Text Based Banking Applications

Authors: K. P. Sandesh, M. H. Suman

Abstract:

Text processing plays an important role in information retrieval, data-mining, and web search. Measuring the similarity between the documents is an important operation in the text processing field. In this project, a new similarity measure is proposed. To compute the similarity between two documents with respect to a feature the proposed measure takes the following three cases into account: (1) The feature appears in both documents; (2) The feature appears in only one document and; (3) The feature appears in none of the documents. The proposed measure is extended to gauge the similarity between two sets of documents. The effectiveness of our measure is evaluated on several real-world data sets for text classification and clustering problems, especially in banking and health sectors. The results show that the performance obtained by the proposed measure is better than that achieved by the other measures.

Keywords: document classification, document clustering, entropy, accuracy, classifiers, clustering algorithms

Procedia PDF Downloads 505
2165 The Challenge of Navigating Long Tunnels

Authors: Ali Mohammadi

Abstract:

One of the concerns that employers and contractors have in creating long tunnels is that when the excavation is completed, the tunnel will be exited in the correct position according to designed, the deviation of the tunnel from its path can have many costs for the employer and the contractor, lack of correct calculations by the surveying engineer or the employer and contractors lack of importance to the surveying team in guiding the tunnel can cause the tunnel to deviate from its path and this deviation becomes a disaster. But employers are able to make the right decisions so that the tunnel is guided with the highest precision if they consider some points. We are investigating two tunnels with lengths of 12 and 18 kilometers that were dug by Tunnel boring machine machines to transfer water, how the contractor’s decision to control the 12 kilometer tunnel caused the most accuracy of one centimeter to the next part of the tunnel will be connected. We will also investigate the reasons for the deviation of axis in the 18 km tunnel about 20 meters. Also we review the calculations of surveyor engineers in both tunnels and what challenges there will be in the calculations and teach how to solve these challenges. Surveying calculations are the most important part in controlling long tunnels.

Keywords: UTM, localization, scale factor, traverse

Procedia PDF Downloads 59
2164 Application of Double Side Approach Method on Super Elliptical Winkler Plate

Authors: Hsiang-Wen Tang, Cheng-Ying Lo

Abstract:

In this study, the static behavior of super elliptical Winkler plate is analyzed by applying the double side approach method. The lack of information about super elliptical Winkler plates is the motivation of this study and we use the double side approach method to solve this problem because of its superior ability on efficiently treating problems with complex boundary shape. The double side approach method has the advantages of high accuracy, easy calculation procedure and less calculation load required. Most important of all, it can give the error bound of the approximate solution. The numerical results not only show that the double side approach method works well on this problem but also provide us the knowledge of static behavior of super elliptical Winkler plate in practical use.

Keywords: super elliptical winkler plate, double side approach method, error bound, mechanic

Procedia PDF Downloads 337
2163 Heuristic Approaches for Injury Reductions by Reduced Car Use in Urban Areas

Authors: Stig H. Jørgensen, Trond Nordfjærn, Øyvind Teige Hedenstrøm, Torbjørn Rundmo

Abstract:

The aim of the paper is to estimate and forecast road traffic injuries in the coming 10-15 years given new targets in urban transport policy and shifts of mode of transport, including injury cross-effects of mode changes. The paper discusses possibilities and limitations in measuring and quantifying possible injury reductions. Injury data (killed and seriously injured road users) from six urban areas in Norway from 1998-2012 (N= 4709 casualties) form the basis for estimates of changing injury patterns. For the coming period calculation of number of injuries and injury rates by type of road user (categories of motorized versus non-motorized) by sex, age and type of road are made. A prognosticated population increase (25 %) in total population within 2025 in the six urban areas will curb the proceeded fall in injury figures. However, policy strategies and measures geared towards a stronger modal shift from use of private vehicles to safer public transport (bus, train) will modify this effect. On the other side will door to door transport (pedestrians on their way to/from public transport nodes) imply a higher exposure for pedestrians (bikers) converting from private vehicle use (including fall accidents not registered as traffic accidents). The overall effect is the sum of these modal shifts in the increasing urban population and in addition diminishing return to the majority of road safety countermeasures has also to be taken into account. The paper demonstrates how uncertainties in the various estimates (prediction factors) on increasing injuries as well as decreasing injury figures may partly offset each other. The paper discusses road safety policy and welfare consequences of transport mode shift, including reduced use of private vehicles, and further environmental impacts. In this regard, safety and environmental issues will as a rule concur. However pursuing environmental goals (e.g. improved air quality, reduced co2 emissions) encouraging more biking may generate more biking injuries. The study was given financial grants from the Norwegian Research Council’s Transport Safety Program.

Keywords: road injuries, forecasting, reduced private care use, urban, Norway

Procedia PDF Downloads 225
2162 Vibration Analysis of Functionally Graded Engesser-Timoshenko Beams Subjected to Axial Load Located on a Continuous Elastic Foundation

Authors: M. Karami Khorramabadi, A. R. Nezamabadi

Abstract:

This paper studies free vibration of functionally graded beams Subjected to Axial Load that is simply supported at both ends lies on a continuous elastic foundation. The displacement field of beam is assumed based on Engesser-Timoshenko beam theory. The Young's modulus of beam is assumed to be graded continuously across the beam thickness. Applying the Hamilton's principle, the governing equation is established. Resulting equation is solved using the Euler's Equation. The effects of the constituent volume fractions and foundation coefficient on the vibration frequency are presented. To investigate the accuracy of the present analysis, a compression study is carried out with a known data.

Keywords: functionally graded beam, free vibration, elastic foundation, Engesser-Timoshenko beam theory

Procedia PDF Downloads 402
2161 Function Approximation with Radial Basis Function Neural Networks via FIR Filter

Authors: Kyu Chul Lee, Sung Hyun Yoo, Choon Ki Ahn, Myo Taeg Lim

Abstract:

Recent experimental evidences have shown that because of a fast convergence and a nice accuracy, neural networks training via extended Kalman filter (EKF) method is widely applied. However, as to an uncertainty of the system dynamics or modeling error, the performance of the method is unreliable. In order to overcome this problem in this paper, a new finite impulse response (FIR) filter based learning algorithm is proposed to train radial basis function neural networks (RBFN) for nonlinear function approximation. Compared to the EKF training method, the proposed FIR filter training method is more robust to those environmental conditions. Furthermore, the number of centers will be considered since it affects the performance of approximation.

Keywords: extended Kalman filter, classification problem, radial basis function networks (RBFN), finite impulse response (FIR) filter

Procedia PDF Downloads 444
2160 Influence of Temperature and Immersion on the Behavior of a Polymer Composite

Authors: Quentin C.P. Bourgogne, Vanessa Bouchart, Pierre Chevrier, Emmanuel Dattoli

Abstract:

This study presents an experimental and theoretical work conducted on a PolyPhenylene Sulfide reinforced with 40%wt of short glass fibers (PPS GF40) and its matrix. Thermoplastics are widely used in the automotive industry to lightweight automotive parts. The replacement of metallic parts by thermoplastics is reaching under-the-hood parts, near the engine. In this area, the parts are subjected to high temperatures and are immersed in cooling liquid. This liquid is composed of water and glycol and can affect the mechanical properties of the composite. The aim of this work was thus to quantify the evolution of mechanical properties of the thermoplastic composite, as a function of temperature and liquid aging effects, in order to develop a reliable design of parts. An experimental campaign in the tensile mode was carried out at different temperatures and for various glycol proportions in the cooling liquid, for monotonic and cyclic loadings on a neat and a reinforced PPS. The results of these tests allowed to highlight some of the main physical phenomena occurring during these solicitations under tough hydro-thermal conditions. Indeed, the performed tests showed that temperature and liquid cooling aging can affect the mechanical behavior of the material in several ways. The more the cooling liquid contains water, the more the mechanical behavior is affected. It was observed that PPS showed a higher sensitivity to absorption than to chemical aggressiveness of the cooling liquid, explaining this dominant sensitivity. Two kinds of behaviors were noted: an elasto-plastic type under the glass transition temperature and a visco-pseudo-plastic one above it. It was also shown that viscosity is the leading phenomenon above the glass transition temperature for the PPS and could also be important under this temperature, mostly under cyclic conditions and when the stress rate is low. Finally, it was observed that soliciting this composite at high temperatures is decreasing the advantages of the presence of fibers. A new phenomenological model was then built to take into account these experimental observations. This new model allowed the prediction of the evolution of mechanical properties as a function of the loading environment, with a reduced number of parameters compared to precedent studies. It was also shown that the presented approach enables the description and the prediction of the mechanical response with very good accuracy (2% of average error at worst), over a wide range of hydrothermal conditions. A temperature-humidity equivalence principle was underlined for the PPS, allowing the consideration of aging effects within the proposed model. Then, a limit of improvement of the reachable accuracy was determinate for all models using this set of data by the application of an artificial intelligence-based model allowing a comparison between artificial intelligence-based models and phenomenological based ones.

Keywords: aging, analytical modeling, mechanical testing, polymer matrix composites, sequential model, thermomechanical

Procedia PDF Downloads 102
2159 Molecular Dynamics Simulations of the Structural, Elastic, and Thermodynamic Properties of Cubic AlBi

Authors: M. Zemouli, K. Amara, M. Elkeurti, Y. Benallou

Abstract:

We present a theoretical study of the structural, elastic and thermodynamic properties of the zinc-blende AlBi for a wide temperature range. The simulation calculation is performed in the framework of the molecular dynamics method using the three-body Tersoff potential which reproduces provide, with reasonable accuracy, the lattice constants and elastic constants. Our results for the lattice constant, the bulk modulus and cohesive energy are in good agreement with other theoretical available works. Other thermodynamic properties such as the specific heat and the lattice thermal expansion can also be predicted. In addition, this method allows us to check its ability to predict the phase transition of this compound. In particular, the transition pressure to the rock-salt phase is calculated and the results are compared with other available works.

Keywords: aluminium compounds, molecular dynamics simulations, interatomic potential, thermodynamic properties, structural phase transition

Procedia PDF Downloads 293
2158 The Resistance Reader Program Based on Image Processing

Authors: Janpen Srijan, Nahathai Tanmang, Thanit Purathanang, Anun Dowchern, Saksit Summart, Seangduan Kampimpa

Abstract:

This paper presents the resistance reader program based on image processing by using MATLAB. The proposed program is divided into six parts; the first part is the web camera; the second part is a watt selection before shooting the resistor; the third part is a part of finding the position of the color on the mid-point of resistor; the fourth part is a part of identifying color code of the resistor; the fifth part is a part of taking the number of values for each color for resistance calculation and the last part is a part of displaying result of resistance value. The experimental result of the resistance reader program based on image processing was able to display the resistance value of resistor. The accuracy of proposed program is 85 percent for 1 watt resistor. It has 15 percent of reading error because a problem with the color code of some resistor was too bright.

Keywords: resistance reader program, image processing, resistor, MATLAB

Procedia PDF Downloads 369
2157 Annual Water Level Simulation Using Support Vector Machine

Authors: Maryam Khalilzadeh Poshtegal, Seyed Ahmad Mirbagheri, Mojtaba Noury

Abstract:

In this paper, by application of the input yearly data of rainfall, temperature and flow to the Urmia Lake, the simulation of water level fluctuation were applied by means of three models. According to the climate change investigation the fluctuation of lakes water level are of high interest. This study investigate data-driven models, support vector machines (SVM), SVM method which is a new regression procedure in water resources are applied to the yearly level data of Lake Urmia that is the biggest and the hyper saline lake in Iran. The evaluated lake levels are found to be in good correlation with the observed values. The results of SVM simulation show better accuracy and implementation. The mean square errors, mean absolute relative errors and determination coefficient statistics are used as comparison criteria.

Keywords: simulation, water level fluctuation, urmia lake, support vector machine

Procedia PDF Downloads 350
2156 An Earth Mover’s Distance Algorithm Based DDoS Detection Mechanism in SDN

Authors: Yang Zhou, Kangfeng Zheng, Wei Ni, Ren Ping Liu

Abstract:

Software-defined networking (SDN) provides a solution for scalable network framework with decoupled control and data plane. However, this architecture also induces a particular distributed denial-of-service (DDoS) attack that can affect or even overwhelm the SDN network. DDoS attack detection problem has to date been mostly researched as entropy comparison problem. However, this problem lacks the utilization of SDN, and the results are not accurate. In this paper, we propose a DDoS attack detection method, which interprets DDoS detection as a signature matching problem and is formulated as Earth Mover’s Distance (EMD) model. Considering the feasibility and accuracy, we further propose to define the cost function of EMD to be a generalized Kullback-Leibler divergence. Simulation results show that our proposed method can detect DDoS attacks by comparing EMD values with the ones computed in the case without attacks. Moreover, our method can significantly increase the true positive rate of detection.

Keywords: DDoS detection, EMD, relative entropy, SDN

Procedia PDF Downloads 320
2155 Computational Fluid Dynamics Study on Water Soot Blower Direction in Tangentially Fired Pulverized-Coal Boiler

Authors: Teewin Plangsrinont, Wasawat Nakkiew

Abstract:

In this study, computational fluid dynamics (CFD) was utilized to simulate and predict the path of water from water soot blower through an ambient flow field in 300-megawatt tangentially burned pulverized coal boiler that utilizes a water soot blower as a cleaning device. To predict the position of the impact of water on the opposite side of the water soot blower under identical conditions, the nozzle size and water flow rate were fixed in this investigation. The simulation findings demonstrated a high degree of accuracy in predicting the direction of water flow to the boiler's water wall tube, which was validated by comparison to experimental data. Results show maximum deviation value of the water jet trajectory is 10.2 percent.

Keywords: computational fluid dynamics, tangentially fired boiler, thermal power plant, water soot blower

Procedia PDF Downloads 193
2154 Temperature Coefficients of the Refractive Index for Ge Film

Authors: Lingmao Xu, Hui Zhou

Abstract:

Ge film is widely used in infrared optical systems. Because of the special requirements of space application, it is usually used in low temperature. The refractive index of Ge film is always changed with the temperature which has a great effect on the manufacture of high precision infrared optical film. Specimens of Ge single film were deposited at ZnSe substrates by EB-PVD method. During temperature range 80K ~ 300K, the transmittance of Ge single film within 2 ~ 15 μm were measured every 20K by PerkinElmer FTIR cryogenic testing system. By the full spectrum inversion method fitting, the relationship between refractive index and wavelength within 2 ~ 12μm at different temperatures was received. It can be seen the relationship consistent with the formula Cauchy, which can be fitted. Then the relationship between refractive index of the Ge film and temperature/wavelength was obtained by fitting method based on formula Cauchy. Finally, the designed value obtained by the formula and the measured spectrum were compared to verify the accuracy of the formula.

Keywords: infrared optical film, low temperature, thermal refractive coefficient, Ge film

Procedia PDF Downloads 286
2153 Use of Interpretable Evolved Search Query Classifiers for Sinhala Documents

Authors: Prasanna Haddela

Abstract:

Document analysis is a well matured yet still active research field, partly as a result of the intricate nature of building computational tools but also due to the inherent problems arising from the variety and complexity of human languages. Breaking down language barriers is vital in enabling access to a number of recent technologies. This paper investigates the application of document classification methods to new Sinhalese datasets. This language is geographically isolated and rich with many of its own unique features. We will examine the interpretability of the classification models with a particular focus on the use of evolved Lucene search queries generated using a Genetic Algorithm (GA) as a method of document classification. We will compare the accuracy and interpretability of these search queries with other popular classifiers. The results are promising and are roughly in line with previous work on English language datasets.

Keywords: evolved search queries, Sinhala document classification, Lucene Sinhala analyzer, interpretable text classification, genetic algorithm

Procedia PDF Downloads 100
2152 Development of Typical Meteorological Year for Passive Cooling Applications Using World Weather Data

Authors: Nasser A. Al-Azri

Abstract:

The effectiveness of passive cooling techniques is assessed based on bioclimatic charts that require the typical meteorological year (TMY) for a specified location for their development. However, TMYs are not always available; mainly due to the scarcity of records of solar radiation which is an essential component used in developing common TMYs intended for general uses. Since solar radiation is not required in the development of the bioclimatic chart, this work suggests developing TMYs based solely on the relevant parameters. This approach improves the accuracy of the developed TMY since only the relevant parameters are considered and it also makes the development of the TMY more accessible since solar radiation data are not used. The presented paper will also discuss the development of the TMY from the raw data available at the NOAA-NCDC archive of world weather data and the construction of the bioclimatic charts for some randomly selected locations around the world.

Keywords: bioclimatic charts, passive cooling, TMY, weather data

Procedia PDF Downloads 229
2151 Evaluation of Turbulence Modelling of Gas-Liquid Two-Phase Flow in a Venturi

Authors: Mengke Zhan, Cheng-Gang Xie, Jian-Jun Shu

Abstract:

A venturi flowmeter is a common device used in multiphase flow rate measurement in the upstream oil and gas industry. Having a robust computational model for multiphase flow in a venturi is desirable for understanding the gas-liquid and fluid-pipe interactions and predicting pressure and phase distributions under various flow conditions. A steady Eulerian-Eulerian framework is used to simulate upward gas-liquid flow in a vertical venturi. The simulation results are compared with experimental measurements of venturi differential pressure and chord-averaged gas holdup in the venturi throat section. The choice of turbulence model is nontrivial in the multiphase flow modelling in a venturi. The performance cross-comparison of the k-ϵ model, Reynolds stress model (RSM) and shear-stress transport (SST) k-ω turbulence model is made in the study. In terms of accuracy and computational cost, the SST k-ω turbulence model is observed to be the most efficient.

Keywords: computational fluid dynamics (CFD), gas-liquid flow, turbulence modelling, venturi

Procedia PDF Downloads 158
2150 A Supervised Face Parts Labeling Framework

Authors: Khalil Khan, Ikram Syed, Muhammad Ehsan Mazhar, Iran Uddin, Nasir Ahmad

Abstract:

Face parts labeling is the process of assigning class labels to each face part. A face parts labeling method (FPL) which divides a given image into its constitutes parts is proposed in this paper. A database FaceD consisting of 564 images is labeled with hand and make publically available. A supervised learning model is built through extraction of features from the training data. The testing phase is performed with two semantic segmentation methods, i.e., pixel and super-pixel based segmentation. In pixel-based segmentation class label is provided to each pixel individually. In super-pixel based method class label is assigned to super-pixel only – as a result, the same class label is given to all pixels inside a super-pixel. Pixel labeling accuracy reported with pixel and super-pixel based methods is 97.68 % and 93.45% respectively.

Keywords: face labeling, semantic segmentation, classification, face segmentation

Procedia PDF Downloads 240
2149 Pricing European Continuous-Installment Options under Regime-Switching Models

Authors: Saghar Heidari

Abstract:

In this paper, we study the valuation problem of European continuous-installment options under Markov-modulated models with a partial differential equation approach. Due to the opportunity for continuing or stopping to pay installments, the valuation problem under regime-switching models can be formulated as coupled partial differential equations (CPDE) with free boundary features. To value the installment options, we express the truncated CPDE as a linear complementarity problem (LCP), then a finite element method is proposed to solve the resulted variational inequality. Under some appropriate assumptions, we establish the stability of the method and illustrate some numerical results to examine the rate of convergence and accuracy of the proposed method for the pricing problem under the regime-switching model.

Keywords: continuous-installment option, European option, regime-switching model, finite element method

Procedia PDF Downloads 127
2148 A Nonlinear Feature Selection Method for Hyperspectral Image Classification

Authors: Pei-Jyun Hsieh, Cheng-Hsuan Li, Bor-Chen Kuo

Abstract:

For hyperspectral image classification, feature reduction is an important pre-processing for avoiding the Hughes phenomena due to the difficulty for collecting training samples. Hence, lots of researches developed feature selection methods such as F-score, HSIC (Hilbert-Schmidt Independence Criterion), and etc., to improve hyperspectral image classification. However, most of them only consider the class separability in the original space, i.e., a linear class separability. In this study, we proposed a nonlinear class separability measure based on kernel trick for selecting an appropriate feature subset. The proposed nonlinear class separability was formed by a generalized RBF kernel with different bandwidths with respect to different features. Moreover, it considered the within-class separability and the between-class separability. A genetic algorithm was applied to tune these bandwidths such that the smallest with-class separability and the largest between-class separability simultaneously. This indicates the corresponding feature space is more suitable for classification. In addition, the corresponding nonlinear classification boundary can separate classes very well. These optimal bandwidths also show the importance of bands for hyperspectral image classification. The reciprocals of these bandwidths can be viewed as weights of bands. The smaller bandwidth, the larger weight of the band, and the more importance for classification. Hence, the descending order of the reciprocals of the bands gives an order for selecting the appropriate feature subsets. In the experiments, three hyperspectral image data sets, the Indian Pine Site data set, the PAVIA data set, and the Salinas A data set, were used to demonstrate the selected feature subsets by the proposed nonlinear feature selection method are more appropriate for hyperspectral image classification. Only ten percent of samples were randomly selected to form the training dataset. All non-background samples were used to form the testing dataset. The support vector machine was applied to classify these testing samples based on selected feature subsets. According to the experiments on the Indian Pine Site data set with 220 bands, the highest accuracies by applying the proposed method, F-score, and HSIC are 0.8795, 0.8795, and 0.87404, respectively. However, the proposed method selects 158 features. F-score and HSIC select 168 features and 217 features, respectively. Moreover, the classification accuracies increase dramatically only using first few features. The classification accuracies with respect to feature subsets of 10 features, 20 features, 50 features, and 110 features are 0.69587, 0.7348, 0.79217, and 0.84164, respectively. Furthermore, only using half selected features (110 features) of the proposed method, the corresponding classification accuracy (0.84168) is approximate to the highest classification accuracy, 0.8795. For other two hyperspectral image data sets, the PAVIA data set and Salinas A data set, we can obtain the similar results. These results illustrate our proposed method can efficiently find feature subsets to improve hyperspectral image classification. One can apply the proposed method to determine the suitable feature subset first according to specific purposes. Then researchers can only use the corresponding sensors to obtain the hyperspectral image and classify the samples. This can not only improve the classification performance but also reduce the cost for obtaining hyperspectral images.

Keywords: hyperspectral image classification, nonlinear feature selection, kernel trick, support vector machine

Procedia PDF Downloads 253
2147 A Quantitative Evaluation of Text Feature Selection Methods

Authors: B. S. Harish, M. B. Revanasiddappa

Abstract:

Due to rapid growth of text documents in digital form, automated text classification has become an important research in the last two decades. The major challenge of text document representations are high dimension, sparsity, volume and semantics. Since the terms are only features that can be found in documents, selection of good terms (features) plays an very important role. In text classification, feature selection is a strategy that can be used to improve classification effectiveness, computational efficiency and accuracy. In this paper, we present a quantitative analysis of most widely used feature selection (FS) methods, viz. Term Frequency-Inverse Document Frequency (tfidf ), Mutual Information (MI), Information Gain (IG), CHISquare (x2), Term Frequency-Relevance Frequency (tfrf ), Term Strength (TS), Ambiguity Measure (AM) and Symbolic Feature Selection (SFS) to classify text documents. We evaluated all the feature selection methods on standard datasets like 20 Newsgroups, 4 University dataset and Reuters-21578.

Keywords: classifiers, feature selection, text classification

Procedia PDF Downloads 446
2146 Dynamic Foot Pressure Measurement System Using Optical Sensors

Authors: Tanapon Keatsamarn, Chuchart Pintavirooj

Abstract:

Foot pressure measurement provides necessary information for diagnosis diseases, foot insole design, disorder prevention and other application. In this paper, dynamic foot pressure measurement is presented for pressure measuring with high resolution and accuracy. The dynamic foot pressure measurement system consists of hardware and software system. The hardware system uses a transparent acrylic plate and uses steel as the base. The glossy white paper is placed on the top of the transparent acrylic plate and covering with a black acrylic on the system to block external light. Lighting from LED strip entering around the transparent acrylic plate. The optical sensors, the digital cameras, are underneath the acrylic plate facing upwards. They have connected with software system to process and record foot pressure video in avi file. Visual Studio 2017 is used for software system using OpenCV library.

Keywords: foot, foot pressure, image processing, optical sensors

Procedia PDF Downloads 234
2145 Revolutionizing Accounting: Unleashing the Power of Artificial Intelligence

Authors: Sogand Barghi

Abstract:

The integration of artificial intelligence (AI) in accounting practices is reshaping the landscape of financial management. This paper explores the innovative applications of AI in the realm of accounting, emphasizing its transformative impact on efficiency, accuracy, decision-making, and financial insights. By harnessing AI's capabilities in data analysis, pattern recognition, and automation, accounting professionals can redefine their roles, elevate strategic decision-making, and unlock unparalleled value for businesses. This paper delves into AI-driven solutions such as automated data entry, fraud detection, predictive analytics, and intelligent financial reporting, highlighting their potential to revolutionize the accounting profession. Artificial intelligence has swiftly emerged as a game-changer across industries, and accounting is no exception. This paper seeks to illuminate the profound ways in which AI is reshaping accounting practices, transcending conventional boundaries, and propelling the profession toward a new era of efficiency and insight-driven decision-making. One of the most impactful applications of AI in accounting is automation. Tasks that were once labor-intensive and time-consuming, such as data entry and reconciliation, can now be streamlined through AI-driven algorithms. This not only reduces the risk of errors but also allows accountants to allocate their valuable time to more strategic and analytical tasks. AI's ability to analyze vast amounts of data in real time enables it to detect irregularities and anomalies that might go unnoticed by traditional methods. Fraud detection algorithms can continuously monitor financial transactions, flagging any suspicious patterns and thereby bolstering financial security. AI-driven predictive analytics can forecast future financial trends based on historical data and market variables. This empowers organizations to make informed decisions, optimize resource allocation, and develop proactive strategies that enhance profitability and sustainability. Traditional financial reporting often involves extensive manual effort and data manipulation. With AI, reporting becomes more intelligent and intuitive. Automated report generation not only saves time but also ensures accuracy and consistency in financial statements. While the potential benefits of AI in accounting are undeniable, there are challenges to address. Data privacy and security concerns, the need for continuous learning to keep up with evolving AI technologies, and potential biases within algorithms demand careful attention. The convergence of AI and accounting marks a pivotal juncture in the evolution of financial management. By harnessing the capabilities of AI, accounting professionals can transcend routine tasks, becoming strategic advisors and data-driven decision-makers. The applications discussed in this paper underline the transformative power of AI, setting the stage for an accounting landscape that is smarter, more efficient, and more insightful than ever before. The future of accounting is here, and it's driven by artificial intelligence.

Keywords: artificial intelligence, accounting, automation, predictive analytics, financial reporting

Procedia PDF Downloads 52
2144 Information Technology Application for Knowledge Management in Medium-Size Businesses

Authors: S. Thongchai

Abstract:

Result of the study on knowledge management systems in businesses was shown that the most of these businesses provide internet accessibility for their employees in order to study new knowledge via internet, corporate website, electronic mail, and electronic learning system. These business organizations use information technology application for knowledge management because of convenience, time saving, ease of use, accuracy of information and knowledge usefulness. The result indicated prominent improvements for corporate knowledge management systems as the following; 1) administrations must support corporate knowledge management system 2) the goal of corporate knowledge management must be clear 3) corporate culture should facilitate the exchange and sharing of knowledge within the organization 4) cooperation of personnel of all levels must be obtained 5) information technology infrastructure must be provided 6) they must develop the system regularly and constantly.

Keywords: business organizations, information technology application, knowledge management systems, prominent improvements

Procedia PDF Downloads 374
2143 Investigation of New Gait Representations for Improving Gait Recognition

Authors: Chirawat Wattanapanich, Hong Wei

Abstract:

This study presents new gait representations for improving gait recognition accuracy on cross gait appearances, such as normal walking, wearing a coat and carrying a bag. Based on the Gait Energy Image (GEI), two ideas are implemented to generate new gait representations. One is to append lower knee regions to the original GEI, and the other is to apply convolutional operations to the GEI and its variants. A set of new gait representations are created and used for training multi-class Support Vector Machines (SVMs). Tests are conducted on the CASIA dataset B. Various combinations of the gait representations with different convolutional kernel size and different numbers of kernels used in the convolutional processes are examined. Both the entire images as features and reduced dimensional features by Principal Component Analysis (PCA) are tested in gait recognition. Interestingly, both new techniques, appending the lower knee regions to the original GEI and convolutional GEI, can significantly contribute to the performance improvement in the gait recognition. The experimental results have shown that the average recognition rate can be improved from 75.65% to 87.50%.

Keywords: convolutional image, lower knee, gait

Procedia PDF Downloads 190
2142 Investigation of the Progressive Collapse Potential in Steel Buildings with Composite Floor System

Authors: Pouya Kaafi, Gholamreza Ghodrati Amiri

Abstract:

Abnormal loads due to natural events, implementation errors and some other issues can lead to occurrence of progressive collapse in structures. Most of the past researches consist of 2- Dimensional (2D) models of steel frames without consideration of the floor system effects, which reduces the accuracy of the modeling. While employing a 3-Dimensional (3D) model and modeling the concrete slab system for the floors have a crucial role in the progressive collapse evaluation. In this research, a 3D finite element model of a 5-story steel building is modeled by the ABAQUS software once with modeling the slabs, and the next time without considering them. Then, the progressive collapse potential is evaluated. The results of the analyses indicate that the lack of the consideration of the slabs during the analyses, can lead to inaccuracy in assessing the progressive failure potential of the structure.

Keywords: abnormal loads, composite floor system, intermediate steel moment resisting frame system, progressive collapse

Procedia PDF Downloads 449