Search results for: Short Time Fourier Transform.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7626

Search results for: Short Time Fourier Transform.

5646 Bayesian Networks for Earthquake Magnitude Classification in a Early Warning System

Authors: G. Zazzaro, F.M. Pisano, G. Romano

Abstract:

During last decades, worldwide researchers dedicated efforts to develop machine-based seismic Early Warning systems, aiming at reducing the huge human losses and economic damages. The elaboration time of seismic waveforms is to be reduced in order to increase the time interval available for the activation of safety measures. This paper suggests a Data Mining model able to correctly and quickly estimate dangerousness of the running seismic event. Several thousand seismic recordings of Japanese and Italian earthquakes were analyzed and a model was obtained by means of a Bayesian Network (BN), which was tested just over the first recordings of seismic events in order to reduce the decision time and the test results were very satisfactory. The model was integrated within an Early Warning System prototype able to collect and elaborate data from a seismic sensor network, estimate the dangerousness of the running earthquake and take the decision of activating the warning promptly.

Keywords: Bayesian Networks, Decision Support System, Magnitude Classification, Seismic Early Warning System

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3573
5645 Dynamic Analysis of Nonlinear Models with Infinite Extension by Boundary Elements

Authors: Delfim Soares Jr., Webe J. Mansur

Abstract:

The Time-Domain Boundary Element Method (TDBEM) is a well known numerical technique that handles quite properly dynamic analyses considering infinite dimension media. However, when these analyses are also related to nonlinear behavior, very complex numerical procedures arise considering the TD-BEM, which may turn its application prohibitive. In order to avoid this drawback and model nonlinear infinite media, the present work couples two BEM formulations, aiming to achieve the best of two worlds. In this context, the regions expected to behave nonlinearly are discretized by the Domain Boundary Element Method (D-BEM), which has a simpler mathematical formulation but is unable to deal with infinite domain analyses; the TD-BEM is employed as in the sense of an effective non-reflexive boundary. An iterative procedure is considered for the coupling of the TD-BEM and D-BEM, which is based on a relaxed renew of the variables at the common interfaces. Elastoplastic models are focused and different time-steps are allowed to be considered by each BEM formulation in the coupled analysis.

Keywords: Boundary Element Method, Dynamic Elastoplastic Analysis, Iterative Coupling, Multiple Time-Steps.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1516
5644 An Asymptotic Solution for the Free Boundary Parabolic Equations

Authors: Hsuan-Ku Liu, Ming Long Liu

Abstract:

In this paper, we investigate the solution of a two dimensional parabolic free boundary problem. The free boundary of this problem is modelled as a nonlinear integral equation (IE). For this integral equation, we propose an asymptotic solution as time is near to maturity and develop an integral iterative method. The computational results reveal that our asymptotic solution is very close to the numerical solution as time is near to maturity.

Keywords: Integral equation, asymptotic solution, free boundary problem, American exchange option.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1451
5643 Improving Co-integration Trading Rule Profitability with Forecasts from an Artificial Neural Network

Authors: Paul Lajbcygier, Seng Lee

Abstract:

Co-integration models the long-term, equilibrium relationship of two or more related financial variables. Even if cointegration is found, in the short run, there may be deviations from the long run equilibrium relationship. The aim of this work is to forecast these deviations using neural networks and create a trading strategy based on them. A case study is used: co-integration residuals from Australian Bank Bill futures are forecast and traded using various exogenous input variables combined with neural networks. The choice of the optimal exogenous input variables chosen for each neural network, undertaken in previous work [1], is validated by comparing the forecasts and corresponding profitability of each, using a trading strategy.

Keywords: Artificial neural networks, co-integration, forecasting, trading rule.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1225
5642 Prioritization Method in the Fuzzy Analytic Network Process by Fuzzy Preferences Programming Method

Authors: Tarifa S. Almulhim, Ludmil Mikhailov, Dong-Ling Xu

Abstract:

In this paper, a method for deriving a group priority vector in the Fuzzy Analytic Network Process (FANP) is proposed. By introducing importance weights of multiple decision makers (DMs) based on their experiences, the Fuzzy Preferences Programming Method (FPP) is extended to a fuzzy group prioritization problem in the FANP. Additionally, fuzzy pair-wise comparison judgments are presented rather than exact numerical assessments in order to model the uncertainty and imprecision in the DMs- judgments and then transform the fuzzy group prioritization problem into a fuzzy non-linear programming optimization problem which maximize the group satisfaction. Unlike the known fuzzy prioritization techniques, the new method proposed in this paper can easily derive crisp weights from incomplete and inconsistency fuzzy set of comparison judgments and does not require additional aggregation producers. Detailed numerical examples are used to illustrate the implement of our approach and compare with the latest fuzzy prioritization method.

Keywords: Fuzzy Analytic Network Process (FANP), Fuzzy Non-linear Programming, Fuzzy Preferences Programming Method (FPP), Multiple Criteria Decision-Making (MCDM), Triangular Fuzzy Number.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2366
5641 Data Transformation Services (DTS): Creating Data Mart by Consolidating Multi-Source Enterprise Operational Data

Authors: J. D. D. Daniel, K. N. Goh, S. M. Yusop

Abstract:

Trends in business intelligence, e-commerce and remote access make it necessary and practical to store data in different ways on multiple systems with different operating systems. As business evolve and grow, they require efficient computerized solution to perform data update and to access data from diverse enterprise business applications. The objective of this paper is to demonstrate the capability of DTS [1] as a database solution for automatic data transfer and update in solving business problem. This DTS package is developed for the sales of variety of plants and eventually expanded into commercial supply and landscaping business. Dimension data modeling is used in DTS package to extract, transform and load data from heterogeneous database systems such as MySQL, Microsoft Access and Oracle that consolidates into a Data Mart residing in SQL Server. Hence, the data transfer from various databases is scheduled to run automatically every quarter of the year to review the efficient sales analysis. Therefore, DTS is absolutely an attractive solution for automatic data transfer and update which meeting today-s business needs.

Keywords: Data Transformation Services (DTS), ObjectLinking and Embedding Database (OLEDB), Data Mart, OnlineAnalytical Processing (OLAP), Online Transactional Processing(OLTP).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2011
5640 Wavelet-Based Classification of Myocardial Ischemia, Arrhythmia, Congestive Heart Failure and Sleep Apnea

Authors: Santanu Chattopadhyay, Gautam Sarkar, Arabinda Das

Abstract:

This paper presents wavelet based classification of various heart diseases. Electrocardiogram signals of different heart patients have been studied. Statistical natures of electrocardiogram signals for different heart diseases have been compared with the statistical nature of electrocardiograms for normal persons. Under this study four different heart diseases have been considered as follows: Myocardial Ischemia (MI), Congestive Heart Failure (CHF), Arrhythmia and Sleep Apnea. Statistical nature of electrocardiograms for each case has been considered in terms of kurtosis values of two types of wavelet coefficients: approximate and detail. Nine wavelet decomposition levels have been considered in each case. Kurtosis corresponding to both approximate and detail coefficients has been considered for decomposition level one to decomposition level nine. Based on significant difference, few decomposition levels have been chosen and then used for classification.

Keywords: Arrhythmia, congestive heart failure, discrete wavelet transform, electrocardiogram, myocardial ischemia, sleep apnea.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 644
5639 Financing Decision and Productivity Growth for the Venture Capital Industry Using High-Order Fuzzy Time Series

Authors: Shang-En Yu

Abstract:

Human society, there are many uncertainties, such as economic growth rate forecast of the financial crisis, many scholars have, since the the Song Chissom two scholars in 1993 the concept of the so-called fuzzy time series (Fuzzy Time Series)different mode to deal with these problems, a previous study, however, usually does not consider the relevant variables selected and fuzzy process based solely on subjective opinions the fuzzy semantic discrete, so can not objectively reflect the characteristics of the data set, in addition to carrying outforecasts are often fuzzy rules as equally important, failed to consider the importance of each fuzzy rule. For these reasons, the variable selection (Factor Selection) through self-organizing map (Self-Organizing Map, SOM) and proposed high-end weighted multivariate fuzzy time series model based on fuzzy neural network (Fuzzy-BPN), and using the the sequential weighted average operator (Ordered Weighted Averaging operator, OWA) weighted prediction. Therefore, in order to verify the proposed method, the Taiwan stock exchange (Taiwan Stock Exchange Corporation) Taiwan Weighted Stock Index (Taiwan Stock Exchange Capitalization Weighted Stock Index, TAIEX) as experimental forecast target, in order to filter the appropriate variables in the experiment Finally, included in other studies in recent years mode in conjunction with this study, the results showed that the predictive ability of this study further improve.

Keywords: Heterogeneity, residential mortgage loans, foreclosure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1369
5638 Copy-Move Image Forgery Detection in Virtual Electrostatic Field

Authors: Michael Zimba, Darlison Nyirenda

Abstract:

A novel copy-move image forgery, CMIF, detection method is proposed. The proposed method presents a new approach which relies on electrostatic field theory, EFT. Solely for the purpose of reducing the dimension of a suspicious image, the proposed algorithm firstly performs discrete wavelet transform, DWT, of the suspicious image and extracts only the approximation subband. The extracted subband is then bijectively mapped onto a virtual electrostatic field where concepts of EFT are utilized to extract robust features. The extracted features are invariant to additive noise, JPEG compression, and affine transformation. Finally, same affine transformation selection, SATS, a duplication verification method, is applied to detect duplicated regions. SATS is a better option than the common shift vector method because SATS is insensitive to affine transformation. Consequently, the proposed CMIF algorithm is not only fast but also more robust to attacks compared to the existing related CMIF algorithms. The experimental results show high detection rates, as high as 100% in some cases.

Keywords: Affine transformation, Radix sort, SATS, Virtual electrostatic field.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1796
5637 Functional and Efficient Query Interpreters: Principle, Application and Performances’ Comparison

Authors: Laurent Thiry, Michel Hassenforder

Abstract:

This paper presents a general approach to implement efficient queries’ interpreters in a functional programming language. Indeed, most of the standard tools actually available use an imperative and/or object-oriented language for the implementation (e.g. Java for Jena-Fuseki) but other paradigms are possible with, maybe, better performances. To proceed, the paper first explains how to model data structures and queries in a functional point of view. Then, it proposes a general methodology to get performances (i.e. number of computation steps to answer a query) then it explains how to integrate some optimization techniques (short-cut fusion and, more important, data transformations). It then compares the functional server proposed to a standard tool (Fuseki) demonstrating that the first one can be twice to ten times faster to answer queries.

Keywords: Data transformation, functional programming, information server, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 725
5636 Hydrolysis of Hull-Less Pumpkin Oil Cake Protein Isolate by Pepsin

Authors: Ivan Živanović, Žužana Vaštag, Senka Popović, Ljiljana Popović, Draginja Peričin

Abstract:

The present work represents an investigation of the hydrolysis of hull-less pumpkin (Cucurbita Pepo L.) oil cake protein isolate (PuOC PI) by pepsin. To examine the effectiveness and suitability of pepsin towards PuOC PI the kinetic parameters for pepsin on PuOC PI were determined and then, the hydrolysis process was studied using Response Surface Methodology (RSM). The hydrolysis was carried out at temperature of 30°C and pH 3.00. Time and initial enzyme/substrate ratio (E/S) at three levels were selected as the independent parameters. The degree of hydrolysis, DH, was mesuared after 20, 30 and 40 minutes, at initial E/S of 0.7, 1 and 1.3 mA/mg proteins. Since the proposed second-order polynomial model showed good fit with the experimental data (R2 = 0.9822), the obtained mathematical model could be used for monitoring the hydrolysis of PuOC PI by pepsin, under studied experimental conditions, varying the time and initial E/S. To achieve the highest value of DH (39.13 %), the obtained optimum conditions for time and initial E/S were 30 min and 1.024 mA/mg proteins.

Keywords: Enzymatic hydrolysis, Pepsin, Pumpkin (CucurbitaPepo L.) oil cake protein isolate, Response surface methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2155
5635 Negative Selection as a Means of Discovering Unknown Temporal Patterns

Authors: Wanli Ma, Dat Tran, Dharmendra Sharma

Abstract:

The temporal nature of negative selection is an under exploited area. In a negative selection system, newly generated antibodies go through a maturing phase, and the survivors of the phase then wait to be activated by the incoming antigens after certain number of matches. These without having enough matches will age and die, while these with enough matches (i.e., being activated) will become active detectors. A currently active detector may also age and die if it cannot find any match in a pre-defined (lengthy) period of time. Therefore, what matters in a negative selection system is the dynamics of the involved parties in the current time window, not the whole time duration, which may be up to eternity. This property has the potential to define the uniqueness of negative selection in comparison with the other approaches. On the other hand, a negative selection system is only trained with “normal" data samples. It has to learn and discover unknown “abnormal" data patterns on the fly by itself. Consequently, it is more appreciate to utilize negation selection as a system for pattern discovery and recognition rather than just pattern recognition. In this paper, we study the potential of using negative selection in discovering unknown temporal patterns.

Keywords: Artificial Immune Systems, ComputationalIntelligence, Negative Selection, Pattern Discovery.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1643
5634 Integrating Big Island Layout with Pull System for Production Optimization

Authors: M. H. M. Rusli, A. Jaffar, M. T. Ali, S. Muhamud @ Kayat

Abstract:

Lean manufacturing is a production philosophy made popular by Toyota Motor Corporation (TMC). It is globally known as the Toyota Production System (TPS) and has the ultimate aim of reducing cost by thoroughly eliminating wastes or muda. TPS embraces the Just-in-time (JIT) manufacturing; achieving cost reduction through lead time reduction. JIT manufacturing can be achieved by implementing Pull system in the production. Furthermore, TPS aims to improve productivity and creating continuous flow in the production by arranging the machines and processes in cellular configurations. This is called as Cellular Manufacturing Systems (CMS). This paper studies on integrating the CMS with the Pull system to establish a Big Island-Pull system production for High Mix Low Volume (HMLV) products in an automotive component industry. The paper will use the build-in JIT system steps adapted from TMC to create the Pull system production and also create a shojinka line which, according to takt time, has the flexibility to adapt to demand changes simply by adding and taking out manpower. This will lead to optimization in production.

Keywords: Big Island layout, Lean manufacturing, Material and Information Flow Chart, Pull system production, TPS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2567
5633 Amplitude and Phase Analysis of EEG Signal by Complex Demodulation

Authors: Sun K. Yoo, Hee Cheol Kang

Abstract:

Analysis of amplitude and phase characteristics for delta, theta, and alpha bands at localized time instant from EEG signals is important for the characterizing information processing in the brain. In this paper, complex demodulation method was used to analyze EEG (Electroencephalographic) signal, particularly for auditory evoked potential response signal, with sufficient time resolution and designated frequency bandwidth resolution required. The complex demodulation decomposes raw EEG signal into 3 designated delta, theta, and alpha bands with complex EEG signal representation at sampled time instant, which can enable the extraction of amplitude envelope and phase information. Throughout simulated test data, and real EEG signal acquired during auditory attention task, it can extract the phase offset, phase and frequency changing instant and decomposed amplitude envelope for delta, theta, and alpha bands. The complex demodulation technique can be efficiently used in brain signal analysis in case of phase, and amplitude information required.

Keywords: EEG, Complex Demodulation, Amplitude, Phase.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4728
5632 Phylogenetic Characterization of Atrazine-Degrading Bacteria Isolated from Agricultural Soil in Eastern Thailand

Authors: Sawangjit Sopid

Abstract:

In this study sugarcane field soils with a long history of atrazine application in Chachoengsao and Chonburi provinces have been explored for their potential of atrazine biodegradation. For the atrazine degrading bacteria isolation, the soils used in this study named ACS and ACB were inoculated in MS-medium containing atrazine. Six short rod and gram-negative bacterial isolates, which were able to use this herbicide as a sole source of nitrogen, were isolated and named as ACS1, ACB1, ACB3, ACB4, ACB5 and ACB6. From the 16S rDNA nucleotide sequence analysis, the isolated bacteria ACS1 and ACB4 were identified as Rhizobium sp. with 89.1-98.7% nucleotide identity, ACB1 and ACB5 were identified as Stenotrophomonas sp. with 91.0-92.8% nucleotide identity, whereas ACB3 and ACB6 were Klebsiella sp. with 97.4-97.8% nucleotide identity.

Keywords: Atrazine-degrading bacteria, bioremediation, Thai isolate bacteria.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2188
5631 Nonlinear Dynamical Characterization of Heart Rate Variability Time Series of Meditation

Authors: B. S. Raghavendra, D. Narayana Dutt

Abstract:

Many recent electrophysiological studies have revealed the importance of investigating meditation state in order to achieve an increased understanding of autonomous control of cardiovascular functions. In this paper, we characterize heart rate variability (HRV) time series acquired during meditation using nonlinear dynamical parameters. We have computed minimum embedding dimension (MED), correlation dimension (CD), largest Lyapunov exponent (LLE), and nonlinearity scores (NLS) from HRV time series of eight Chi and four Kundalini meditation practitioners. The pre-meditation state has been used as a baseline (control) state to compare the estimated parameters. The chaotic nature of HRV during both pre-meditation and meditation is confirmed by MED. The meditation state showed a significant decrease in the value of CD and increase in the value of LLE of HRV, in comparison with premeditation state, indicating a less complex and less predictable nature of HRV. In addition, it was shown that the HRV of meditation state is having highest NLS than pre-meditation state. The study indicated highly nonlinear dynamic nature of cardiac states as revealed by HRV during meditation state, rather considering it as a quiescent state.

Keywords: Correlation dimension, Embedding dimension, Heartrate variability, Largest Lyapunov exponent, Meditation, Nonlinearity score.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1882
5630 Spatial-Temporal Clustering Characteristics of Dengue in the Northern Region of Sri Lanka, 2010-2013

Authors: Sumiko Anno, Keiji Imaoka, Takeo Tadono, Tamotsu Igarashi, Subramaniam Sivaganesh, Selvam Kannathasan, Vaithehi Kumaran, Sinnathamby Noble Surendran

Abstract:

Dengue outbreaks are affected by biological, ecological, socio-economic and demographic factors that vary over time and space. These factors have been examined separately and still require systematic clarification. The present study aimed to investigate the spatial-temporal clustering relationships between these factors and dengue outbreaks in the northern region of Sri Lanka. Remote sensing (RS) data gathered from a plurality of satellites were used to develop an index comprising rainfall, humidity and temperature data. RS data gathered by ALOS/AVNIR-2 were used to detect urbanization, and a digital land cover map was used to extract land cover information. Other data on relevant factors and dengue outbreaks were collected through institutions and extant databases. The analyzed RS data and databases were integrated into geographic information systems, enabling temporal analysis, spatial statistical analysis and space-time clustering analysis. Our present results showed that increases in the number of the combination of ecological factor and socio-economic and demographic factors with above the average or the presence contribute to significantly high rates of space-time dengue clusters.

Keywords: ALOS/AVNIR-2, Dengue, Space-time clustering analysis, Sri Lanka.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2261
5629 Affordable and Environmental Friendly Small Commuter Aircraft Improving European Mobility

Authors: Diego Giuseppe Romano, Gianvito Apuleo, Jiri Duda

Abstract:

Mobility is one of the most important societal needs for amusement, business activities and health. Thus, transport needs are continuously increasing, with the consequent traffic congestion and pollution increase. Aeronautic effort aims at smarter infrastructures use and in introducing greener concepts. A possible solution to address the abovementioned topics is the development of Small Air Transport (SAT) system, able to guarantee operability from today underused airfields in an affordable and green way, helping meanwhile travel time reduction, too. In the framework of Horizon2020, EU (European Union) has funded the Clean Sky 2 SAT TA (Transverse Activity) initiative to address market innovations able to reduce SAT operational cost and environmental impact, ensuring good levels of operational safety. Nowadays, most of the key technologies to improve passenger comfort and to reduce community noise, DOC (Direct Operating Costs) and pilot workload for SAT have reached an intermediate level of maturity TRL (Technology Readiness Level) 3/4. Thus, the key technologies must be developed, validated and integrated on dedicated ground and flying aircraft demonstrators to reach higher TRL levels (5/6). Particularly, SAT TA focuses on the integration at aircraft level of the following technologies [1]: 1)    Low-cost composite wing box and engine nacelle using OoA (Out of Autoclave) technology, LRI (Liquid Resin Infusion) and advance automation process. 2) Innovative high lift devices, allowing aircraft operations from short airfields (< 800 m). 3) Affordable small aircraft manufacturing of metallic fuselage using FSW (Friction Stir Welding) and LMD (Laser Metal Deposition). 4)       Affordable fly-by-wire architecture for small aircraft (CS23 certification rules). 5) More electric systems replacing pneumatic and hydraulic systems (high voltage EPGDS -Electrical Power Generation and Distribution System-, hybrid de-ice system, landing gear and brakes). 6) Advanced avionics for small aircraft, reducing pilot workload. 7) Advanced cabin comfort with new interiors materials and more comfortable seats. 8) New generation of turboprop engine with reduced fuel consumption, emissions, noise and maintenance costs for 19 seats aircraft. (9) Alternative diesel engine for 9 seats commuter aircraft. To address abovementioned market innovations, two different platforms have been designed: Reference and Green aircraft. Reference aircraft is a virtual aircraft designed considering 2014 technologies with an existing engine assuring requested take-off power; Green aircraft is designed integrating the technologies addressed in Clean Sky 2. Preliminary integration of the proposed technologies shows an encouraging reduction of emissions and operational costs of small: about 20% CO2 reduction, about 24% NOx reduction, about 10 db (A) noise reduction at measurement point and about 25% DOC reduction. Detailed description of the performed studies, analyses and validations for each technology as well as the expected benefit at aircraft level are reported in the present paper.

Keywords: Affordable, European, green, mobility, technologies development, travel time reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 503
5628 A Programmer’s Survey of the Quantum Computing Paradigm

Authors: Philippe Jorrand

Abstract:

Research in quantum computation is looking for the consequences of having information encoding, processing and communication exploit the laws of quantum physics, i.e. the laws which govern the ultimate knowledge that we have, today, of the foreign world of elementary particles, as described by quantum mechanics. This paper starts with a short survey of the principles which underlie quantum computing, and of some of the major breakthroughs brought by the first ten to fifteen years of research in this domain; quantum algorithms and quantum teleportation are very biefly presented. The next sections are devoted to one among the many directions of current research in the quantum computation paradigm, namely quantum programming languages and their semantics. A few other hot topics and open problems in quantum information processing and communication are mentionned in few words in the concluding remarks, the most difficult of them being the physical implementation of a quantum computer. The interested reader will find a list of useful references at the end of the paper.

Keywords: Quantum information processing, quantum algorithms, quantum programming languages.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1983
5627 Retrospective Reconstruction of Time Series Data for Integrated Waste Management

Authors: A. Buruzs, M. F. Hatwágner, A. Torma, L. T. Kóczy

Abstract:

The development, operation and maintenance of Integrated Waste Management Systems (IWMS) affects essentially the sustainable concern of every region. The features of such systems have great influence on all of the components of sustainability. In order to reach the optimal way of processes, a comprehensive mapping of the variables affecting the future efficiency of the system is needed such as analysis of the interconnections among the components and modeling of their interactions. The planning of a IWMS is based fundamentally on technical and economical opportunities and the legal framework. Modeling the sustainability and operation effectiveness of a certain IWMS is not in the scope of the present research. The complexity of the systems and the large number of the variables require the utilization of a complex approach to model the outcomes and future risks. This complex method should be able to evaluate the logical framework of the factors composing the system and the interconnections between them. The authors of this paper studied the usability of the Fuzzy Cognitive Map (FCM) approach modeling the future operation of IWMS’s. The approach requires two input data set. One is the connection matrix containing all the factors affecting the system in focus with all the interconnections. The other input data set is the time series, a retrospective reconstruction of the weights and roles of the factors. This paper introduces a novel method to develop time series by content analysis.

Keywords: Content analysis, factors, integrated waste management system, time series.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2003
5626 View-Point Insensitive Human Pose Recognition using Neural Network and CUDA

Authors: Sanghyeok Oh, Keechul Jung

Abstract:

Although lots of research work has been done for human pose recognition, the view-point of cameras is still critical problem of overall recognition system. In this paper, view-point insensitive human pose recognition is proposed. The aims of the proposed system are view-point insensitivity and real-time processing. Recognition system consists of feature extraction module, neural network and real-time feed forward calculation. First, histogram-based method is used to extract feature from silhouette image and it is suitable for represent the shape of human pose. To reduce the dimension of feature vector, Principle Component Analysis(PCA) is used. Second, real-time processing is implemented by using Compute Unified Device Architecture(CUDA) and this architecture improves the speed of feed-forward calculation of neural network. We demonstrate the effectiveness of our approach with experiments on real environment.

Keywords: computer vision, neural network, pose recognition, view-point insensitive, PCA, CUDA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1320
5625 Support Vector Machine based Intelligent Watermark Decoding for Anticipated Attack

Authors: Syed Fahad Tahir, Asifullah Khan, Abdul Majid, Anwar M. Mirza

Abstract:

In this paper, we present an innovative scheme of blindly extracting message bits from an image distorted by an attack. Support Vector Machine (SVM) is used to nonlinearly classify the bits of the embedded message. Traditionally, a hard decoder is used with the assumption that the underlying modeling of the Discrete Cosine Transform (DCT) coefficients does not appreciably change. In case of an attack, the distribution of the image coefficients is heavily altered. The distribution of the sufficient statistics at the receiving end corresponding to the antipodal signals overlap and a simple hard decoder fails to classify them properly. We are considering message retrieval of antipodal signal as a binary classification problem. Machine learning techniques like SVM is used to retrieve the message, when certain specific class of attacks is most probable. In order to validate SVM based decoding scheme, we have taken Gaussian noise as a test case. We generate a data set using 125 images and 25 different keys. Polynomial kernel of SVM has achieved 100 percent accuracy on test data.

Keywords: Bit Correct Ratio (BCR), Grid Search, Intelligent Decoding, Jackknife Technique, Support Vector Machine (SVM), Watermarking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1649
5624 Adsorption of Lead(II) and Cadmium(II) Ions from Aqueous Solutions by Adsorption on Activated Carbon Prepared from Cashew Nut Shells

Authors: S. Tangjuank, N. Insuk , J. Tontrakoon , V. Udeye

Abstract:

Cashew nut shells were converted into activated carbon powders using KOH activation plus CO2 gasification at 1027 K. The increase both of impregnation ratio and activation time, there was swiftly the development of mesoporous structure with increasing of mesopore volume ratio from 20-28% and 27-45% for activated carbon with ratio of KOH per char equal to 1 and 4, respectively. Activated carbon derived from KOH/char ratio equal to 1 and CO2 gasification time from 20 to 150 minutes were exhibited the BET surface area increasing from 222 to 627 m2.g-1. And those were derived from KOH/char ratio of 4 with activation time from 20 to 150 minutes exhibited high BET surface area from 682 to 1026 m2.g-1. The adsorption of Lead(II) and Cadmium(II) ion was investigated. This adsorbent exhibited excellent adsorption for Lead(II) and Cadmium(II) ion. Maximum adsorption presented at 99.61% at pH 6.5 and 98.87% at optimum conditions. The experimental data was calculated from Freundlich isotherm and Langmuir isotherm model. The maximum capacity of Pb2+ and Cd2+ ions was found to be 28.90 m2.g-1 and 14.29 m2.g-1, respectively.

Keywords: Activated carbon, cashew nut shell, heavy metals, adsorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3349
5623 A New Approach to Face Recognition Using Dual Dimension Reduction

Authors: M. Almas Anjum, M. Younus Javed, A. Basit

Abstract:

In this paper a new approach to face recognition is presented that achieves double dimension reduction, making the system computationally efficient with better recognition results and out perform common DCT technique of face recognition. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results change with change in face image resolution and provide optimal results when arriving at a certain resolution level. In the proposed model of face recognition, initially image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to increased computational speed and feature extraction potential of Discrete Cosine Transform (DCT), it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A tradeoff between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL , Yale and EME color database.

Keywords: Biometrics, DCT, Face Recognition, Illumination, Computation, Feature extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1663
5622 Historical and Future Rainfall Variations in Bangladesh

Authors: M. M. Hossain, M. Z. Hasan, M. Alauddin, S. Akhter

Abstract:

Climate change has become a major concern across the world as the intensity along with quantity of the rainfall, mean surface temperature and other climatic parameters have been changed not only in Bangladesh but also in the entire globe. Bangladesh has already experienced many natural hazards. Among them changing of rainfall pattern, erratic and heavy rainfalls are very common. But changes of rainfall pattern and its amount is still in question to some extent. This study aimed to unfold how the historical rainfalls varied over time and how would be their future trends. In this context, historical rainfall data (1975-2014) were collected from Bangladesh Metrological Department (BMD) and then a time series model was developed using Box-Jenkins algorithm in IBM SPSS to forecast the future rainfall. From the historical data analysis, this study revealed that the amount of rainfall decreased over the time and shifted to the post monsoons. Forecasted rainfall shows that the pre-monsoon and early monsoon will get drier in future whereas late monsoon and post monsoon will show huge fluctuations in rainfall magnitudes with temporal variations which means Bangladesh will get comparatively drier seasons in future which may be a serious problem for the country as it depends on agriculture.

Keywords: Monsoon, Pre-monsoon, rainfall, pattern, variations, IBM-SPSS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1309
5621 A Dynamic Decision Model for Vertical Handoffs across Heterogeneous Wireless Networks

Authors: Pramod Goyal, S. K. Saxena

Abstract:

The convergence of heterogeneous wireless access technologies characterizes the 4G wireless networks. In such converged systems, the seamless and efficient handoff between different access technologies (vertical handoff) is essential and remains a challenging problem. The heterogeneous co-existence of access technologies with largely different characteristics creates a decision problem of determining the “best" available network at “best" time to reduce the unnecessary handoffs. This paper proposes a dynamic decision model to decide the “best" network at “best" time moment to handoffs. The proposed dynamic decision model make the right vertical handoff decisions by determining the “best" network at “best" time among available networks based on, dynamic factors such as “Received Signal Strength(RSS)" of network and “velocity" of mobile station simultaneously with static factors like Usage Expense, Link capacity(offered bandwidth) and power consumption. This model not only meets the individual user needs but also improve the whole system performance by reducing the unnecessary handoffs.

Keywords: Dynamic decision model, Seamless handoff, Vertical handoff, Wireless networks

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2030
5620 Generation Scheduling Optimization of Multi-Hydroplants: A Case Study

Authors: Shuangquan Liu, Jinwen Wang, Dada Wang

Abstract:

A case study of the generation scheduling optimization of the multi-hydroplants on the Yuan River Basin in China is reported in this paper. Concerning the uncertainty of the inflows, the long/mid-term generation scheduling (LMTGS) problem is solved by a stochastic model in which the inflows are considered as stochastic variables. For the short-term generation scheduling (STGS) problem, a constraint violation priority is defined in case not all constraints are satisfied. Provided the stage-wise separable condition and low dimensions, the hydroplant-based operational region schedules (HBORS) problem is solved by dynamic programming (DP). The coordination of LMTGS and STGS is presented as well. The feasibility and the effectiveness of the models and solution methods are verified by the numerical results.

Keywords: generation scheduling, multi-hydroplants, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1528
5619 Numerical Analysis of Fractured Process in Locomotive Steel Wheels

Authors: J. Alizadeh K., R. S. Ashofteh, A. Asadi Lari

Abstract:

Railway vehicle wheels are designed to operate in harsh environments and to withstand high hydrostatic contact pressures. This situation may result in critical circumstances, in particular wheel breakage. This paper presents a time history of a series of broken wheels during a time interval [2007-2008] belongs to locomotive fleet on Iranian Railways. Such fractures in locomotive wheels never reported before. Due to the importance of this issue, a research study has been launched to find the potential reasons of this problem. The authors introduce a FEM model to indicate how and where the wheels could have been affected during their operation. Then, the modeling results are presented and discussed in detail.

Keywords: Crack, fatigue, FE analysis, wheel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2341
5618 Computational Fluid Dynamics Simulation Approach for Developing a Powder Dispensing Device

Authors: Rallapalli Revanth, Shivakumar Bhavi, Vijay Kumar Turaga

Abstract:

Dispensing powders manually can be difficult as it requires to gradually pour and check the amount on the scale to be dispensed. Current systems are manual and non-continuous in nature and is user dependent and it is also difficult to control powder dispensation. Recurrent dosing of powdered medicines in precise amounts quickly and accurately has been an all-time challenge. Various powder dispensing mechanisms are being designed to overcome these challenges. Battery operated screw conveyor mechanism is being innovated to overcome above problems faced. These inventions are numerically evaluated at concept development level by employing Computational Fluid Dynamics (CFD) of gas-solids multiphase flow systems. CFD has been very helpful in the development of such devices, saving time and money by reducing the number of prototypes and testing. In this study, powder dispensation from the trocar's end is simulated by using the Dense Discrete Phase Model technique along with Kinetic Theory of Granular Flow. The powder is viewed as a secondary flow in air (DDPM-KTGF). By considering the volume fraction of powder as 50%, the transportation side is done by rotation of the screw conveyor. The performance is calculated for 1 sec time frame in an unsteady computation manner. This methodology will help designers in developing design concepts to improve the dispensation and the effective area within a quick turnaround time frame.

Keywords: Multiphase flow, screw conveyor, transient, DDPM - KTGF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 343
5617 An Investigation on Material Removal Rate of EDM Process: A Response Surface Methodology Approach

Authors: Azhar Equbal, Anoop Kumar Sood, M. Asif Equbal, M. Israr Equbal

Abstract:

In the present work response surface methodology (RSM) based central composite design (CCD) is used for analyzing the electrical discharge machining (EDM) process. For experimentation, mild steel is selected as work piece and copper is used as electrode. Three machining parameters namely current (I), spark on time (Ton) and spark off time (Toff) are selected as the input variables. The output or response chosen is material removal rate (MRR) which is to be maximized. To reduce the number of runs face centered central composite design (FCCCD) was used. ANOVA was used to determine the significance of parameter and interactions. The suitability of model is tested using Anderson darling (AD) plot. The results conclude that different parameters considered i.e. current, pulse on and pulse off time; all have dominant effect on the MRR. At last, the optimized parameter setting for maximizing MRR is found through main effect plot analysis.

Keywords: Electrical discharge machining, electrode, MRR, RSM, ANOVA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1151