Search results for: teaching-learning based optimization
8635 DTC-SVM Scheme for Induction Motors Fedwith a Three-level Inverter
Authors: Ehsan Hassankhan, Davood A. Khaburi
Abstract:
Direct Torque Control is a control technique in AC drive systems to obtain high performance torque control. The conventional DTC drive contains a pair of hysteresis comparators. DTC drives utilizing hysteresis comparators suffer from high torque ripple and variable switching frequency. The most common solution to those problems is to use the space vector depends on the reference torque and flux. In this Paper The space vector modulation technique (SVPWM) is applied to 2 level inverter control in the proposed DTC-based induction motor drive system, thereby dramatically reducing the torque ripple. Then the controller based on space vector modulation is designed to be applied in the control of Induction Motor (IM) with a three-level Inverter. This type of Inverter has several advantages over the standard two-level VSI, such as a greater number of levels in the output voltage waveforms, Lower dV/dt, less harmonic distortion in voltage and current waveforms and lower switching frequencies. This paper proposes a general SVPWM algorithm for three-level based on standard two-level SVPWM. The proposed scheme is described clearly and simulation results are reported to demonstrate its effectiveness. The entire control scheme is implemented with Matlab/Simulink.Keywords: Direct torque control, space vector Pulsewidthmodulation(SVPWM), neutral point clamped(NPC), two-levelinverter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43968634 Application New Approach with Two Networks Slow and Fast on the Asynchronous Machine
Authors: Samia Salah, M’hamed Hadj Sadok, Abderrezak Guessoum
Abstract:
In this paper, we propose a new modular approach called neuroglial consisting of two neural networks slow and fast which emulates a biological reality recently discovered. The implementation is based on complex multi-time scale systems; validation is performed on the model of the asynchronous machine. We applied the geometric approach based on the Gerschgorin circles for the decoupling of fast and slow variables, and the method of singular perturbations for the development of reductions models.
This new architecture allows for smaller networks with less complexity and better performance in terms of mean square error and convergence than the single network model.
Keywords: Gerschgorin’s Circles, Neuroglial Network, Multi time scales systems, Singular perturbation method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16058633 Soil Remediation Technologies towards Green Remediation Strategies
Authors: G. Petruzzelli, F. Pedron, M. Grifoni, M. Barbafieri, I. Rosellini, B. Pezzarossa
Abstract:
As a result of diverse industrial activities, pollution from numerous contaminant affects both groundwater and soils. Many contaminated sites have been discovered in industrialized countries and their remediation is a priority in environmental legislations. The aim of this paper is to provide the evolution of remediation from consolidated invasive technologies to environmental friendly green strategies. Many clean-up technologies have been used. Nowadays the technologies selection is no longer exclusively based on eliminating the source of pollution, but the aim of remediation includes also the recovery of soil quality. “Green remediation”, a strategy based on “soft technologies”, appears the key to tackle the issue of remediation of contaminated sites with the greatest attention to environmental quality, including the preservation of soil functionality.
Keywords: Bioremediation, green remediation, phytoremediation, remediation technologies, soil.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18588632 Topology-Based Character Recognition Method for Coin Date Detection
Authors: Xingyu Pan, Laure Tougne
Abstract:
For recognizing coins, the graved release date is important information to identify precisely its monetary type. However, reading characters in coins meets much more obstacles than traditional character recognition tasks in the other fields, such as reading scanned documents or license plates. To address this challenging issue in a numismatic context, we propose a training-free approach dedicated to detection and recognition of the release date of the coin. In the first step, the date zone is detected by comparing histogram features; in the second step, a topology-based algorithm is introduced to recognize coin numbers with various font types represented by binary gradient map. Our method obtained a recognition rate of 92% on synthetic data and of 44% on real noised data.
Keywords: Coin, detection, character recognition, topology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14778631 Fractal - Wavelet Based Techniques for Improving the Artificial Neural Network Models
Authors: Reza Bazargan Lari, Mohammad H. Fattahi
Abstract:
Natural resources management including water resources requires reliable estimations of time variant environmental parameters. Small improvements in the estimation of environmental parameters would result in grate effects on managing decisions. Noise reduction using wavelet techniques is an effective approach for preprocessing of practical data sets. Predictability enhancement of the river flow time series are assessed using fractal approaches before and after applying wavelet based preprocessing. Time series correlation and persistency, the minimum sufficient length for training the predicting model and the maximum valid length of predictions were also investigated through a fractal assessment.
Keywords: Wavelet, de-noising, predictability, time series fractal analysis, valid length, ANN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20638630 A System of Automatic Speech Recognition based on the Technique of Temporal Retiming
Authors: Samir Abdelhamid, Noureddine Bouguechal
Abstract:
We report in this paper the procedure of a system of automatic speech recognition based on techniques of the dynamic programming. The technique of temporal retiming is a technique used to synchronize between two forms to compare. We will see how this technique is adapted to the field of the automatic speech recognition. We will expose, in a first place, the theory of the function of retiming which is used to compare and to adjust an unknown form with a whole of forms of reference constituting the vocabulary of the application. Then we will give, in the second place, the various algorithms necessary to their implementation on machine. The algorithms which we will present were tested on part of the corpus of words in Arab language Arabdic-10 [4] and gave whole satisfaction. These algorithms are effective insofar as we apply them to the small ones or average vocabularies.Keywords: Continuous speech recognition, temporal retiming, phonetic decoding, algorithms, vocal signal, dynamic programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13488629 Web Usability : A Fuzzy Approach to the Navigation Structure Enhancement in a Website System, Case of Iranian Civil Aviation Organization Website
Authors: Hamed Qahri Saremi, Gholam Ali Montazer
Abstract:
With the proliferation of World Wide Web, development of web-based technologies and the growth in web content, the structure of a website becomes more complex and web navigation becomes a critical issue to both web designers and users. In this paper we define the content and web pages as two important and influential factors in website navigation and paraphrase the enhancement in the website navigation as making some useful changes in the link structure of the website based on the aforementioned factors. Then we suggest a new method for proposing the changes using fuzzy approach to optimize the website architecture. Applying the proposed method to a real case of Iranian Civil Aviation Organization (CAO) website, we discuss the results of the novel approach at the final section.Keywords: Web content, Web navigation, Website system, Webusage mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17888628 Information Technologies in Automotive Assembly Industry in Thailand
Authors: Jirarat Teeravaraprug, Usawadee Inklay
Abstract:
This paper gave an attempt in prioritizing information technologies that organizations should give concentration. The case study was organizations in the automotive assembly industry in Thailand. Data were first collected to gather all information technologies known and used in the automotive assembly industry in Thailand. Five experts from the industries were surveyed based on the concept of fuzzy DEMATEL. The information technologies were categorized into six groups, which were communication, transaction, planning, organization management, warehouse management, and transportation. The cause groups of information technologies for each group were analyzed and presented. Moreover, the relationship between the used and the significant information technologies was given. Discussions based on the used information technologies and the research results are given.
Keywords: Information technology, automotive assembly industry, fuzzy DEMATEL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17198627 A New Knapsack Public-Key Cryptosystem Based on Permutation Combination Algorithm
Authors: Min-Shiang Hwang, Cheng-Chi Lee, Shiang-Feng Tzeng
Abstract:
A new secure knapsack cryptosystem based on the Merkle-Hellman public key cryptosystem will be proposed in this paper. Although it is common sense that when the density is low, the knapsack cryptosystem turns vulnerable to the low-density attack. The density d of a secure knapsack cryptosystem must be larger than 0.9408 to avoid low-density attack. In this paper, we investigate a new Permutation Combination Algorithm. By exploiting this algorithm, we shall propose a novel knapsack public-key cryptosystem. Our proposed scheme can enjoy a high density to avoid the low-density attack. The density d can also exceed 0.9408 to avoid the low-density attack.Keywords: Public key, Knapsack problem, Knapsack cryptosystem, low-density attack.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19568626 Molecular Electronic Devices based on Carotenoid Derivatives
Authors: Vicente F. P. Aleixo, Augusto C. F. Saraiva, Jordan Del Nero
Abstract:
The production of devices in nanoscale with specific molecular rectifying function is one of the most significant goals in state-of-art technology. In this work we show by ab initio quantum mechanics calculations coupled with non-equilibrium Green function, the design of an organic two-terminal device. These molecular structures have molecular source and drain with several bridge length (from five up to 11 double bonds). Our results are consistent with significant features as a molecular rectifier and can be raised up as: (a) it can be used as bi-directional symmetrical rectifier; (b) two devices integrated in one (FET with one operational region, and Thyristor thiristor); (c) Inherent stability due small intrinsic capacitance under forward/reverse bias. We utilize a scheme for the transport mechanism based on previous properties of ¤Ç bonds type that can be successfully utilized to construct organic nanodevices.Keywords: ab initio, Carotenoid, Charge Transfer, Nanodevice
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16248625 Principal Component Analysis for the Characterization in the Application of Some Soil Properties
Authors: Kamolchanok Panishkan, Kanokporn Swangjang, Natdhera Sanmanee, Daoroong Sungthong
Abstract:
The objective of this research is to study principal component analysis for classification of 67 soil samples collected from different agricultural areas in the western part of Thailand. Six soil properties were measured on the soil samples and are used as original variables. Principal component analysis is applied to reduce the number of original variables. A model based on the first two principal components accounts for 72.24% of total variance. Score plots of first two principal components were used to map with agricultural areas divided into horticulture, field crops and wetland. The results showed some relationships between soil properties and agricultural areas. PCA was shown to be a useful tool for agricultural areas classification based on soil properties.Keywords: soil organic matter, soil properties, classification, principal components
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41148624 Real-Time Data Stream Partitioning over a Sliding Window in Real-Time Spatial Big Data
Authors: Sana Hamdi, Emna Bouazizi, Sami Faiz
Abstract:
In recent years, real-time spatial applications, like location-aware services and traffic monitoring, have become more and more important. Such applications result dynamic environments where data as well as queries are continuously moving. As a result, there is a tremendous amount of real-time spatial data generated every day. The growth of the data volume seems to outspeed the advance of our computing infrastructure. For instance, in real-time spatial Big Data, users expect to receive the results of each query within a short time period without holding in account the load of the system. But with a huge amount of real-time spatial data generated, the system performance degrades rapidly especially in overload situations. To solve this problem, we propose the use of data partitioning as an optimization technique. Traditional horizontal and vertical partitioning can increase the performance of the system and simplify data management. But they remain insufficient for real-time spatial Big data; they can’t deal with real-time and stream queries efficiently. Thus, in this paper, we propose a novel data partitioning approach for real-time spatial Big data named VPA-RTSBD (Vertical Partitioning Approach for Real-Time Spatial Big data). This contribution is an implementation of the Matching algorithm for traditional vertical partitioning. We find, firstly, the optimal attribute sequence by the use of Matching algorithm. Then, we propose a new cost model used for database partitioning, for keeping the data amount of each partition more balanced limit and for providing a parallel execution guarantees for the most frequent queries. VPA-RTSBD aims to obtain a real-time partitioning scheme and deals with stream data. It improves the performance of query execution by maximizing the degree of parallel execution. This affects QoS (Quality Of Service) improvement in real-time spatial Big Data especially with a huge volume of stream data. The performance of our contribution is evaluated via simulation experiments. The results show that the proposed algorithm is both efficient and scalable, and that it outperforms comparable algorithms.Keywords: Real-Time Spatial Big Data, Quality Of Service, Vertical partitioning, Horizontal partitioning, Matching algorithm, Hamming distance, Stream query.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10568623 Secure Image Retrieval Based On Orthogonal Decomposition under Cloud Environment
Authors: Yanyan Xu, Lizhi Xiong, Zhengquan Xu, Li Jiang
Abstract:
In order to protect data privacy, image with sensitive or private information needs to be encrypted before being outsourced to the cloud. However, this causes difficulties in image retrieval and data management. A secure image retrieval method based on orthogonal decomposition is proposed in the paper. The image is divided into two different components, for which encryption and feature extraction are executed separately. As a result, cloud server can extract features from an encrypted image directly and compare them with the features of the queried images, so that the user can thus obtain the image. Different from other methods, the proposed method has no special requirements to encryption algorithms. Experimental results prove that the proposed method can achieve better security and better retrieval precision.
Keywords: Secure image retrieval, secure search, orthogonal decomposition, secure cloud computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21148622 Taguchi-Based Six Sigma Approach to Optimize Surface Roughness for Milling Processes
Authors: Sky Chou, Joseph C. Chen
Abstract:
This paper focuses on using Six Sigma methodologies to improve the surface roughness of a manufactured part produced by the CNC milling machine. It presents a case study where the surface roughness of milled aluminum is required to reduce or eliminate defects and to improve the process capability index Cp and Cpk for a CNC milling process. The six sigma methodology, DMAIC (design, measure, analyze, improve, and control) approach, was applied in this study to improve the process, reduce defects, and ultimately reduce costs. The Taguchi-based six sigma approach was applied to identify the optimized processing parameters that led to the targeted surface roughness specified by our customer. A L9 orthogonal array was applied in the Taguchi experimental design, with four controllable factors and one non-controllable/noise factor. The four controllable factors identified consist of feed rate, depth of cut, spindle speed, and surface roughness. The noise factor is the difference between the old cutting tool and the new cutting tool. The confirmation run with the optimal parameters confirmed that the new parameter settings are correct. The new settings also improved the process capability index. The purpose of this study is that the Taguchi–based six sigma approach can be efficiently used to phase out defects and improve the process capability index of the CNC milling process.
Keywords: CNC machining, Six Sigma, Surface roughness, Taguchi methodology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10568621 Novel Schemes of Pilot-Aided Integer Frequency Offset Estimation for OFDM-Based DVB-T Systems
Authors: Youngyoon Lee, Dahae Chong, Myungsoo Lee, Seongho Chae, Seokho Yoon
Abstract:
This paper proposes two novel schemes for pilot-aided integer frequency offset (IFO) estimation in orthogonal frequency division multiplexing (OFDM)-based digital video broadcastingterrestrial (DVB-T) systems. The conventional scheme proposed for estimating the IFO uses only partial information of combinations that pilots can provide, which stems from a rigorous assumption that the channel responses of pilots used for estimating the IFO change very rapidly. Thus, in this paper, we propose the novel IFO estimation schemes exploiting all information of combinations that pilots can provide to improve the performance of IFO estimation. The simulation results show that the proposed schemes are highly accurate in terms of the IFO detection probability.Keywords: OFDM, DVB-T, pilot, IFO, estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17938620 Intelligent Recognition of Diabetes Disease via FCM Based Attribute Weighting
Authors: Kemal Polat
Abstract:
In this paper, an attribute weighting method called fuzzy C-means clustering based attribute weighting (FCMAW) for classification of Diabetes disease dataset has been used. The aims of this study are to reduce the variance within attributes of diabetes dataset and to improve the classification accuracy of classifier algorithm transforming from non-linear separable datasets to linearly separable datasets. Pima Indians Diabetes dataset has two classes including normal subjects (500 instances) and diabetes subjects (268 instances). Fuzzy C-means clustering is an improved version of K-means clustering method and is one of most used clustering methods in data mining and machine learning applications. In this study, as the first stage, fuzzy C-means clustering process has been used for finding the centers of attributes in Pima Indians diabetes dataset and then weighted the dataset according to the ratios of the means of attributes to centers of theirs. Secondly, after weighting process, the classifier algorithms including support vector machine (SVM) and k-NN (k- nearest neighbor) classifiers have been used for classifying weighted Pima Indians diabetes dataset. Experimental results show that the proposed attribute weighting method (FCMAW) has obtained very promising results in the classification of Pima Indians diabetes dataset.
Keywords: Fuzzy C-means clustering, Fuzzy C-means clustering based attribute weighting, Pima Indians diabetes dataset, SVM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17638619 A Novel GNSS Integrity Augmentation System for Civil and Military Aircraft
Authors: Roberto Sabatini, Terry Moore, Chris Hill
Abstract:
This paper presents a novel Global Navigation Satellite System (GNSS) Avionics Based Integrity Augmentation (ABIA) system architecture suitable for civil and military air platforms, including Unmanned Aircraft Systems (UAS). Taking the move from previous research on high-accuracy Differential GNSS (DGNSS) systems design, integration and experimental flight test activities conducted at the Italian Air Force Flight Test Centre (CSV-RSV), our research focused on the development of a novel approach to the problem of GNSS ABIA for mission- and safety-critical air vehicle applications and for multi-sensor avionics architectures based on GNSS. Detailed mathematical models were developed to describe the main causes of GNSS signal outages and degradation in flight, namely: antenna obscuration, multipath, fading due to adverse geometry and Doppler shift. Adopting these models in association with suitable integrity thresholds and guidance algorithms, the ABIA system is able to generate integrity cautions (predictive flags) and warnings (reactive flags), as well as providing steering information to the pilot and electronic commands to the aircraft/UAS flight control systems. These features allow real-time avoidance of safety-critical flight conditions and fast recovery of the required navigation performance in case of GNSS data losses. In other words, this novel ABIA system addresses all three cornerstones of GNSS integrity augmentation in mission- and safety-critical applications: prediction (caution flags), reaction (warning flags) and correction (alternate flight path computation).
Keywords: Global Navigation Satellite Systems (GNSS), Integrity Augmentation, Unmanned Aircraft Systems, Aircraft Based Augmentation, Avionics Based Integrity Augmentation, Safety-Critical Applications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32448618 An Elin Load Tap Changer Diagnosis by DGA
Authors: Hoda Molavi, Alireza Zahiri, Katayoon Anvarizadeh
Abstract:
Dissolved gas analysis has been accepted as a sensitive, informative and reliable technique for incipient faults detection in power transformers and is widely used. In the last few years this method, which has been recommended by IEEE Power & Energy society, has been applied for fault detection in load tap changers. Regarding the critical role of load tap changers in electrical network and essential of catastrophic failures prevention, it is necessary to choose "condition based preventative maintenance strategy" which leads to reduction in costs, the number of unnecessary visits as well as the probability of interruptions and also increment in equipment reliability. In current work, considering the condition based preventative maintenance strategy, condition assessment of an Elin tap changer was carried out using dissolved gas analysis.
Keywords: Condition Assessment, Dissolved Gas Analysis, Load Tap Changer
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37168617 Genetic Characterization of Barley Genotypes via Inter-Simple Sequence Repeat
Authors: Mustafa Yorgancılar, Emine Atalay, Necdet Akgün, Ali Topal
Abstract:
In this study, polymerase chain reaction based Inter-simple sequence repeat (ISSR) from DNA fingerprinting techniques were used to investigate the genetic relationships among barley crossbreed genotypes in Turkey. It is important that selection based on the genetic base in breeding programs via ISSR, in terms of breeding time. 14 ISSR primers generated a total of 97 bands, of which 81 (83.35%) were polymorphic. The highest total resolution power (RP) value was obtained from the F2 (0.53) and M16 (0.51) primers. According to the ISSR result, the genetic similarity index changed between 0.64–095; Lane 3 with Line 6 genotypes were the closest, while Line 36 were the most distant ones. The ISSR markers were found to be promising for assessing genetic diversity in barley crossbreed genotypes.
Keywords: Barley, crossbreed, genetic similarity, ISSR.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9198616 Using Emotional Learning in Rescue Simulation Environment
Authors: Maziar Ahmad Sharbafi, Caro Lucas, Abolfazel Toroghi Haghighat, Omid AmirGhiasvand, Omid Aghazade
Abstract:
RoboCup Rescue simulation as a large-scale Multi agent system (MAS) is one of the challenging environments for keeping coordination between agents to achieve the objectives despite sensing and communication limitations. The dynamicity of the environment and intensive dependency between actions of different kinds of agents make the problem more complex. This point encouraged us to use learning-based methods to adapt our decision making to different situations. Our approach is utilizing reinforcement leaning. Using learning in rescue simulation is one of the current ways which has been the subject of several researches in recent years. In this paper we present an innovative learning method implemented for Police Force (PF) Agent. This method can cope with the main difficulties that exist in other learning approaches. Different methods used in the literature have been examined. Their drawbacks and possible improvements have led us to the method proposed in this paper which is fast and accurate. The Brain Emotional Learning Based Intelligent Controller (BELBIC) is our solution for learning in this environment. BELBIC is a physiologically motivated approach based on a computational model of amygdale and limbic system. The paper presents the results obtained by the proposed approach, showing the power of BELBIC as a decision making tool in complex and dynamic situation.Keywords: Emotional learning, rescue, simulation environment, RoboCup, multi-agent system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16218615 Bangla Vowel Characterization Based on Analysis by Synthesis
Authors: Syed Akhter Hossain, M. Lutfar Rahman, Farruk Ahmed
Abstract:
Bangla Vowel characterization determines the spectral properties of Bangla vowels for efficient synthesis as well as recognition of Bangla vowels. In this paper, Bangla vowels in isolated word have been analyzed based on speech production model within the framework of Analysis-by-Synthesis. This has led to the extraction of spectral parameters for the production model in order to produce different Bangla vowel sounds. The real and synthetic spectra are compared and a weighted square error has been computed along with the error in the formant bandwidths for efficient representation of Bangla vowels. The extracted features produced good representation of targeted Bangla vowel. Such a representation also plays essential role in low bit rate speech coding and vocoders.
Keywords: Speech, vowel, formant, synthesis, spectrum, LPC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23718614 Entanglement-based Quantum Computing by Diagrams of States
Authors: Sara Felloni, Giuliano Strini
Abstract:
We explore entanglement in composite quantum systems and how its peculiar properties are exploited in quantum information and communication protocols by means of Diagrams of States, a novel method to graphically represent and analyze how quantum information is elaborated during computations performed by quantum circuits. We present quantum diagrams of states for Bell states generation, measurements and projections, for dense coding and quantum teleportation, for probabilistic quantum machines designed to perform approximate quantum cloning and universal NOT and, finally, for quantum privacy amplification based on entanglement purification. Diagrams of states prove to be a useful approach to analyze quantum computations, by offering an intuitive graphic representation of the processing of quantum information. They also help in conceiving novel quantum computations, from describing the desired information processing to deriving the final implementation by quantum gate arrays.Keywords: Diagrams of states, entanglement, quantum circuits, quantum information.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16588613 Curvelet Transform Based Two Class Motor Imagery Classification
Authors: Nebi Gedik
Abstract:
One of the important parts of the brain-computer interface (BCI) studies is the classification of motor imagery (MI) obtained by electroencephalography (EEG). The major goal is to provide non-muscular communication and control via assistive technologies to people with severe motor disorders so that they can communicate with the outside world. In this study, an EEG signal classification approach based on multiscale and multi-resolution transform method is presented. The proposed approach is used to decompose the EEG signal containing motor image information (right- and left-hand movement imagery). The decomposition process is performed using curvelet transform which is a multiscale and multiresolution analysis method, and the transform output was evaluated as feature data. The obtained feature set is subjected to feature selection process to obtain the most effective ones using t-test methods. SVM and k-NN algorithms are assigned for classification.
Keywords: motor imagery, EEG, curvelet transform, SVM, k-NN
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6228612 Load Transfer Mechanism Based Unified Strut-and-Tie Modeling for Design of Concrete Beams
Authors: Ahmed, M., Yasser A., Mahmoud H., Ahmed, A., Abdulla M. S., Nazar, S.
Abstract:
Strut-and-Tie Models (STM) for the design of concrete beams, comprising of struts, ties, nodes as the basic tools, is conceptually simple, but its realization for complex concrete structure is not straightforward and depends on flow of internal forces in the structure. STM technique has won wide acceptance for deep member and shear design. STM technique is a unified approach that considers all load effects (bending, axial, shear, and torsion) simultaneously, not just applicable to shear loading only. The present study is to portray Strut-and-Tie Modeling based on Load-Transfer-Mechanisms as a unified method to analyze, design and detailing for deep and slender concrete beams. Three shear span- effective depth ratio (a/ d) are recommended for the modeling of STM elements corresponding to dominant load paths. The study also discusses the research work conduct on effective stress of concrete, tie end anchorage, and transverse reinforcement demand under different load transfer mechanism. It is also highlighted that to make the STM versatile tool for design of beams applicable to all shear spans, the effective stress of concrete and, transverse reinforcement demand, inclined angle of strut, and anchorage requirements of tie bars is required to be correlated with respect to load transfer mechanism. The country code provisions are to be modified and updated to apply for generalized design of concrete deep and slender member using load transfer mechanism based STM technique. Examples available in literature are reanalyzed with refined STM based on load transfer mechanisms and results are compared. It is concluded from the results that proposed approach will require true reinforcement demand depending on dominant force transfer action in concrete beam.
Keywords: Deep member, Load transfer mechanism, Strut-and-Tie Model, Strut, Truss.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 59868611 Video Data Mining based on Information Fusion for Tamper Detection
Authors: Girija Chetty, Renuka Biswas
Abstract:
In this paper, we propose novel algorithmic models based on information fusion and feature transformation in crossmodal subspace for different types of residue features extracted from several intra-frame and inter-frame pixel sub-blocks in video sequences for detecting digital video tampering or forgery. An evaluation of proposed residue features – the noise residue features and the quantization features, their transformation in cross-modal subspace, and their multimodal fusion, for emulated copy-move tamper scenario shows a significant improvement in tamper detection accuracy as compared to single mode features without transformation in cross-modal subspace.Keywords: image tamper detection, digital forensics, correlation features image fusion
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19008610 Demographics Are Not Enough: Targeting and Segmentation of Anti-Obesity Campaigns in Mexico
Authors: D. Wrzecionkowska
Abstract:
Mass media campaigns against obesity are often designed to impact large audiences. This usually means that their audience is defined based on general demographic characteristics like age, gender, occupation etc., not taking into account psychographics like behavior, motivations, wants, etc. Using psychographics, as the base for the audience segmentation, is a common practice in case of successful campaigns, as it allows developing more relevant messages. It also serves a purpose of identifying key segments, those that generate the best return on investment. For a health campaign, that would be segments that have the best chance of being converted into healthy lifestyle at the lowest cost. This paper presents the limitations of the demographic targeting, based on the findings from the reception study of IMSS (Mexican Social Security Institute) antiobesity TV commercials and proposes mothers as the first level of segmentation, in the process of identifying the key segment for these campaigns.Keywords: Anti-obesity campaigns, mothers, segmentation, targeting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20718609 Intention to Use Digital Library based on Modified UTAUT Model: Perspectives of Malaysian Postgraduate Students
Authors: Abd Latif Abdul Rahman, Adnan Jamaludin, Zamalia Mahmud
Abstract:
Unified Theory of Acceptance and Use of Technology (UTAUT) model has demonstrated the influencing factors for generic information systems use such as tablet personal computer (TPC) and mobile communication. However, in the context of digital library system, there has been very little effort to determine factors affecting the intention to use digital library based on the UTAUT model. This paper investigates factors that are expected to influence the intention of postgraduate students to use digital library based on modified UTAUT model. The modified model comprises of constructs represented by several latent variables, namely performance expectancy (PE), effort expectancy (EE), information quality (IQ) and service quality (SQ) and moderated by age, gender and experience in using digital library. Results show that performance expectancy, effort expectancy and information quality are positively related to the intention to use digital library, while service quality is negatively related to the intention to use digital library. Age and gender have shown no evidence of any significant interactions, while experience in using digital library significantly interacts with effort expectancy and intention to use digital library. This has provided the evidence of a moderating effect of experience in the intention to use digital library. It is expected that this research will shed new lights into research of acceptance and intention to use the library in a digital environment.Keywords: Intention to use digital library, UTAUT model, performance expectancy, effort expectancy, information quality, service quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 48708608 Lean Healthcare: Barriers and Enablers in the Colombian Context
Authors: Erika Ruiz, Nestor Ortiz
Abstract:
Lean philosophy has evolved over time and has been implemented both in manufacturing and services, more recently lean has been integrated in the companies of the health sector. Currently it is important to understand the successful way to implement this philosophy and try to identify barriers and enablers to the sustainability of lean healthcare. The main purpose of this research is to identify the barriers and enablers in the implementation of Lean Healthcare based on case studies of Colombian healthcare centers. In order to do so, we conducted semi-structured interviews based on a maturity model. The main results indicate that the success of Lean implementation depends on its adaptation to contextual factors. In addition, in the Colombian context were identified new factors such as organizational culture, management models, integration of the care and administrative departments and triple helix relationship.Keywords: Barriers, enablers, implementation, lean healthcare, sustainability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16478607 The Frequency Graph for the Traveling Salesman Problem
Authors: Y. Wang
Abstract:
Traveling salesman problem (TSP) is hard to resolve when the number of cities and routes become large. The frequency graph is constructed to tackle the problem. A frequency graph maintains the topological relationships of the original weighted graph. The numbers on the edges are the frequencies of the edges emulated from the local optimal Hamiltonian paths. The simplest kind of local optimal Hamiltonian paths are computed based on the four vertices and three lines inequality. The search algorithm is given to find the optimal Hamiltonian circuit based on the frequency graph. The experiments show that the method can find the optimal Hamiltonian circuit within several trials.Keywords: Traveling salesman problem, frequency graph, local optimal Hamiltonian path, four vertices and three lines inequality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17718606 A Psychophysiological Evaluation of an Effective Recognition Technique Using Interactive Dynamic Virtual Environments
Authors: Mohammadhossein Moghimi, Robert Stone, Pia Rotshtein
Abstract:
Recording psychological and physiological correlates of human performance within virtual environments and interpreting their impacts on human engagement, ‘immersion’ and related emotional or ‘effective’ states is both academically and technologically challenging. By exposing participants to an effective, real-time (game-like) virtual environment, designed and evaluated in an earlier study, a psychophysiological database containing the EEG, GSR and Heart Rate of 30 male and female gamers, exposed to 10 games, was constructed. Some 174 features were subsequently identified and extracted from a number of windows, with 28 different timing lengths (e.g. 2, 3, 5, etc. seconds). After reducing the number of features to 30, using a feature selection technique, K-Nearest Neighbour (KNN) and Support Vector Machine (SVM) methods were subsequently employed for the classification process. The classifiers categorised the psychophysiological database into four effective clusters (defined based on a 3-dimensional space – valence, arousal and dominance) and eight emotion labels (relaxed, content, happy, excited, angry, afraid, sad, and bored). The KNN and SVM classifiers achieved average cross-validation accuracies of 97.01% (±1.3%) and 92.84% (±3.67%), respectively. However, no significant differences were found in the classification process based on effective clusters or emotion labels.
Keywords: Virtual Reality, effective computing, effective VR, emotion-based effective physiological database.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 994