Search results for: data mining technique
8551 Person-Environment Fit (PE Fit): Evidence from Brazil
Authors: Jucelia Appio, Danielle Deimling De Carli, Bruno Henrique Rocha Fernandes, Nelson Natalino Frizon
Abstract:
The purpose of this paper is to investigate if there are positive and significant correlations between the dimensions of Person-Environment Fit (Person-Job, Person-Organization, Person-Group and Person-Supervisor) at the “Best Companies to Work for” in Brazil in 2017. For that, a quantitative approach was used with a descriptive method being defined as a research sample the "150 Best Companies to Work for", according to data base collected in 2017 and provided by Fundação Instituto of Administração (FIA) of the University of São Paulo (USP). About the data analysis procedures, asymmetry and kurtosis, factorial analysis, Kaiser-Meyer-Olkin (KMO) tests, Bartlett sphericity and Cronbach's alpha were used for the 69 research variables, and as a statistical technique for the purpose of analyzing the hypothesis, Pearson's correlation analysis was performed. As a main result, we highlight that there was a positive and significant correlation between the dimensions of Person-Environment Fit, corroborating the H1 hypothesis that there is a positive and significant correlation between Person-Job Fit, Person-Organization Fit, Person-Group Fit and Person-Supervisor Fit.
Keywords: Human resource management, person-environment fit, strategic people management, best companies to work for.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9988550 Weigh-in-Motion Data Analysis Software for Developing Traffic Data for Mechanistic Empirical Pavement Design
Authors: M. A. Hasan, M. R. Islam, R. A. Tarefder
Abstract:
Currently, there are few user friendly Weigh-in- Motion (WIM) data analysis softwares available which can produce traffic input data for the recently developed AASHTOWare pavement Mechanistic-Empirical (ME) design software. However, these softwares have only rudimentary Quality Control (QC) processes. Therefore, they cannot properly deal with erroneous WIM data. As the pavement performance is highly sensible to the quality of WIM data, it is highly recommended to use more refined QC process on raw WIM data to get a good result. This study develops a userfriendly software, which can produce traffic input for the ME design software. This software takes the raw data (Class and Weight data) collected from the WIM station and processes it with a sophisticated QC procedure. Traffic data such as traffic volume, traffic distribution, axle load spectra, etc. can be obtained from this software; which can directly be used in the ME design software.Keywords: Weigh-in-motion, software, axle load spectra, traffic distribution, AASHTOWare.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18968549 A Hybrid Approach for Color Image Quantization Using K-means and Firefly Algorithms
Authors: Parisut Jitpakdee, Pakinee Aimmanee, Bunyarit Uyyanonvara
Abstract:
Color Image quantization (CQ) is an important problem in computer graphics, image and processing. The aim of quantization is to reduce colors in an image with minimum distortion. Clustering is a widely used technique for color quantization; all colors in an image are grouped to small clusters. In this paper, we proposed a new hybrid approach for color quantization using firefly algorithm (FA) and K-means algorithm. Firefly algorithm is a swarmbased algorithm that can be used for solving optimization problems. The proposed method can overcome the drawbacks of both algorithms such as the local optima converge problem in K-means and the early converge of firefly algorithm. Experiments on three commonly used images and the comparison results shows that the proposed algorithm surpasses both the base-line technique k-means clustering and original firefly algorithm.Keywords: Clustering, Color quantization, Firefly algorithm, Kmeans.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22188548 Audio Watermarking Based on Compression-expansion Technique
Authors: Say Wei Foo, Qi Dong
Abstract:
A novel robust audio watermarking scheme is proposed in this paper. In the proposed scheme, the host audio signals are segmented into frames. Two consecutive frames are assessed if they are suitable to represent a watermark bit. If so, frequency transform is performed on these two frames. The compressionexpansion technique is adopted to generate distortion over the two frames. The distortion is used to represent one watermark bit. Psychoacoustic model is applied to calculate local auditory mask to ensure that the distortion is not audible. The watermarking schemes using mono and stereo audio signals are designed differently. The correlation-based detection method is used to detect the distortion and extract embedded watermark bits. The experimental results show that the quality degradation caused by the embedded watermarks is perceptually transparent and the proposed schemes are very robust against different types of attacks.Keywords: Audio watermarking, Compression-expansion, Stereo signals, Robustness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16458547 The Significance of the Radiography Technique in the Non-Destructive Evaluation of the Integrity and Reliability of Cast Interconnects
Authors: Keshav Pujeri, Pranesh Jain, Krutibas Panda
Abstract:
Significant changes in oil and gas drilling have emphasized the need to verify the integrity and reliability of drill stem components. Defects are inevitable in cast components, regardless of application; but if these defects go undetected, any severe defect could cause down-hole failure. One such defect is shrinkage porosity. Castings with lower level shrinkage porosity (CB levels 1 and 2) have scattered pores and do not occupy large volumes; so pressure testing and helium leak testing (HLT) are sufficient for qualifying the castings. However, castings with shrinkage porosity of CB level 3 and higher, behave erratically under pressure testing and HLT making these techniques insufficient for evaluating the castings- integrity. This paper presents a case study to highlight how the radiography technique is much more effective than pressure testing and HLT.Keywords: Casting Defects, Interconnects, Leak Check, Pressure Test, Radiography.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35148546 Electric Load Forecasting Using Genetic Based Algorithm, Optimal Filter Estimator and Least Error Squares Technique: Comparative Study
Authors: Khaled M. EL-Naggar, Khaled A. AL-Rumaih
Abstract:
This paper presents performance comparison of three estimation techniques used for peak load forecasting in power systems. The three optimum estimation techniques are, genetic algorithms (GA), least error squares (LS) and, least absolute value filtering (LAVF). The problem is formulated as an estimation problem. Different forecasting models are considered. Actual recorded data is used to perform the study. The performance of the above three optimal estimation techniques is examined. Advantages of each algorithms are reported and discussed.
Keywords: Forecasting, Least error squares, Least absolute Value, Genetic algorithms
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27238545 Human Growth Curve Estimation through a Combination of Longitudinal and Cross-sectional Data
Authors: Sedigheh Mirzaei S., Debasis Sengupta
Abstract:
Parametric models have been quite popular for studying human growth, particularly in relation to biological parameters such as peak size velocity and age at peak size velocity. Longitudinal data are generally considered to be vital for fittinga parametric model to individual-specific data, and for studying the distribution of these biological parameters in a human population. However, cross-sectional data are easier to obtain than longitudinal data. In this paper, we present a method of combining longitudinal and cross-sectional data for the purpose of estimating the distribution of the biological parameters. We demonstrate, through simulations in the special case ofthePreece Baines model, how estimates based on longitudinal data can be improved upon by harnessing the information contained in cross-sectional data.We study the extent of improvement for different mixes of the two types of data, and finally illustrate the use of the method through data collected by the Indian Statistical Institute.Keywords: Preece-Baines growth model, MCMC method, Mixed effect model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21398544 Modern Vibration Signal Processing Techniques for Vehicle Gearbox Fault Diagnosis
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
This paper presents modern vibration signalprocessing techniques for vehicle gearbox fault diagnosis, via the wavelet analysis and the Squared Envelope (SE) technique. The wavelet analysis is regarded as a powerful tool for the detection of sudden changes in non-stationary signals. The Squared Envelope (SE) technique has been extensively used for rolling bearing diagnostics. In the present work a scheme of using the Squared Envelope technique for early detection of gear tooth pit. The pitting defect is manufactured on the tooth side of a fifth speed gear on the intermediate shaft of a vehicle gearbox. The objective is to supplement the current techniques of gearbox fault diagnosis based on using the raw vibration and ordered signals. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of output joint shafts. The gearbox used for experimental measurements is the type most commonly used in modern small to mid-sized passenger cars with transversely mounted powertrain and front wheel drive; a five-speed gearbox with final drive gear and front wheel differential. The results show that the approaches methods are effective for detecting and diagnosing localized gear faults in early stage under different operation conditions, and are more sensitive and robust than current gear diagnostic techniques.
Keywords: Wavelet analysis, Squared Envelope, gear faults.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25768543 Loop Back Connected Component Labeling Algorithm and Its Implementation in Detecting Face
Authors: A. Rakhmadi, M. S. M. Rahim, A. Bade, H. Haron, I. M. Amin
Abstract:
In this study, a Loop Back Algorithm for component connected labeling for detecting objects in a digital image is presented. The approach is using loop back connected component labeling algorithm that helps the system to distinguish the object detected according to their label. Deferent than whole window scanning technique, this technique reduces the searching time for locating the object by focusing on the suspected object based on certain features defined. In this study, the approach was also implemented for a face detection system. Face detection system is becoming interesting research since there are many devices or systems that require detecting the face for certain purposes. The input can be from still image or videos, therefore the sub process of this system has to be simple, efficient and accurate to give a good result.Keywords: Image processing, connected components labeling, face detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22998542 Computer Generated Hologram for SemiFragile Watermarking with Encrypted Images
Authors: G. Schirripa Spagnolo, M. De Santis
Abstract:
The protection of the contents of digital products is referred to as content authentication. In some applications, to be able to authenticate a digital product could be extremely essential. For example, if a digital product is used as a piece of evidence in the court, its integrity could mean life or death of the accused. Generally, the problem of content authentication can be solved using semifragile digital watermarking techniques. Recently many authors have proposed Computer Generated Hologram Watermarking (CGHWatermarking) techniques. Starting from these studies, in this paper a semi-fragile Computer Generated Hologram coding technique is proposed, which is able to detect malicious tampering while tolerating some incidental distortions. The proposed technique uses as watermark an encrypted image, and it is well suitable for digital image authentication.Keywords: Asymmetric cryptography, Semi-Fragile watermarking, Image authentication, Hologram watermark, Public- Key Cryptography, RSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16098541 Semantic Support for Hypothesis-Based Research from Smart Environment Monitoring and Analysis Technologies
Authors: T. S. Myers, J. Trevathan
Abstract:
Improvements in the data fusion and data analysis phase of research are imperative due to the exponential growth of sensed data. Currently, there are developments in the Semantic Sensor Web community to explore efficient methods for reuse, correlation and integration of web-based data sets and live data streams. This paper describes the integration of remotely sensed data with web-available static data for use in observational hypothesis testing and the analysis phase of research. The Semantic Reef system combines semantic technologies (e.g., well-defined ontologies and logic systems) with scientific workflows to enable hypothesis-based research. A framework is presented for how the data fusion concepts from the Semantic Reef architecture map to the Smart Environment Monitoring and Analysis Technologies (SEMAT) intelligent sensor network initiative. The data collected via SEMAT and the inferred knowledge from the Semantic Reef system are ingested to the Tropical Data Hub for data discovery, reuse, curation and publication.
Keywords: Information architecture, Semantic technologies Sensor networks, Ontologies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17158540 Data Migration between Document-Oriented and Relational Databases
Authors: Bogdan Walek, Cyril Klimes
Abstract:
Current tools for data migration between documentoriented and relational databases have several disadvantages. We propose a new approach for data migration between documentoriented and relational databases. During data migration the relational schema of the target (relational database) is automatically created from collection of XML documents. Proposed approach is verified on data migration between document-oriented database IBM Lotus/ Notes Domino and relational database implemented in relational database management system (RDBMS) MySQL.Keywords: data migration, database, document-oriented database, XML, relational schema
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35258539 The CEO Mission II, Rescue Robot with Multi-Joint Mechanical Arm
Authors: Amon Tunwannarux, Supanunt Tunwannarux
Abstract:
This paper presents design features of a rescue robot, named CEO Mission II. Its body is designed to be the track wheel type with double front flippers for climbing over the collapse and the rough terrain. With 125 cm. long, 5-joint mechanical arm installed on the robot body, it is deployed not only for surveillance from the top view but also easier and faster access to the victims to get their vital signs. Two cameras and sensors for searching vital signs are set up at the tip of the multi-joint mechanical arm. The third camera is at the back of the robot for driving control. Hardware and software of the system, which controls and monitors the rescue robot, are explained. The control system is used for controlling the robot locomotion, the 5-joint mechanical arm, and for turning on/off devices. The monitoring system gathers all information from 7 distance sensors, IR temperature sensors, 3 CCD cameras, voice sensor, robot wheels encoders, yawn/pitch/roll angle sensors, laser range finder and 8 spare A/D inputs. All sensors and controlling data are communicated with a remote control station via IEEE 802.11b Wi-Fi. The audio and video data are compressed and sent via another IEEE 802.11g Wi-Fi transmitter for getting real-time response. At remote control station site, the robot locomotion and the mechanical arm are controlled by joystick. Moreover, the user-friendly GUI control program is developed based on the clicking and dragging method to easily control the movement of the arm. Robot traveling map is plotted from computing the information of wheel encoders and the yawn/pitch data. 2D Obstacle map is plotted from data of the laser range finder. The concept and design of this robot can be adapted to suit many other applications. As the Best Technique awardee from Thailand Rescue Robot Championship 2006, all testing results are satisfied.Keywords: Controlling, monitoring, rescue robot, mechanicalarm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19728538 An Appraisal of Coal Fly Ash Soil Amendment Technology (FASAT) of Central Institute of Mining and Fuel Research (CIMFR)
Authors: L.C. Ram, R.E. Masto, Smriti Singh, R.C. Tripathi, S.K. Jha, N.K. Srivastava, A.K. Sinha, V.A. Selvi, A. Sinha
Abstract:
Coal will continue to be the predominant source of global energy for coming several decades. The huge generation of fly ash (FA) from combustion of coal in thermal power plants (TPPs) is apprehended to pose the concerns of its disposal and utilization. FA application based on its typical characteristics as soil ameliorant for agriculture and forestry is the potential area, and hence the global attempt. The inferences drawn suffer from the variations of ash characteristics, soil types, and agro-climatic conditions; thereby correlating the effects of ash between various plant species and soil types is difficult. Indian FAs have low bulk density, high water holding capacity and porosity, rich silt-sized particles, alkaline nature, negligible solubility, and reasonable plant nutrients. Findings of the demonstrations trials for more than two decades from lab/pot to field scale long-term experiments are developed as FA soil amendment technology (FASAT) by Central Institute of Mining and Fuel Research (CIMFR), Dhanbad. Performance of different crops and plant species in cultivable and problematic soils, are encouraging, eco-friendly, and being adopted by the farmers. FA application includes ash alone and in combination with inorganic/organic amendments; combination treatments including bio-solids perform better than FA alone. Optimum dose being up to 100 t/ha for cultivable land and up to/ or above 200 t/ha of FA for waste/degraded land/mine refuse, depending on the characteristics of ash and soil. The elemental toxicity in Indian FA is usually not of much concern owing to alkaline ashes, oxide forms of elements, and elemental concentration within the threshold limits for soil application. Combating toxicity, if any, is possible through combination treatments with organic materials and phytoremediation. Government initiatives through extension programme involving farmers and ash generating organizations need to be acceleratedKeywords: Fly ash, soil quality, CIMFR, FASAT, agriculture, forestry, toxicity, remediation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30648537 Time-Delay Estimation Using Cross-ΨB-Energy Operator
Authors: Z. Saidi, A.O. Boudraa, J.C. Cexus, S. Bourennane
Abstract:
In this paper, a new time-delay estimation technique based on the cross IB-energy operator [5] is introduced. This quadratic energy detector measures how much a signal is present in another one. The location of the peak of the energy operator, corresponding to the maximum of interaction between the two signals, is the estimate of the delay. The method is a fully data-driven approach. The discrete version of the continuous-time form of the cross IBenergy operator, for its implementation, is presented. The effectiveness of the proposed method is demonstrated on real underwater acoustic signals arriving from targets and the results compared to the cross-correlation method.Keywords: Teager-Kaiser energy operator, Cross-energyoperator, Time-Delay, Underwater acoustic signals.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 56478536 An Automated Stock Investment System Using Machine Learning Techniques: An Application in Australia
Authors: Carol Anne Hargreaves
Abstract:
A key issue in stock investment is how to select representative features for stock selection. The objective of this paper is to firstly determine whether an automated stock investment system, using machine learning techniques, may be used to identify a portfolio of growth stocks that are highly likely to provide returns better than the stock market index. The second objective is to identify the technical features that best characterize whether a stock’s price is likely to go up and to identify the most important factors and their contribution to predicting the likelihood of the stock price going up. Unsupervised machine learning techniques, such as cluster analysis, were applied to the stock data to identify a cluster of stocks that was likely to go up in price – portfolio 1. Next, the principal component analysis technique was used to select stocks that were rated high on component one and component two – portfolio 2. Thirdly, a supervised machine learning technique, the logistic regression method, was used to select stocks with a high probability of their price going up – portfolio 3. The predictive models were validated with metrics such as, sensitivity (recall), specificity and overall accuracy for all models. All accuracy measures were above 70%. All portfolios outperformed the market by more than eight times. The top three stocks were selected for each of the three stock portfolios and traded in the market for one month. After one month the return for each stock portfolio was computed and compared with the stock market index returns. The returns for all three stock portfolios was 23.87% for the principal component analysis stock portfolio, 11.65% for the logistic regression portfolio and 8.88% for the K-means cluster portfolio while the stock market performance was 0.38%. This study confirms that an automated stock investment system using machine learning techniques can identify top performing stock portfolios that outperform the stock market.
Keywords: Machine learning, stock market trading, logistic principal component analysis, automated stock investment system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10988535 Characterization of Fabricated A 384.1-MgO Based Metal Matrix Composite and Optimization of Tensile Strength using Taguchi Techniques
Authors: Nripjit, Anand K Tyagi, Nirmal Singh
Abstract:
The present work consecutively on synthesis and characterization of composites, Al/Al alloy A 384.1 as matrix in which the main ingredient as Al/Al-5% MgO alloy based metal matrix composite. As practical implications the low cost processing route for the fabrication of Al alloy A 384.1 and operational difficulties of presently available manufacturing processes based in liquid manipulation methods. As all new developments, complete understanding of the influence of processing variables upon the final quality of the product. And the composite is applied comprehensively to the acquaintance for achieving superiority of information concerning the specific heat measurement of a material through the aid of thermographs. Products are evaluated concerning relative particle size and mechanical behavior under tensile strength. Furthermore, Taguchi technique was employed to examine the experimental optimum results are achieved, owing to effectiveness of this approach.Keywords: MMC, Thermographs, Tensile strength, Taguchi technique, Optimal parameters
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16388534 Power Saving System in Green Data Center
Authors: Joon-young Jung, Dong-oh Kang, Chang-seok Bae
Abstract:
Power consumption is rapidly increased in data centers because the number of data center is increased and more the scale of data center become larger. Therefore, it is one of key research items to reduce power consumption in data center. The peak power of a typical server is around 250 watts. When a server is idle, it continues to use around 60% of the power consumed when in use, though vendors are putting effort into reducing this “idle" power load. Servers tend to work at only around a 5% to 20% utilization rate, partly because of response time concerns. An average of 10% of servers in their data centers was unused. In those reason, we propose dynamic power management system to reduce power consumption in green data center. Experiment result shows that about 55% power consumption is reduced at idle time.Keywords: Data Center, Green IT, Management Server, Power Saving.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16288533 Deployment of Beyond 4G Wireless Communication Networks with Carrier Aggregation
Authors: Bahram Khan, Anderson Rocha Ramos, Rui R. Paulo, Fernando J. Velez
Abstract:
With the growing demand for a new blend of applications, the users dependency on the internet is increasing day by day. Mobile internet users are giving more attention to their own experiences, especially in terms of communication reliability, high data rates and service stability on move. This increase in the demand is causing saturation of existing radio frequency bands. To address these challenges, researchers are investigating the best approaches, Carrier Aggregation (CA) is one of the newest innovations, which seems to fulfill the demands of the future spectrum, also CA is one the most important feature for Long Term Evolution - Advanced (LTE-Advanced). For this purpose to get the upcoming International Mobile Telecommunication Advanced (IMT-Advanced) mobile requirements (1 Gb/s peak data rate), the CA scheme is presented by 3GPP, which would sustain a high data rate using widespread frequency bandwidth up to 100 MHz. Technical issues such as aggregation structure, its implementations, deployment scenarios, control signal techniques, and challenges for CA technique in LTE-Advanced, with consideration of backward compatibility, are highlighted in this paper. Also, performance evaluation in macro-cellular scenarios through a simulation approach is presented, which shows the benefits of applying CA, low-complexity multi-band schedulers in service quality, system capacity enhancement and concluded that enhanced multi-band scheduler is less complex than the general multi-band scheduler, which performs better for a cell radius longer than 1800 m (and a PLR threshold of 2%).Keywords: Component carrier, carrier aggregation, LTE-Advanced, scheduling, spectrum management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5628532 Guided Wave Sensitivity for De-Bond Defects in Aluminum Skin-Honeycomb Core
Authors: A. Satour, F. Boubenider, R. Halimi, A. Badidibouda
Abstract:
Sandwich plates are finding an increasing range of application in the aircraft industry. The inspection of honeycomb composite structure by conventional ultrasonic technique is complex and very time consuming. The present study demonstrates a technique using guided Lamb waves at low frequencies to predict de-bond defects in aluminum skin-honeycomb core sandwich structure used in aeronautics. The numerical method was investigated for drawing the dispersion and displacement curves of ultrasonic Lamb wave propagated in Aluminum plate. An experimental study was carried out to check the theoretical prediction. The detection of unsticking between the skin and the core was tested by the two first modes for a low frequency. It was found that A0 mode is more sensitive to delamination defect compared to S0 mode.
Keywords: Damage detection, delamination, guided waves, Sandwich structure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23248531 Generation of Numerical Data for the Facilitation of the Personalized Hyperthermic Treatment of Cancer with An Interstital Antenna Array Using the Method of Symmetrical Components
Authors: Prodromos E. Atlamazoglou
Abstract:
The method of moments combined with the method of symmetrical components is used for the analysis of interstitial hyperthermia applicators. The basis and testing functions are both piecewise sinusoids, qualifying our technique as a Galerkin one. The dielectric coatings are modeled by equivalent volume polarization currents, which are simply related to the conduction current distribution, avoiding in that way the introduction of additional unknowns or numerical integrations. The results of our method for a four dipole circular array, are in agreement with those already published in literature for a same hyperthermia configuration. Apart from being accurate, our approach is more general, more computationally efficient and takes into account the coupling between the antennas.Keywords: Hyperthermia, integral equations, insulated antennas, method of symmetrical components.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6948530 Orthogonal Polynomial Density Estimates: Alternative Representation and Degree Selection
Authors: Serge B. Provost, Min Jiang
Abstract:
The density estimates considered in this paper comprise a base density and an adjustment component consisting of a linear combination of orthogonal polynomials. It is shown that, in the context of density approximation, the coefficients of the linear combination can be determined either from a moment-matching technique or a weighted least-squares approach. A kernel representation of the corresponding density estimates is obtained. Additionally, two refinements of the Kronmal-Tarter stopping criterion are proposed for determining the degree of the polynomial adjustment. By way of illustration, the density estimation methodology advocated herein is applied to two data sets.Keywords: kernel density estimation, orthogonal polynomials, moment-based methodologies, density approximation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23698529 Generating Class-Based Test Cases for Interface Classes of Object-Oriented Gray-Box Frameworks
Authors: Jehad Al Dallal, Paul Sorenson
Abstract:
An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define Framework Interface Classes (FICs) and their possible specifications, which helps in building reusable test cases for the implementations of these classes. In applications developed using gray-box frameworks, FICs inherit framework classes or use them without inheritance. In this paper, a test-case generation technique is extended to build test cases for FICs built for gray-box frameworks. A tool is developed to automate the introduced technique.Keywords: Class testing, object-oriented framework, reusable test case.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15178528 A Low Cost and High Quality Duty-Cycle Modulation Scheme and Applications
Authors: B. Lonla Moffo, J. Mbihi, L. Nneme Nneme
Abstract:
In this paper, a low cost duty-cycle modulation scheme is studied in depth and compared to the standard pulse width modulation technique. Using a mix of analytical reasoning and electronics simulation tools, it is shown that under the same operating conditions, most characteristics of the proposed duty-cycle modulation scheme are better than those provided by a standard pulse width modulation technique. The simulation results obtained when testing both modulation control policies on prototyping systems, indicate that the proposed duty-cycle modulation approach, appears to be a high quality control policy in a wide variety of application areas, including A/D and D/A conversion, signal transmission and switching control in power electronics.
Keywords: Duty-cycle Modulation, Operational amplifiers, Pulse width modulation, Power electronics, Signal processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27458527 Concept of Transforaminal Lumbar Interbody Fusion Cage Insertion Device
Authors: Sangram A. Sathe, Neha A. Madgulkar, Shruti S. Raut, S. P. Wadkar
Abstract:
Transforaminal lumbar interbody fusion (TLIF) surgeries have nowadays became popular for treatment of degenerated spinal disorders. The interbody fusion technique like TLIF maintains load bearing capacity of the spine and a suitable disc height. Currently many techniques have been introduced to cure Spondylolisthesis. This surgery provides greater rehabilitation of degenerative spines. While performing this TLIF surgery existing methods use guideway, which is a troublesome surgery technique as the use of two separate instruments is required to perform this surgery. This paper presents a concept which eliminates the use of guideway. This concept also eliminates problems that occur like reverting the cage. The concept discussed in this paper also gives high accuracy while performing surgery.Keywords: Degenerative disc diseases, pedicle screw, spine, spondylolisthesis, transforaminal lumbar interbody fusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13768526 The Optimal Indirect Vector Controller Design via an Adaptive Tabu Search Algorithm
Authors: P. Sawatnatee, S. Udomsuk, K-N. Areerak, K-L. Areerak, A. Srikaew
Abstract:
The paper presents how to design the indirect vector control of three-phase induction motor drive systems using the artificial intelligence technique called the adaptive tabu search. The results from the simulation and the experiment show that the drive system with the controller designed from the proposed method can provide the best output speed response compared with those of the conventional method. The controller design using the proposed technique can be used to create the software package for engineers to achieve the optimal controller design of the induction motor speed control based on the indirect vector concept.
Keywords: Indirect Vector Control, Induction Motor, Adaptive Tabu Search, Control Design, Artificial Intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19338525 Solving Directional Overcurrent Relay Coordination Problem Using Artificial Bees Colony
Authors: M. H. Hussain, I. Musirin, A. F. Abidin, S. R. A. Rahim
Abstract:
This paper presents the implementation of Artificial Bees Colony (ABC) algorithm in solving Directional OverCurrent Relays (DOCRs) coordination problem for near-end faults occurring in fixed network topology. The coordination optimization of DOCRs is formulated as linear programming (LP) problem. The objective function is introduced to minimize the operating time of the associated relay which depends on the time multiplier setting. The proposed technique is to taken as a technique for comparison purpose in order to highlight its superiority. The proposed algorithms have been tested successfully on 8 bus test system. The simulation results demonstrated that the ABC algorithm which has been proved to have good search ability is capable in dealing with constraint optimization problems.
Keywords: Artificial bees colony, directional overcurrent relay coordination problem, relay settings, time multiplier setting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35328524 Granulation using Clustering and Rough Set Theory and its Tree Representation
Authors: Girish Kumar Singh, Sonajharia Minz
Abstract:
Granular computing deals with representation of information in the form of some aggregates and related methods for transformation and analysis for problem solving. A granulation scheme based on clustering and Rough Set Theory is presented with focus on structured conceptualization of information has been presented in this paper. Experiments for the proposed method on four labeled data exhibit good result with reference to classification problem. The proposed granulation technique is semi-supervised imbibing global as well as local information granulation. To represent the results of the attribute oriented granulation a tree structure is proposed in this paper.Keywords: Granular computing, clustering, Rough sets, datamining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17198523 Optimal Dynamic Economic Load Dispatch Using Artificial Immune System
Authors: I. A. Farhat
Abstract:
The The dynamic economic dispatch (DED) problem is one of the complex constrained optimization problems that have nonlinear, con-convex and non-smooth objective functions. The purpose of the DED is to determine the optimal economic operation of the committed units while meeting the load demand. Associated to this constrained problem there exist highly nonlinear and non-convex practical constraints to be satisfied. Therefore, classical and derivative-based methods are likely not to converge to an optimal or near optimal solution to such a dynamic and large-scale problem. In this paper, an Artificial Immune System technique (AIS) is implemented and applied to solve the DED problem considering the transmission power losses and the valve-point effects in addition to the other operational constraints. To demonstrate the effectiveness of the proposed technique, two case studies are considered. The results obtained using the AIS are compared to those obtained by other methods reported in the literature and found better.
Keywords: Artificial Immune System (AIS), Dynamic Economic Dispatch (DED).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18848522 Improve Safety Performance of Un-Signalized Intersections in Oman
Authors: Siham G. Farag
Abstract:
The main objective of this paper is to provide a new methodology for road safety assessment in Oman through the development of suitable accident prediction models. GLM technique with Poisson or NBR using SAS package was carried out to develop these models. The paper utilized the accidents data of 31 un-signalized T-intersections during three years. Five goodness-of-fit measures were used to assess the overall quality of the developed models. Two types of models were developed separately; the flow-based models including only traffic exposure functions, and the full models containing both exposure functions and other significant geometry and traffic variables. The results show that, traffic exposure functions produced much better fit to the accident data. The most effective geometric variables were major-road mean speed, minor-road 85th percentile speed, major-road lane width, distance to the nearest junction, and right-turn curb radius. The developed models can be used for intersection treatment or upgrading and specify the appropriate design parameters of T-intersections. Finally, the models presented in this thesis reflect the intersection conditions in Oman and could represent the typical conditions in several countries in the middle east area, especially gulf countries.
Keywords: Accidents Prediction Models (APMs), Generalized Linear Model (GLM), T-intersections, Oman.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2063