Search results for: Multivariate Statistical Analysis.
7623 Numerical Modeling of Artisanal and Small-Scale Mining of Coltan in the African Great Lakes Region
Authors: Sergio Perez Rodriguez
Abstract:
Findings of a production model of Artisanal and Small-Scale Mining (ASM) of coltan ore by an average Democratic Republic of Congo (DRC) mineworker are presented in this paper. These can be used as a reference for a similar characterization of the daily labor of counterparts from other countries in the Africa's Great Lakes region. To that end, the Fundamental Equation of Mineral Production has been applied in this paper, considering a miner's average daily output of coltan, estimated in the base of gross statistical data gathered from reputable sources. Results indicate daily yields of individual miners in the order of 300 g of coltan ore, with hourly peaks of production in the range of 30 to 40 g of the mineral. Yields are expected to be in the order of 5 g or less during the least productive hours. These outputs are expected to be achieved during the halves of the eight to 10 hours of daily working sessions that these artisanal laborers can attend during the mining season.
Keywords: Coltan, mineral production, Production to Reserve ratio, artisanal mining, small-scale mining, ASM, human work, Great Lakes region, Democratic Republic of Congo.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1947622 Environmental Impacts of Point and Non-Point Source Pollution in Krishnagiri Reservoir: A Case Study in South India
Authors: N. K. Ambujam, V. Sudha
Abstract:
Reservoirs are being contaminated all around the world with point source and Non-Point Source (NPS) pollution. The most common NPS pollutants are sediments and nutrients. Krishnagiri Reservoir (KR) has been chosen for the present case study, which is located in the tropical semi-arid climatic zone of Tamil Nadu, South India. It is the main source of surface water in Krishnagiri district to meet the freshwater demands. The reservoir has lost about 40% of its water holding capacity due to sedimentation over the period of 50 years. Hence, from the research and management perspective, there is a need for a sound knowledge on the spatial and seasonal variations of KR water quality. The present study encompasses the specific objectives as (i) to investigate the longitudinal heterogeneity and seasonal variations of physicochemical parameters, nutrients and biological characteristics of KR water and (ii) to examine the extent of degradation of water quality in KR. 15 sampling points were identified by uniform stratified method and a systematic monthly sampling strategy was selected due to high dynamic nature in its hydrological characteristics. The physicochemical parameters, major ions, nutrients and Chlorophyll a (Chl a) were analysed. Trophic status of KR was classified by using Carlson's Trophic State Index (TSI). All statistical analyses were performed by using Statistical Package for Social Sciences programme, version-16.0. Spatial maps were prepared for Chl a using Arc GIS. Observations in KR pointed out that electrical conductivity and major ions are highly variable factors as it receives inflow from the catchment with different land use activities. The study of major ions in KR exhibited different trends in their values and it could be concluded that as the monsoon progresses the major ions in the water decreases or water quality stabilizes. The inflow point of KR showed comparatively higher concentration of nutrients including nitrate, soluble reactive phosphorus (SRP), total phosphors (TP), total suspended phosphorus (TSP) and total dissolved phosphorus (TDP) during monsoon seasons. This evidently showed the input of significant amount of nutrients from the catchment side through agricultural runoff. High concentration of TDP and TSP at the lacustrine zone of the reservoir during summer season evidently revealed that there was a significant release of phosphorus from the bottom sediments. Carlson’s TSI of KR ranged between 81 and 92 during northeast monsoon and summer seasons. High and permanent Cyanobacterial bloom in KR could be mainly due to the internal loading of phosphorus from the bottom sediments. According to Carlson’s TSI classification Krishnagiri reservoir was ranked in the hyper-eutrophic category. This study provides necessary basic data on the spatio-temporal variations of water quality in KR and also proves the impact of point and NPS pollution from the catchment area. High TSI warrants a greater threat for the recovery of internal P loading and hyper-eutrophic condition of KR. Several expensive internal measures for the reduction of internal loading of P were introduced by many scientists. However, the outcome of the present research suggests for the innovative algae harvesting technique for the removal of sediment nutrients.
Keywords: Hyper-eutrophication, Krishnagiri reservoir, nutrients, NPS pollution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16227621 Parallelization of Ensemble Kalman Filter (EnKF) for Oil Reservoirs with Time-lapse Seismic Data
Authors: Md Khairullah, Hai-Xiang Lin, Remus G. Hanea, Arnold W. Heemink
Abstract:
In this paper we describe the design and implementation of a parallel algorithm for data assimilation with ensemble Kalman filter (EnKF) for oil reservoir history matching problem. The use of large number of observations from time-lapse seismic leads to a large turnaround time for the analysis step, in addition to the time consuming simulations of the realizations. For efficient parallelization it is important to consider parallel computation at the analysis step. Our experiments show that parallelization of the analysis step in addition to the forecast step has good scalability, exploiting the same set of resources with some additional efforts.
Keywords: EnKF, Data assimilation, Parallel computing, Parallel efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22817620 Thermal Analysis of Toroidal Transformers Using Finite Element Method
Authors: Adrian T.
Abstract:
In this paper a three dimensional thermal model of a power toroidal transformer is proposed for both steady-state or transient conditions. The influence of electric current and ambient temperature on the temperature distribution, has been investigated. To validate the three dimensional thermal model, some experimental tests have been done. There is a good correlation between experimental and simulation results.Keywords: Temperature distribution, thermal analysis, toroidal transformer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35787619 Power Flow Analysis for Radial Distribution System Using Backward/Forward Sweep Method
Authors: J. A. Michline Rupa, S. Ganesh
Abstract:
This paper proposes a backward/forward sweep method to analyze the power flow in radial distribution systems. The distribution system has radial structure and high R/X ratios. So the newton-raphson and fast decoupled methods are failed with distribution system. The proposed method presents a load flow study using backward/forward sweep method, which is one of the most effective methods for the load-flow analysis of the radial distribution system. By using this method, power losses for each bus branch and voltage magnitudes for each bus node are determined. This method has been tested on IEEE 33-bus radial distribution system and effective results are obtained using MATLAB.
Keywords: Backward/Forward sweep method, Distribution system, Load flow analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 175217618 Review of Surface Electromyogram Signals: Its Analysis and Applications
Authors: Anjana Goen, D. C. Tiwari
Abstract:
Electromyography (EMG) is the study of muscles function through analysis of electrical activity produced from muscles. This electrical activity which is displayed in the form of signal is the result of neuromuscular activation associated with muscle contraction. The most common techniques of EMG signal recording are by using surface and needle/wire electrode where the latter is usually used for interest in deep muscle. This paper will focus on surface electromyogram (SEMG) signal. During SEMG recording, several problems had to been countered such as noise, motion artifact and signal instability. Thus, various signal processing techniques had been implemented to produce a reliable signal for analysis. SEMG signal finds broad application particularly in biomedical field. It had been analyzed and studied for various interests such as neuromuscular disease, enhancement of muscular function and human-computer interface.
Keywords: Evolvable hardware (EHW), Functional Electrical Simulation (FES), Hidden Markov Model (HMM), Hjorth Time Domain (HTD).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35167617 Study on Performance of Wigner Ville Distribution for Linear FM and Transient Signal Analysis
Authors: Azeemsha Thacham Poyil, Nasimudeen KM
Abstract:
This research paper presents some methods to assess the performance of Wigner Ville Distribution for Time-Frequency representation of non-stationary signals, in comparison with the other representations like STFT, Spectrogram etc. The simultaneous timefrequency resolution of WVD is one of the important properties which makes it preferable for analysis and detection of linear FM and transient signals. There are two algorithms proposed here to assess the resolution and to compare the performance of signal detection. First method is based on the measurement of area under timefrequency plot; in case of a linear FM signal analysis. A second method is based on the instantaneous power calculation and is used in case of transient, non-stationary signals. The implementation is explained briefly for both methods with suitable diagrams. The accuracy of the measurements is validated to show the better performance of WVD representation in comparison with STFT and Spectrograms.
Keywords: WVD: Wigner Ville Distribution, STFT: Short Time Fourier Transform, FT: Fourier Transform, TFR: Time-Frequency Representation, FM: Frequency Modulation, LFM Signal: Linear FM Signal, JTFA: Joint time frequency analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24237616 Bayesian Inference for Phase Unwrapping Using Conjugate Gradient Method in One and Two Dimensions
Authors: Yohei Saika, Hiroki Sakaematsu, Shota Akiyama
Abstract:
We investigated statistical performance of Bayesian inference using maximum entropy and MAP estimation for several models which approximated wave-fronts in remote sensing using SAR interferometry. Using Monte Carlo simulation for a set of wave-fronts generated by assumed true prior, we found that the method of maximum entropy realized the optimal performance around the Bayes-optimal conditions by using model of the true prior and the likelihood representing optical measurement due to the interferometer. Also, we found that the MAP estimation regarded as a deterministic limit of maximum entropy almost achieved the same performance as the Bayes-optimal solution for the set of wave-fronts. Then, we clarified that the MAP estimation perfectly carried out phase unwrapping without using prior information, and also that the MAP estimation realized accurate phase unwrapping using conjugate gradient (CG) method, if we assumed the model of the true prior appropriately.
Keywords: Bayesian inference using maximum entropy, MAP estimation using conjugate gradient method, SAR interferometry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17517615 A Study on Early Prediction of Fault Proneness in Software Modules using Genetic Algorithm
Authors: Parvinder S. Sandhu, Sunil Khullar, Satpreet Singh, Simranjit K. Bains, Manpreet Kaur, Gurvinder Singh
Abstract:
Fault-proneness of a software module is the probability that the module contains faults. To predict faultproneness of modules different techniques have been proposed which includes statistical methods, machine learning techniques, neural network techniques and clustering techniques. The aim of proposed study is to explore whether metrics available in the early lifecycle (i.e. requirement metrics), metrics available in the late lifecycle (i.e. code metrics) and metrics available in the early lifecycle (i.e. requirement metrics) combined with metrics available in the late lifecycle (i.e. code metrics) can be used to identify fault prone modules using Genetic Algorithm technique. This approach has been tested with real time defect C Programming language datasets of NASA software projects. The results show that the fusion of requirement and code metric is the best prediction model for detecting the faults as compared with commonly used code based model.Keywords: Genetic Algorithm, Fault Proneness, Software Faultand Software Quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19847614 Predicting DHF Incidence in Northern Thailand using Time Series Analysis Technique
Authors: S. Wongkoon, M. Pollar, M. Jaroensutasinee, K. Jaroensutasinee
Abstract:
This study aimed at developing a forecasting model on the number of Dengue Haemorrhagic Fever (DHF) incidence in Northern Thailand using time series analysis. We developed Seasonal Autoregressive Integrated Moving Average (SARIMA) models on the data collected between 2003-2006 and then validated the models using the data collected between January-September 2007. The results showed that the regressive forecast curves were consistent with the pattern of actual values. The most suitable model was the SARIMA(2,0,1)(0,2,0)12 model with a Akaike Information Criterion (AIC) of 12.2931 and a Mean Absolute Percent Error (MAPE) of 8.91713. The SARIMA(2,0,1)(0,2,0)12 model fitting was adequate for the data with the Portmanteau statistic Q20 = 8.98644 ( x20,95= 27.5871, P>0.05). This indicated that there was no significant autocorrelation between residuals at different lag times in the SARIMA(2,0,1)(0,2,0)12 model.
Keywords: Dengue, SARIMA, Time Series Analysis, Northern Thailand.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19907613 Stress and Strain Analysis of Notched Bodies Subject to Non-Proportional Loadings
Authors: A. Ince
Abstract:
In this paper, an analytical simplified method for calculating elasto-plastic stresses strains of notched bodies subject to non-proportional loading paths is discussed. The method was based on the Neuber notch correction, which relates the incremental elastic and elastic-plastic strain energy densities at the notch root and the material constitutive relationship. The validity of the method was presented by comparing computed results of the proposed model against finite element numerical data of notched shaft. The comparison showed that the model estimated notch-root elasto-plastic stresses strains with good accuracy using linear-elastic stresses. The prosed model provides more efficient and simple analysis method preferable to expensive experimental component tests and more complex and time consuming incremental non-linear FE analysis. The model is particularly suitable to perform fatigue life and fatigue damage estimates of notched components subjected to nonproportional loading paths.Keywords: Elasto-plastic, stress-strain, notch analysis, nonprortional loadings, cyclic plasticity, fatigue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25577612 Harmonic Analysis and Performance Improvement of a Wind Energy Conversions System with Double Output Induction Generator
Authors: M. Sedighizadeh, A. Rezazadeh
Abstract:
Wind turbines with double output induction generators can operate at variable speed permitting conversion efficiency maximization over a wide range of wind velocities. This paper presents the performance analysis of a wind driven double output induction generator (DOIG) operating at varying shafts speed. A periodic transient state analysis of DOIG equipped with two converters is carried out using a hybrid induction machine model. This paper simulates the harmonic content of waveforms in various points of drive at different speeds, based on the hybrid model (dqabc). Then the sinusoidal and trapezoidal pulse-width–modulation control techniques are used in order to improve the power factor of the machine and to weaken the injected low order harmonics to the supply. Based on the frequency spectrum, total harmonics distortion, distortion factor and power factor. Finally advantages of sinusoidal and trapezoidal pulse width modulation techniques are compared.Keywords: DOIG, Harmonic Analysis, Wind.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18017611 Predicting Crack Initiation Due to Ratchetting in Rail Heads Using Critical Element Analysis
Authors: I. U. Wickramasinghe, D. J. Hargreaves, D. V. De Pellegrin
Abstract:
This paper presents a strategy to predict the lifetime of rails subjected to large rolling contact loads that induce ratchetting strains in the rail head. A critical element concept is used to calculate the number of loading cycles needed for crack initiation to occur in the rail head surface. In this technique the finite element method (FEM) is used to determine the maximum equivalent ratchetting strain per load cycle, which is calculated by combining longitudinal and shear stains in the critical element. This technique builds on a previously developed critical plane concept that has been used to calculate the number of cycles to crack initiation in rolling contact fatigue under ratchetting failure conditions. The critical element concept simplifies the analytical difficulties of critical plane analysis. Finite element analysis (FEA) is used to identify the critical element in the mesh, and then the strain values of the critical element are used to calculate the ratchetting rate analytically. Finally, a ratchetting criterion is used to calculate the number of cycles to crack initiation from the ratchetting rate calculated.
Keywords: Critical element analysis, finite element modeling (FEM), wheel/rail contact.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29337610 Reliability Analysis of Computer Centre at Yobe State University Using LRU Algorithm
Authors: V. V. Singh, Yusuf Ibrahim Gwanda, Rajesh Prasad
Abstract:
In this paper, we focus on the reliability and performance analysis of Computer Centre (CC) at Yobe State University, Damaturu, Nigeria. The CC consists of three servers: one database mail server, one redundant and one for sharing with the client computers in the CC (called as a local server). Observing the different possibilities of the functioning of the CC, the analysis has been done to evaluate the various popular measures of reliability such as availability, reliability, mean time to failure (MTTF), profit analysis due to the operation of the system. The system can ultimately fail due to the failure of router, redundant server before repairing the mail server and switch failure. The system can also partially fail when a local server fails. The failed devices have restored according to Least Recently Used (LRU) techniques. The system can also fail entirely due to a cooling failure of the server, electricity failure or some natural calamity like earthquake, fire tsunami, etc. All the failure rates are assumed to be constant and follow exponential time distribution, while the repair follows two types of distributions: i.e. general and Gumbel-Hougaard family copula distribution.Keywords: Reliability, availability Gumbel-Hougaard family copula, MTTF, internet data center.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8707609 Parameter Sensitivity Analysis of Artificial Neural Network for Predicting Water Turbidity
Authors: Chia-Ling Chang, Chung-Sheng Liao
Abstract:
The present study focuses on the discussion over the parameter of Artificial Neural Network (ANN). Sensitivity analysis is applied to assess the effect of the parameters of ANN on the prediction of turbidity of raw water in the water treatment plant. The result shows that transfer function of hidden layer is a critical parameter of ANN. When the transfer function changes, the reliability of prediction of water turbidity is greatly different. Moreover, the estimated water turbidity is less sensitive to training times and learning velocity than the number of neurons in the hidden layer. Therefore, it is important to select an appropriate transfer function and suitable number of neurons in the hidden layer in the process of parameter training and validation.Keywords: Artificial Neural Network (ANN), sensitivity analysis, turbidity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28137608 Fighter Aircraft Selection Using Neutrosophic Multiple Criteria Decision Making Analysis
Authors: C. Ardil
Abstract:
Fuzzy set and intuitionistic fuzzy set are dealing with the imprecision and uncertainty inherent in a complex decision problem. However, sometimes these theories are not sufficient to model indeterminate and inconsistent information encountered in real-life problems. To overcome this insufficiency, the neutrosophic set, which is useful in practical applications, is proposed, triangular neutrosophic numbers and trapezoidal neutrosophic numbers are examined, their definitions and applications are discussed. In this study, a decision making algorithm is developed using neutrosophic set processes and an application is given in fighter aircraft selection as an example of a decision making problem. The estimation of the fighter aircraft selection with the neutrosophic multiple criteria decision analysis method is examined.
Keywords: neutrosophic set, multiple criteria decision making analysis, fighter aircraft selection, MCDMA, neutrosophic numbers
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9347607 Using Project MIND - Math Is Not Difficult Strategies to Help Children with Autism Improve Mathematics Skills
Authors: Hui Fang Huang Su, Leanne Lai, Pei-Fen Li, Mei-Hwei Ho, Yu-Wen Chiu
Abstract:
This study aimed to provide a practical, systematic, and comprehensive intervention for children with Autism Spectrum Disorder (ASD). A pilot study of quasi-experimental pre-post intervention with control group design was conducted to evaluate if the mathematical intervention (Project MIND - Math Is Not Difficult) increases the math comprehension of children with ASD Children with ASD in the primary grades (K-1, 2) participated in math interventions to enhance their math comprehension and cognitive ability. The Bracken basic concept scale was used to evaluate subjects’ language skills, cognitive development, and school readiness. The study found that our systemic interventions of Project MIND significantly improved the mathematical and cognitive abilities in children with autism. The results of this study may lead to a major change in effective and adequate health care services for children with ASD and their families. All statistical analyses were performed with the IBM SPSS Statistics Version 25 for Windows. The significant level was set at 0.05 P-value.Keywords: Young Children, Autism, Mathematics, Curriculum, teaching and learning, children with special needs, Project MIND.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9277606 A Monte Carlo Method to Data Stream Analysis
Authors: Kittisak Kerdprasop, Nittaya Kerdprasop, Pairote Sattayatham
Abstract:
Data stream analysis is the process of computing various summaries and derived values from large amounts of data which are continuously generated at a rapid rate. The nature of a stream does not allow a revisit on each data element. Furthermore, data processing must be fast to produce timely analysis results. These requirements impose constraints on the design of the algorithms to balance correctness against timely responses. Several techniques have been proposed over the past few years to address these challenges. These techniques can be categorized as either dataoriented or task-oriented. The data-oriented approach analyzes a subset of data or a smaller transformed representation, whereas taskoriented scheme solves the problem directly via approximation techniques. We propose a hybrid approach to tackle the data stream analysis problem. The data stream has been both statistically transformed to a smaller size and computationally approximated its characteristics. We adopt a Monte Carlo method in the approximation step. The data reduction has been performed horizontally and vertically through our EMR sampling method. The proposed method is analyzed by a series of experiments. We apply our algorithm on clustering and classification tasks to evaluate the utility of our approach.Keywords: Data Stream, Monte Carlo, Sampling, DensityEstimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14177605 A Simulated Design and Analysis of a Solar Thermal Parabolic Trough Concentrator
Authors: Fauziah Sulaiman, Nurhayati Abdullah, Balbir Singh Mahinder Singh
Abstract:
In recent years Malaysia has included renewable energy as an alternative fuel to help in diversifying the country-s energy reliance on oil, natural gas, coal and hydropower with biomass and solar energy gaining priority. The scope of this paper is to look at the designing procedures and analysis of a solar thermal parabolic trough concentrator by simulation utilizing meteorological data in several parts of Malaysia. Parameters which include the aperture area, the diameter of the receiver and the working fluid may be varied to optimize the design. Aperture area is determined by considering the width and the length of the concentrator whereas the geometric concentration ratio (CR) is obtained by considering the width and diameter of the receiver. Three types of working fluid are investigated. Theoretically, concentration ratios can be very high in the range of 10 to 40 000 depending on the optical elements used and continuous tracking of the sun. However, a thorough analysis is essential as discussed in this paper where optical precision and thermal analysis must be carried out to evaluate the performance of the parabolic trough concentrator as the theoretical CR is not the only factor that should be considered.Keywords: Parabolic trough concentrator, Concentration ratio, Intercept factor, Efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39817604 Aircraft Selection Process Using Preference Analysis for Reference Ideal Solution (PARIS)
Authors: C. Ardil
Abstract:
Multiple criteria decision making analysis (MCDMA) methods are applied to many real - life problems in different fields of engineering science and technology. The "preference analysis for reference ideal solution (PARIS)" method is proposed for an efficient MCDMA evaluation of decision problems. The multiple criteria aircraft evaluation approach is based on the integrated the mean weight, entropy weight, PARIS, and TOPSIS method, which eliminates the subjective importance weight assignment process. The evaluation criteria were identified from an extensive literature review of aircraft selection process. The aim of this study is to propose an efficient methodology for handling the aircraft selection process in which the proposed method solves effectively the MCDMA problem. A numerical example is presented to demonstrate the applicability and validity of the proposed MCDMA approach.
Keywords: aircraft selection, aircraft, multiple criteria decision making, multiple criteria decision making analysis, mean weight, entropy weight, MCDMA, PARIS, TOPSIS, VIKOR, ELECTRE, PROMETHEE
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5367603 Geospatial Network Analysis Using Particle Swarm Optimization
Authors: Varun Singh, Mainak Bandyopadhyay, Maharana Pratap Singh
Abstract:
The shortest path (SP) problem concerns with finding the shortest path from a specific origin to a specified destination in a given network while minimizing the total cost associated with the path. This problem has widespread applications. Important applications of the SP problem include vehicle routing in transportation systems particularly in the field of in-vehicle Route Guidance System (RGS) and traffic assignment problem (in transportation planning). Well known applications of evolutionary methods like Genetic Algorithms (GA), Ant Colony Optimization, Particle Swarm Optimization (PSO) have come up to solve complex optimization problems to overcome the shortcomings of existing shortest path analysis methods. It has been reported by various researchers that PSO performs better than other evolutionary optimization algorithms in terms of success rate and solution quality. Further Geographic Information Systems (GIS) have emerged as key information systems for geospatial data analysis and visualization. This research paper is focused towards the application of PSO for solving the shortest path problem between multiple points of interest (POI) based on spatial data of Allahabad City and traffic speed data collected using GPS. Geovisualization of results of analysis is carried out in GIS.
Keywords: GIS, Outliers, PSO, Traffic Data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28927602 Trusting Smart Speakers: Analysing the Different Levels of Trust between Technologies
Authors: Alec Wells, Aminu Bello Usman, Justin McKeown
Abstract:
The growing usage of smart speakers raises many privacy and trust concerns compared to other technologies such as smart phones and computers. In this study, a proxy measure of trust is used to gauge users’ opinions on three different technologies based on an empirical study, and to understand which technology most people are most likely to trust. The collected data were analysed using the Kruskal-Wallis H test to determine the statistical differences between the users’ trust level of the three technologies: smart speaker, computer and smart phone. The findings of the study revealed that despite the wide acceptance, ease of use and reputation of smart speakers, people find it difficult to trust smart speakers with their sensitive information via the Direct Voice Input (DVI) and would prefer to use a keyboard or touchscreen offered by computers and smart phones. Findings from this study can inform future work on users’ trust in technology based on perceived ease of use, reputation, perceived credibility and risk of using technologies via DVI.
Keywords: Direct voice input, risk, security, technology and trust.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5927601 Semi-Automatic Artifact Rejection Procedure Based on Kurtosis, Renyi's Entropy and Independent Component Scalp Maps
Authors: Antonino Greco, Nadia Mammone, Francesco Carlo Morabito, Mario Versaci
Abstract:
Artifact rejection plays a key role in many signal processing applications. The artifacts are disturbance that can occur during the signal acquisition and that can alter the analysis of the signals themselves. Our aim is to automatically remove the artifacts, in particular from the Electroencephalographic (EEG) recordings. A technique for the automatic artifact rejection, based on the Independent Component Analysis (ICA) for the artifact extraction and on some high order statistics such as kurtosis and Shannon-s entropy, was proposed some years ago in literature. In this paper we try to enhance this technique proposing a new method based on the Renyi-s entropy. The performance of our method was tested and compared to the performance of the method in literature and the former proved to outperform the latter.
Keywords: Artifact, EEG, Renyi's entropy, kurtosis, independent component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18567600 Process Analysis through Length Consistency
Authors: James E. Ponder
Abstract:
The requirement for consistency in physics can sometimes offer a common ground between disciplines such that their fundamental equations share a common parameter set and mathematical method for equation extraction. The parameter set shared by Relativity and Quantum Wave Mechanics enables an analysis which will be seen to be very straightforward, primarily classical in nature using linear algebra concepts, yet deriving a theoretical estimate of the value of the Gravitational Constant along with dependencies never before known.
Keywords: Gravitational Constant, Physical Consistency, Quantum Mechanics, Relativity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15397599 Investigation on the Stability of Rock Slopes Subjected to Tension Cracks via Limit Analysis
Abstract:
Based on the kinematic approach of limit analysis, a full set of upper bound solutions for the stability of homogeneous rock slopes subjected to tension cracks are obtained. The generalized Hoek-Brown failure criterion is employed to describe the non-linear strength envelope of rocks. In this paper, critical failure mechanisms are determined for cracks of known depth but unspecified location, cracks of known location but unknown depth, and cracks of unspecified location and depth. It is shown that there is a nearly up to 50% drop in terms of the stability factors for the rock slopes intersected by a tension crack compared with intact ones. Tables and charts of solutions in dimensionless forms are presented for ease of use by practitioners.
Keywords: Hoek-Brown failure criterion, limit analysis, rock slope, tension cracks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24527598 Reform-Oriented Teaching of Introductory Statistics in the Health, Social and Behavioral Sciences – Historical Context and Rationale
Authors: Rossi A. Hassad
Abstract:
There is widespread emphasis on reform in the teaching of introductory statistics at the college level. Underpinning this reform is a consensus among educators and practitioners that traditional curricular materials and pedagogical strategies have not been effective in promoting statistical literacy, a competency that is becoming increasingly necessary for effective decision-making and evidence-based practice. This paper explains the historical context of, and rationale for reform-oriented teaching of introductory statistics (at the college level) in the health, social and behavioral sciences (evidence-based disciplines). A firm understanding and appreciation of the basis for change in pedagogical approach is important, in order to facilitate commitment to reform, consensus building on appropriate strategies, and adoption and maintenance of best practices. In essence, reform-oriented pedagogy, in this context, is a function of the interaction among content, pedagogy, technology, and assessment. The challenge is to create an appropriate balance among these domains.
Keywords: Reform-oriented, reform, introductory statistics, health, behavioral sciences, evidence-based, psychology, teaching, learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9967597 Aerodynamic Analysis of Dimple Effect on Aircraft Wing
Authors: E. Livya, G. Anitha, P. Valli
Abstract:
The main objective of aircraft aerodynamics is to enhance the aerodynamic characteristics and maneuverability of the aircraft. This enhancement includes the reduction in drag and stall phenomenon. The airfoil which contains dimples will have comparatively less drag than the plain airfoil. Introducing dimples on the aircraft wing will create turbulence by creating vortices which delays the boundary layer separation resulting in decrease of pressure drag and also increase in the angle of stall. In addition, wake reduction leads to reduction in acoustic emission. The overall objective of this paper is to improve the aircraft maneuverability by delaying the flow separation point at stall and thereby reducing the drag by applying the dimple effect over the aircraft wing. This project includes both computational and experimental analysis of dimple effect on aircraft wing, using NACA 0018 airfoil. Dimple shapes of Semi-sphere, hexagon, cylinder, square are selected for the analysis; airfoil is tested under the inlet velocity of 30m/s and 60m/s at different angle of attack (5˚, 10˚, 15˚, 20˚, and 25˚). This analysis favors the dimple effect by increasing L/D ratio and thereby providing the maximum aerodynamic efficiency, which provides the enhanced performance for the aircraft.
Keywords: Airfoil, Boundary layer, Dimple effect, Flow separation, Stall reduction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 61487596 Performance Analysis of Self Excited Induction Generator Using Artificial Bee Colony Algorithm
Authors: A. K. Sharma, N. P. Patidar, G. Agnihotri, D. K. Palwalia
Abstract:
This paper presents the performance state analysis of Self-Excited Induction Generator (SEIG) using Artificial Bee Colony (ABC) optimization technique. The total admittance of the induction machine is minimized to calculate the frequency and magnetizing reactance corresponding to any rotor speed, load impedance and excitation capacitance. The performance of SEIG is calculated using the optimized parameter found. The results obtained by ABC algorithm are compared with results from numerical method. The results obtained coincide with the numerical method results. This technique proves to be efficient in solving nonlinear constrained optimization problems and analyzing the performance of SEIG.
Keywords: Artificial bee colony, Steady state analysis, Selfexcited induction generator, Nonlinear constrained optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21807595 Seismic Retrofitting of RC Buildings with Soft Storey and Floating Columns
Authors: Vinay Agrawal, Suyash Garg, Ravindra Nagar, Vinay Chandwani
Abstract:
Open ground storey with floating columns is a typical feature in the modern multistory constructions in urban India. Such features are very much undesirable in buildings built in seismically active areas. The present study proposes a feasible solution to mitigate the effects caused due to non-uniformity of stiffness and discontinuity in load path and to simultaneously hold the functional use of the open storey particularly under the floating column, through a combination of various lateral strengthening systems. An investigation is performed on an example building with nine different analytical models to bring out the importance of recognising the presence of open ground storey and floating columns. Two separate analyses on various models of the building namely, the equivalent static analysis and the response spectrum analysis as per IS: 1893-2002 were performed. Various measures such as incorporation of Chevron bracings and shear walls, strengthening the columns in the open ground storey, and their different combinations were examined. The analysis shows that, in comparison to two short ones separated by interconnecting beams, the structural walls are most effective when placed at the periphery of the buildings and used as one long structural wall. Further, it can be shown that the force transfer from floating columns becomes less horizontal when the Chevron Bracings are placed just below them, thereby reducing the shear forces in the beams on which the floating column rests.
Keywords: Equivalent static analysis, floating column, open ground storey, response spectrum analysis, shear wall, stiffness irregularity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15357594 Research on Simulation Model of Collision Force between Floating Ice and Pier
Authors: Tianlai Yu, Zhengguo Yuan, Sidi Shan
Abstract:
Adopting the measured constitutive relationship of stress-strain of river ice, the finite element analysis model of percussive force of river ice and pier is established, by the explicit dynamical analysis software package LS-DYNA. Effects of element types, contact method and arithmetic of ice and pier, coupled modes between different elements, mesh density of pier, and ice sheet in contact area on the collision force are studied. Some of measures for the collision force analysis of river ice and pier are proposed as follows: bridge girder can adopt beam161 element with 3-node; pier below the line of 1.30m above ice surface and ice sheet use solid164 element with 8-node; in order to accomplish the connection of different elements, the rigid body with 0.01-0.05m thickness is defined between solid164 and beam161; the contact type of ice and pier adopts AUTOMATIC_SURFACE_TO_SURFACE, using symmetrical penalty function algorithms; meshing size of pier below the line of 1.30m above ice surface should not less than 0.25×0.25×0.5m3. The simulation results have the advantage of high precision by making a comparison between measured and computed data. The research results can be referred for collision force study between river ice and pier.Keywords: River ice, collision force, simulation analysis, ANSYS/LS-DYNA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2046