Search results for: Location Estimate.
1296 Useful Lifetime Prediction of Rail Pads for High Speed Trains
Authors: Chang Su Woo, Hyun Sung Park
Abstract:
Useful lifetime evaluation of railpads were very important in design procedure to assure the safety and reliability. It is, therefore, necessary to establish a suitable criterion for the replacement period of rail pads. In this study, we performed properties and accelerated heat aging tests of rail pads considering degradation factors and all environmental conditions including operation, and then derived a lifetime prediction equation according to changes in hardness, thickness, and static spring constants in the Arrhenius plot to establish how to estimate the aging of rail pads. With the useful lifetime prediction equation, the lifetime of e-clip pads was 2.5 years when the change in hardness was 10% at 25°C; and that of f-clip pads was 1.7 years. When the change in thickness was 10%, the lifetime of e-clip pads and f-clip pads is 2.6 years respectively. The results obtained in this study to estimate the useful lifetime of rail pads for high speed trains can be used for determining the maintenance and replacement schedule for rail pads.
Keywords: Rail pads, accelerated test, Arrhenius plot, useful lifetime prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28051295 Idiopathic Constipation can be Subdivided in Clinical Subtypes: Data Mining by Cluster Analysis on a Population based Study
Authors: Mauro Giacomini, Stefania Bertone, Carlo Mansi, Pietro Dulbecco, Vincenzo Savarino
Abstract:
The prevalence of non organic constipation differs from country to country and the reliability of the estimate rates is uncertain. Moreover, the clinical relevance of subdividing the heterogeneous functional constipation disorders into pre-defined subgroups is largely unknown.. Aim: to estimate the prevalence of constipation in a population-based sample and determine whether clinical subgroups can be identified. An age and gender stratified sample population from 5 Italian cities was evaluated using a previously validated questionnaire. Data mining by cluster analysis was used to determine constipation subgroups. Results: 1,500 complete interviews were obtained from 2,083 contacted households (72%). Self-reported constipation correlated poorly with symptombased constipation found in 496 subjects (33.1%). Cluster analysis identified four constipation subgroups which correlated to subgroups identified according to pre-defined symptom criteria. Significant differences in socio-demographics and lifestyle were observed among subgroups.Keywords: Cluster analysis, constipation, data mining, statistical analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12981294 A Commercial Building Plug Load Management System That Uses Internet of Things Technology to Automatically Identify Plugged-In Devices and Their Locations
Authors: Amy LeBar, Kim L. Trenbath, Bennett Doherty, William Livingood
Abstract:
Plug and process loads (PPLs) account for a large portion of U.S. commercial building energy use. There is a huge potential to reduce whole building consumption by targeting PPLs for energy savings measures or implementing some form of plug load management (PLM). Despite this potential, there has yet to be a widely adopted commercial PLM technology. This paper describes the Automatic Type and Location Identification System (ATLIS), a PLM system framework with automatic and dynamic load detection (ADLD). ADLD gives PLM systems the ability to automatically identify devices as they are plugged into the outlets of a building. The ATLIS framework takes advantage of smart, connected devices to identify device locations in a building, meter and control their power, and communicate this information to a central database. ATLIS includes five primary capabilities: location identification, communication, control, energy metering, and data storage. A laboratory proof of concept (PoC) demonstrated all but the energy metering capability, and these capabilities were validated using a series of system tests. The PoC was able to identify when a device was plugged into an outlet and the location of the device in the building. When a device was moved, the PoC’s dashboard and database were automatically updated with the new location. The PoC implemented controls to devices from the system dashboard so that devices maintained correct schedules regardless of where they were plugged in within the building. ATLIS’s primary technology application is improved PLM, but other applications include asset management, energy audits, and interoperability for grid-interactive efficient buildings. An ATLIS-based system could also be used to direct power to critical devices, such as ventilators, during a brownout or blackout. Such a framework is an opportunity to make PLM more widespread and reduce the amount of energy consumed by PPLs in current and future commercial buildings.
Keywords: commercial buildings, grid-interactive efficient buildings, miscellaneous electric loads, plug loads, plug load management
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8781293 A Novel Approach of Power Transformer Diagnostic Using 3D FEM Parametrical Model
Authors: M. Brandt, A. Peniak, J. Makarovič, P. Rafajdus
Abstract:
This paper deals with a novel approach of power transformers diagnostics. This approach identifies the exact location and the range of a fault in the transformer and helps to reduce operation costs related to handling of the faulty transformer, its disassembly and repair. The advantage of the approach is a possibility to simulate healthy transformer and also all faults, which can occur in transformer during its operation without its disassembling, which is very expensive in practice. The approach is based on creating frequency dependent impedance of the transformer by sweep frequency response analysis measurements and by 3D FE parametrical modeling of the fault in the transformer. The parameters of the 3D FE model are the position and the range of the axial short circuit. Then, by comparing the frequency dependent impedances of the parametrical models with the measured ones, the location and the range of the fault is identified. The approach was tested on a real transformer and showed high coincidence between the real fault and the simulated one.
Keywords: Fault, finite element method, parametrical model of transformer, sweep frequency response analysis, transformer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20291292 A Study of Students’ Perceptions Regarding the Effectiveness of Semester and Annual Examination System at Institute of Education and Research
Authors: Ayesha Batool, Saghir Ahmad, Abid Hussain Ch.
Abstract:
The art of the examination is probably the most difficult one in the whole range of educational practices. Semester system is the system of examination, which is set with an institute by its own teachers. Annual system is the system of examination, which is constructed and administrated by some agency outside the institute, it enables the teacher to estimate the effectiveness of the instruction, and students to estimate the progress made by them. On the other hand, semester system of examinations requires following the curriculum strictly and methods of teaching are to be employed by the choice of teachers. The main purpose of the study was to investigate university students’ perceptions regarding the effectiveness of semester system and annual system. The study was quantitative in nature. The sample consisted of 200 students. A five point Likert type scale was used to collect the data. The statistical measures like frequencies, mean, standard deviation, and One Way ANOVA test were applied to analyze the data. The major findings of the study indicated that in semester system students do not spend much time in political activities and develop their study habits. It also revealed that annual system of examination does not satisfy the educational aspirations of the students.
Keywords: Effectiveness, semester system, annual system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9421291 Numerical Studies on Thrust Vectoring Using Shock-Induced Self Impinging Secondary Jets
Authors: S. Vignesh, N. Vishnu, S. Vigneshwaran, M. Vishnu Anand, Dinesh Kumar Babu, V. R. Sanal Kumar
Abstract:
Numerical studies have been carried out using a validated two-dimensional standard k-omega turbulence model for the design optimization of a thrust vector control system using shock induced self-impinging supersonic secondary double jet. Parametric analytical studies have been carried out at different secondary injection locations to identifying the highest unsymmetrical distribution of the main gas flow due to shock waves, which produces a desirable side force more lucratively for vectoring. The results from the parametric studies of the case on hand reveal that the shock induced self-impinging supersonic secondary double jet is more efficient in certain locations at the divergent region of a CD nozzle than a case with supersonic single jet with same mass flow rate. We observed that the best axial location of the self-impinging supersonic secondary double jet nozzle with a given jet interaction angle, built-in to a CD nozzle having area ratio 1.797, is 0.991 times the primary nozzle throat diameter from the throat location. We also observed that the flexible steering is possible after invoking ON/OFF facility to the secondary nozzles for meeting the onboard mission requirements. Through our case studies we concluded that the supersonic self-impinging secondary double jet at predesigned jet interaction angle and location can provide more flexible steering options facilitating with 8.81% higher thrust vectoring efficiency than the conventional supersonic single secondary jet without compromising the payload capability of any supersonic aerospace vehicle.Keywords: Fluidic thrust vectoring, rocket steering, self-impinging secondary supersonic jet, TVC in aerospace vehicles.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26811290 Dynamic State Estimation with Optimal PMU and Conventional Measurements for Complete Observability
Authors: M. Ravindra, R. Srinivasa Rao
Abstract:
This paper presents a Generalized Binary Integer Linear Programming (GBILP) method for optimal allocation of Phasor Measurement Units (PMUs) and to generate Dynamic State Estimation (DSE) solution with complete observability. The GBILP method is formulated with Zero Injection Bus (ZIB) constraints to reduce the number of locations for placement of PMUs in the case of normal and single line contingency. The integration of PMU and conventional measurements is modeled in DSE process to estimate accurate states of the system. To estimate the dynamic behavior of the power system with proposed method, load change up to 40% considered at a bus in the power system network. The proposed DSE method is compared with traditional Weighted Least Squares (WLS) state estimation method in presence of load changes to show the impact of PMU measurements. MATLAB simulations are carried out on IEEE 14, 30, 57, and 118 bus systems to prove the validity of the proposed approach.
Keywords: Observability, phasor measurement units, PMU, state estimation, dynamic state estimation, SCADA measurements, zero injection bus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12651289 Spatial Query Localization Method in Limited Reference Point Environment
Authors: Victor Krebss
Abstract:
Task of object localization is one of the major challenges in creating intelligent transportation. Unfortunately, in densely built-up urban areas, localization based on GPS only produces a large error, or simply becomes impossible. New opportunities arise for the localization due to the rapidly emerging concept of a wireless ad-hoc network. Such network, allows estimating potential distance between these objects measuring received signal level and construct a graph of distances in which nodes are the localization objects, and edges - estimates of the distances between pairs of nodes. Due to the known coordinates of individual nodes (anchors), it is possible to determine the location of all (or part) of the remaining nodes of the graph. Moreover, road map, available in digital format can provide localization routines with valuable additional information to narrow node location search. However, despite abundance of well-known algorithms for solving the problem of localization and significant research efforts, there are still many issues that currently are addressed only partially. In this paper, we propose localization approach based on the graph mapped distances on the digital road map data basis. In fact, problem is reduced to distance graph embedding into the graph representing area geo location data. It makes possible to localize objects, in some cases even if only one reference point is available. We propose simple embedding algorithm and sample implementation as spatial queries over sensor network data stored in spatial database, allowing employing effectively spatial indexing, optimized spatial search routines and geometry functions.Keywords: Intelligent Transportation System, Sensor Network, Localization, Spatial Query, GIS, Graph Embedding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15371288 Disturbances of the Normal Operation of Kosovo Power System Regarding Atmospheric Discharges
Authors: B. Prebreza, I. Krasniqi, G. Kabashi, G. Pula, N. Avdiu
Abstract:
This paper discusses aspects of outages in the electric transmission network in the Kosovo Power System caused by the atmospheric discharges.
Frequency and location of the atmospheric discharges in Kosovo territory will be provided by a lightning location system ALARM (Automated Lightning Alert and Risk Management) and from the data from the Meteorological Department in Prishtina International Airport. These data will be used to make comparisons with the actual outages registered in the Kosovo Power System from the Kosovo Transmission, systems and market operator (KOSTT) during a specific time period.
The lines with the worst performance determined, regarding the atmospheric discharges, will be choose for further discussions in terms of over voltages caused by the direct or indirect lightning strokes.
Recommendations for protection in terms of insulator coordination and surge arresters will be given at the end and in this stage dynamic simulation will take part.
Keywords: Atmospheric discharges, dynamic simulations, Kosovo Power System, surge arresters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18391287 Evaluation of Model Evaluation Criterion for Software Development Effort Estimation
Authors: S. K. Pillai, M. K. Jeyakumar
Abstract:
Estimation of model parameters is necessary to predict the behavior of a system. Model parameters are estimated using optimization criteria. Most algorithms use historical data to estimate model parameters. The known target values (actual) and the output produced by the model are compared. The differences between the two form the basis to estimate the parameters. In order to compare different models developed using the same data different criteria are used. The data obtained for short scale projects are used here. We consider software effort estimation problem using radial basis function network. The accuracy comparison is made using various existing criteria for one and two predictors. Then, we propose a new criterion based on linear least squares for evaluation and compared the results of one and two predictors. We have considered another data set and evaluated prediction accuracy using the new criterion. The new criterion is easy to comprehend compared to single statistic. Although software effort estimation is considered, this method is applicable for any modeling and prediction.
Keywords: Software effort estimation, accuracy, Radial Basis Function, linear least squares.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20431286 The Effect of CPU Location in Total Immersion of Microelectronics
Authors: A. Almaneea, N. Kapur, J. L. Summers, H. M. Thompson
Abstract:
Meeting the growth in demand for digital services such as social media, telecommunications, and business and cloud services requires large scale data centres, which has led to an increase in their end use energy demand. Generally, over 30% of data centre power is consumed by the necessary cooling overhead. Thus energy can be reduced by improving the cooling efficiency. Air and liquid can both be used as cooling media for the data centre. Traditional data centre cooling systems use air, however liquid is recognised as a promising method that can handle the more densely packed data centres. Liquid cooling can be classified into three methods; rack heat exchanger, on-chip heat exchanger and full immersion of the microelectronics. This study quantifies the improvements of heat transfer specifically for the case of immersed microelectronics by varying the CPU and heat sink location. Immersion of the server is achieved by filling the gap between the microelectronics and a water jacket with a dielectric liquid which convects the heat from the CPU to the water jacket on the opposite side. Heat transfer is governed by two physical mechanisms, which is natural convection for the fixed enclosure filled with dielectric liquid and forced convection for the water that is pumped through the water jacket. The model in this study is validated with published numerical and experimental work and shows good agreement with previous work. The results show that the heat transfer performance and Nusselt number (Nu) is improved by 89% by placing the CPU and heat sink on the bottom of the microelectronics enclosure.
Keywords: CPU location, data centre cooling, heat sink in enclosures, Immersed microelectronics, turbulent natural convection in enclosures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21751285 Spatially Random Sampling for Retail Food Risk Factors Study
Authors: Guilan Huang
Abstract:
In 2013 and 2014, the U.S. Food and Drug Administration (FDA) collected data from selected fast food restaurants and full service restaurants for tracking changes in the occurrence of foodborne illness risk factors. This paper discussed how we customized spatial random sampling method by considering financial position and availability of FDA resources, and how we enriched restaurants data with location. Location information of restaurants provides opportunity for quantitatively determining random sampling within non-government units (e.g.: 240 kilometers around each data-collector). Spatial analysis also could optimize data-collectors’ work plans and resource allocation. Spatial analytic and processing platform helped us handling the spatial random sampling challenges. Our method fits in FDA’s ability to pinpoint features of foodservice establishments, and reduced both time and expense on data collection.
Keywords: Geospatial technology, restaurant, retail food risk factors study, spatial random sampling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14661284 BasWilCalc – Basket Willow (Salix viminalis) Biomass Yield Calculator
Authors: Wiesław Szulczewski, Wojciech Jakubowski, Andrzej Żyromski, Małgorzata Biniak-Pieróg
Abstract:
The aim of the paper was to elaborate a novel calculator BasWilCalc, that allows to estimate the actual amount of biomass on the basket willow plantations. The proposed method is based on the results of field experiment conducted during years 2011-2013 on basket willow plantation in the south-western part of Poland. As input data the results of destructive measurements of the diameter, length and weight of willow stems and non-destructive biometric measurements of diameter in the middle of stems and their length during the growing season performed at weekly intervals were used. Performed analysis enabled to develop the algorithm which, due to the fact that energy plantations are of known and constant planting structure, allows to estimate the actual amount of willow basket biomass on the plantation with a given probability and accuracy specified by the model, based on the number of stems measured and the age of the plantation.
Keywords: Basket willow (Salix viminalis) biomass, biometric measurements, yield, biomass calculator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16671283 Totally Integrated Smart Energy System through Data Acquisition via Remote Location
Authors: Muhammad Tahir Qadri, M. Irfan Anis, M. Nawaz Irshad Khan
Abstract:
This paper discusses the approach of real-time controlling of the energy management system using the data acquisition tool of LabVIEW. The main idea of this inspiration was to interface the Station (PC) with the system and publish the data on internet using LabVIEW. In this venture, controlling and switching of 3 phase AC loads are effectively and efficiently done. The phases are also sensed through devices. In case of any failure the attached generator starts functioning automatically. The computer sends command to the system and system respond to the request. The modern feature is to access and control the system world-wide using world wide web (internet). This controlling can be done at any time from anywhere to effectively use the energy especially in developing countries where energy management is a big problem. In this system totally integrated devices are used to operate via remote location.Keywords: VI-server, Remote Access, Telemetry, Data Acquisition, web server.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18791282 Bayesian Networks for Earthquake Magnitude Classification in a Early Warning System
Authors: G. Zazzaro, F.M. Pisano, G. Romano
Abstract:
During last decades, worldwide researchers dedicated efforts to develop machine-based seismic Early Warning systems, aiming at reducing the huge human losses and economic damages. The elaboration time of seismic waveforms is to be reduced in order to increase the time interval available for the activation of safety measures. This paper suggests a Data Mining model able to correctly and quickly estimate dangerousness of the running seismic event. Several thousand seismic recordings of Japanese and Italian earthquakes were analyzed and a model was obtained by means of a Bayesian Network (BN), which was tested just over the first recordings of seismic events in order to reduce the decision time and the test results were very satisfactory. The model was integrated within an Early Warning System prototype able to collect and elaborate data from a seismic sensor network, estimate the dangerousness of the running earthquake and take the decision of activating the warning promptly.Keywords: Bayesian Networks, Decision Support System, Magnitude Classification, Seismic Early Warning System
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35981281 Optimal Allocation of FACTS Devices for ATC Enhancement Using Bees Algorithm
Authors: R.Mohamad Idris, A.Khairuddin, M.W.Mustafa
Abstract:
In this paper, a novel method using Bees Algorithm is proposed to determine the optimal allocation of FACTS devices for maximizing the Available Transfer Capability (ATC) of power transactions between source and sink areas in the deregulated power system. The algorithm simultaneously searches the FACTS location, FACTS parameters and FACTS types. Two types of FACTS are simulated in this study namely Thyristor Controlled Series Compensator (TCSC) and Static Var Compensator (SVC). A Repeated Power Flow with FACTS devices including ATC is used to evaluate the feasible ATC value within real and reactive power generation limits, line thermal limits, voltage limits and FACTS operation limits. An IEEE30 bus system is used to demonstrate the effectiveness of the algorithm as an optimization tool to enhance ATC. A Genetic Algorithm technique is used for validation purposes. The results clearly indicate that the introduction of FACTS devices in a right combination of location and parameters could enhance ATC and Bees Algorithm can be efficiently used for this kind of nonlinear integer optimization.
Keywords: ATC, Bees Algorithm, TCSC, SVC
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31681280 Soybean Based Farming System Assessment in Pasuruan East Java Indonesia
Authors: Mohammad Saeri, Noor Rizkiyah, Kambang Vetrani Asie, Titin Apung Atikah
Abstract:
The study aims to assess efficient specific-location soybean farming technology assembly by assisting the farmers in applying the suggested technology. Superimposed trial was conducted to know NPK fertilizer effect toward soybean growth and yield and soybean improved variety test for the dissemination of improved variety. The assessment was conducted at the farmers group of Sumber Rejeki, Kepulungan Village, Gempol Sub-district, Pasuruan Regency as the soybean central at Pasuruan area. The number of farmers involved in the study was 38 people with 25 ha soybean area. This study was held from July to October 2012. The recommended technology package agreed at the socialization time and used in this research were: using Argomulyo variety seeds of 40 kg/ha, planting by drilling, planting by distance of 40x10 cm, deciding the seeds amount of 2-3 seeds per hole, and giving fertilization based on recommendation of East Java AIAT of 50 kg Urea, 100 kg SP-36 and 50 kg KCl. Farmers around the research location were used as control group. Assessment on soybean farming system was considered effective because it could increase the production up to 38%. The farming analysis showed that the result collaborator farmers gained were positively higher than non-collaborator farmers with RC ratio of 2.03 and 1.54, respectively. Argomulyo variety has the prospect to be developed due to the high yield of about 2 tons/ha and the larger seeds. The NPK fertilization test at the soybean plants showed that the fertilization had minor effect on the yield.
Keywords: Farming system, soybean, variety, location specific farming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10691279 Evaluation of the Accuracy of Time of Arrival Source Location Algorithm of Acoustic Emission in Concrete-Mortar Structure
Authors: Hisham A. Elfergani, Ayad A. Abdalla, Ahmed R. Ballil
Abstract:
Acoustic Emission (AE) is one of the most effective non-destructive tests that can be used to detect the defect process as it is occurring. AE techniques can be used to monitor a wide range of structures and materials such as metals, non-metals and combinations of these when load is applied. The current work investigates the effectiveness and accuracy of TOA method in AE tests involving reinforced composite concrete-mortar structures. A series of experimental tests were performed using the Hsu-Neilson (H-N) source to study 2-D location accuracy using this method on concrete-mortar (400×400 mm) specimens. Four AE sensors (R3I – resonant frequency 30 kHz) were mounted to the mortar surface and six sources were performed at each point of preselected locations on the upper surface of the mortar. Results show that the TOA method can be used effectively to locate signals on composite concrete/mortar specimen and has high accuracy.
Keywords: Acoustic emission, time of arrival, composite materials, reinforced concrete.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6381278 Could Thermal Oceanic Hotspot Increase Climate Changes Activities in North Tropical Atlantic: Example of the 2005 Caribbean Coral Bleaching Hotspot and Hurricane Katrina Interaction
Authors: J- L. Siméon
Abstract:
This paper reviews recent studies and particularly the effects of Climate Change in the North Tropical Atlantic by studying atmospheric conditions that prevailed in 2005 ; Coral Bleaching HotSpot and Hurricane Katrina. In the aim to better understand and estimate the impact of the physical phenomenon, i.e. Thermal Oceanic HotSpot (TOHS), isotopic studies of δ18O and δ13C on marine animals from Guadeloupe (French Caribbean Island) were carried out. Recorded measures show Sea Surface Temperature (SST) up to 35°C in August which is much higher than data recorded by NOAA satellites 32°C. After having reviewed the process that led to the creation of Hurricane Katrina which hit New Orleans in August 29, 2005, it will be shown that the climatic conditions in the Caribbean from August to October 2005 have influenced Katrina evolution. This TOHS is a combined effect of various phenomenon which represent an additional factor to estimate future climate changes.Keywords: Climate Change, Thermal Ocean HotSpot, Isotope, Hurricane, Connection, Uncertainty, Sea, Science.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16861277 Neural Networks and Particle Swarm Optimization Based MPPT for Small Wind Power Generator
Authors: Chun-Yao Lee, Yi-Xing Shen, Jung-Cheng Cheng, Yi-Yin Li, Chih-Wen Chang
Abstract:
This paper proposes the method combining artificial neural network (ANN) with particle swarm optimization (PSO) to implement the maximum power point tracking (MPPT) by controlling the rotor speed of the wind generator. First, the measurements of wind speed, rotor speed of wind power generator and output power of wind power generator are applied to train artificial neural network and to estimate the wind speed. Second, the method mentioned above is applied to estimate and control the optimal rotor speed of the wind turbine so as to output the maximum power. Finally, the result reveals that the control system discussed in this paper extracts the maximum output power of wind generator within the short duration even in the conditions of wind speed and load impedance variation.Keywords: Maximum power point tracking, artificial neuralnetwork, particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22091276 An Approach to Measure Snow Depth of Winter Accumulation at Basin Scale Using Satellite Data
Authors: M. Geetha Priya, D. Krishnaveni
Abstract:
Snow depth estimation and monitoring studies have been carried out for decades using empirical relationship or extrapolation of point measurements carried out in field. With the development of advanced satellite based remote sensing techniques, a modified approach is proposed in the present study to estimate the winter accumulated snow depth at basin scale. Assessment of snow depth by differencing Digital Elevation Model (DEM) generated at the beginning and end of winter season can be experimented for the region of interest (Himalayan and polar regions) accounting for winter accumulation (solid precipitation). The proposed approach is based on existing geodetic method that is being used for glacier mass balance estimation. Considering the satellite datasets purely acquired during beginning and end of winter season, it is possible to estimate the change in depth or thickness for the snow that is accumulated during the winter as it takes one year for the snow to get transformed into firn (snow that has survived one summer or one-year old snow).
Keywords: Digital elevation model, snow depth, geodetic method, snow cover.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7161275 Effects of Incident Angle and Distance on Visible Light Communication
Authors: Taegyoo Woo, Jong Kang Park, Jong Tae Kim
Abstract:
Visible Light Communication (VLC) provides wireless communication features in illumination systems. One of the key applications is to recognize the user location by indoor illuminators such as light emitting diodes. For localization of individual receivers in these systems, we usually assume that receivers and transmitters are placed in parallel. However, it is difficult to satisfy this assumption because the receivers move randomly in real case. It is necessary to analyze the case when transmitter is not placed perfectly parallel to receiver. It is also important to identify changes on optical gain by the tilted angles and distances of them against the illuminators. In this paper, we simulate optical gain for various cases where the tilt of the receiver and the distance change. Then, we identified changing patterns of optical gains according to tilted angles of a receiver and distance. These results can help many VLC applications understand the extent of the location errors with regard to optical gains of the receivers and identify the root cause.Keywords: Visible light communication, optical channel, indoor positioning, Lambertian radiation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14641274 Modeling of Pulsatile Blood Flow in a Weak Magnetic Field
Authors: Chee Teck Phua, Gaëlle Lissorgues
Abstract:
Blood pulse is an important human physiological signal commonly used for the understanding of the individual physical health. Current methods of non-invasive blood pulse sensing require direct contact or access to the human skin. As such, the performances of these devices tend to vary with time and are subjective to human body fluids (e.g. blood, perspiration and skin-oil) and environmental contaminants (e.g. mud, water, etc). This paper proposes a simulation model for the novel method of non-invasive acquisition of blood pulse using the disturbance created by blood flowing through a localized magnetic field. The simulation model geometry represents a blood vessel, a permanent magnet, a magnetic sensor, surrounding tissues and air in 2-dimensional. In this model, the velocity and pressure fields in the blood stream are described based on Navier-Stroke equations and the walls of the blood vessel are assumed to have no-slip condition. The blood assumes a parabolic profile considering a laminar flow for blood in major artery near the skin. And the inlet velocity follows a sinusoidal equation. This will allow the computational software to compute the interactions between the magnetic vector potential generated by the permanent magnet and the magnetic nanoparticles in the blood. These interactions are simulated based on Maxwell equations at the location where the magnetic sensor is placed. The simulated magnetic field at the sensor location is found to assume similar sinusoidal waveform characteristics as the inlet velocity of the blood. The amplitude of the simulated waveforms at the sensor location are compared with physical measurements on human subjects and found to be highly correlated.
Keywords: Blood pulse, magnetic sensing, non-invasive measurement, magnetic disturbance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26151273 Image Haze Removal Using Scene Depth Based Spatially Varying Atmospheric Light in Haar Lifting Wavelet Domain
Authors: Prabh Preet Singh, Harpreet Kaur
Abstract:
This paper presents a method for single image dehazing based on dark channel prior (DCP). The property that the intensity of the dark channel gives an approximate thickness of the haze is used to estimate the transmission and atmospheric light. Instead of constant atmospheric light, the proposed method employs scene depth to estimate spatially varying atmospheric light as it truly occurs in nature. Haze imaging model together with the soft matting method has been used in this work to produce high quality haze free image. Experimental results demonstrate that the proposed approach produces better results than the classic DCP approach as color fidelity and contrast of haze free image are improved and no over-saturation in the sky region is observed. Further, lifting Haar wavelet transform is employed to reduce overall execution time by a factor of two to three as compared to the conventional approach.
Keywords: Depth based atmospheric light, dark channel prior, lifting wavelet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5541272 Optimal Distributed Generator Sizing and Placement by Analytical Method and PSO Algorithm Considering Optimal Reactive Power Dispatch
Authors: Kyaw Myo Lin, Pyone Lai Swe, Khine Zin Oo
Abstract:
In this paper, an approach combining analytical method for the distributed generator (DG) sizing and meta-heuristic search for the optimal location of DG has been presented. The optimal size of DG on each bus is estimated by the loss sensitivity factor method while the optimal sites are determined by Particle Swarm Optimization (PSO) based optimal reactive power dispatch for minimizing active power loss. To confirm the proposed approach, it has been tested on IEEE-30 bus test system. The adjustments of operating constraints and voltage profile improvements have also been observed. The obtained results show that the allocation of DGs results in a significant loss reduction with good voltage profiles and the combined approach is competent in keeping the system voltages within the acceptable limits.
Keywords: Analytical approach, distributed generations, optimal size, optimal location, optimal reactive power dispatch, particle swarm optimization algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11761271 A Study on the Location and Range of Obstacle Region in Robot's Point Placement Task based on the Vision Control Algorithm
Authors: Jae Kyung Son, Wan Shik Jang, Sung hyun Shim, Yoon Gyung Sung
Abstract:
This paper is concerned with the application of the vision control algorithm for robot's point placement task in discontinuous trajectory caused by obstacle. The presented vision control algorithm consists of four models, which are the robot kinematic model, vision system model, parameters estimation model, and robot joint angle estimation model.When the robot moves toward a target along discontinuous trajectory, several types of obstacles appear in two obstacle regions. Then, this study is to investigate how these changes will affect the presented vision control algorithm.Thus, the practicality of the vision control algorithm is demonstrated experimentally by performing the robot's point placement task in discontinuous trajectory by obstacle.
Keywords: Vision control algorithm, location of obstacle region, range of obstacle region, point placement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14021270 Conjugate Heat Transfer Analysis of a Combustion Chamber using ANSYS Computational Fluid Dynamics to Estimate the Thermocouple Positioning in a Chamber Wall
Authors: Muzna Tariq, Ihtzaz Qamar
Abstract:
In most engineering cases, the working temperatures inside a combustion chamber are high enough that they lie beyond the operational range of thermocouples. Furthermore, design and manufacturing limitations restrict the use of internal thermocouples in many applications. Heat transfer inside a combustion chamber is caused due to interaction of the post-combustion hot fluid with the chamber wall. Heat transfer that involves an interaction between the fluid and solid is categorized as Conjugate Heat Transfer (CHT). Therefore, to satisfy the needs of CHT, CHT Analysis is performed by using ANSYS CFD tool to estimate theoretically precise thermocouple positions at the combustion chamber wall where excessive temperatures (beyond thermocouple range) can be avoided. In accordance with these Computational Fluid Dynamics (CFD) results, a combustion chamber is designed, and a prototype is manufactured with multiple thermocouple ports positioned at the specified distances so that the temperature of hot gases can be measured on the chamber wall where the temperatures do not exceed the thermocouple working range.
Keywords: Computational Fluid Dynamics, CFD, conduction, conjugate heat transfer, CHT, convection, fluid flow, thermocouples.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6911269 Using Linear Quadratic Gaussian Optimal Control for Lateral Motion of Aircraft
Authors: A. Maddi, A. Guessoum, D. Berkani
Abstract:
The purpose of this paper is to provide a practical example to the Linear Quadratic Gaussian (LQG) controller. This method includes a description and some discussion of the discrete Kalman state estimator. One aspect of this optimality is that the estimator incorporates all information that can be provided to it. It processes all available measurements, regardless of their precision, to estimate the current value of the variables of interest, with use of knowledge of the system and measurement device dynamics, the statistical description of the system noises, measurement errors, and uncertainty in the dynamics models. Since the time of its introduction, the Kalman filter has been the subject of extensive research and application, particularly in the area of autonomous or assisted navigation. For example, to determine the velocity of an aircraft or sideslip angle, one could use a Doppler radar, the velocity indications of an inertial navigation system, or the relative wind information in the air data system. Rather than ignore any of these outputs, a Kalman filter could be built to combine all of this data and knowledge of the various systems- dynamics to generate an overall best estimate of velocity and sideslip angle.Keywords: Aircraft motion, Kalman filter, LQG control, Lateral stability, State estimator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24711268 Congestion Management in a Deregulated Power System with Micro Grid
Authors: Guguloth Ramesh, T. K. Sunil Kumar
Abstract:
This paper presents congestion management in deregulated power systems. In a deregulated environment, every buyer wants to buy power from the cheapest generator available, irrespective of relative geographical location of buyer and seller. As a consequence of this, the transmission corridors evacuating the power of cheaper generators would get overloaded if all such transactions are approved. Congestion management is a mechanism to prioritize the transactions and commit to such a schedule which would not overload the network. The congestions in the transmission lines are determined by Optimal Power Flow (OPF) solution, which is carried by primal liner programming method. Congestion in the transmission lines are alleviated by connected Distributed Generation (DG) of micro grid at load bus. A method to determine the optimal location of DG unit has been suggested based on transmission line relief sensitivity based approach. The effectiveness of proposed method has been demonstrated on modified IEEE-14 and 30 bus test systems.
Keywords: Congestion management, Distribution Generation (DG), Transmission Line Relief (TLR) sensitivity index, OPF.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 38941267 Text Mining Technique for Data Mining Application
Authors: M. Govindarajan
Abstract:
Text Mining is around applying knowledge discovery techniques to unstructured text is termed knowledge discovery in text (KDT), or Text data mining or Text Mining. In decision tree approach is most useful in classification problem. With this technique, tree is constructed to model the classification process. There are two basic steps in the technique: building the tree and applying the tree to the database. This paper describes a proposed C5.0 classifier that performs rulesets, cross validation and boosting for original C5.0 in order to reduce the optimization of error ratio. The feasibility and the benefits of the proposed approach are demonstrated by means of medial data set like hypothyroid. It is shown that, the performance of a classifier on the training cases from which it was constructed gives a poor estimate by sampling or using a separate test file, either way, the classifier is evaluated on cases that were not used to build and evaluate the classifier are both are large. If the cases in hypothyroid.data and hypothyroid.test were to be shuffled and divided into a new 2772 case training set and a 1000 case test set, C5.0 might construct a different classifier with a lower or higher error rate on the test cases. An important feature of see5 is its ability to classifiers called rulesets. The ruleset has an error rate 0.5 % on the test cases. The standard errors of the means provide an estimate of the variability of results. One way to get a more reliable estimate of predictive is by f-fold –cross- validation. The error rate of a classifier produced from all the cases is estimated as the ratio of the total number of errors on the hold-out cases to the total number of cases. The Boost option with x trials instructs See5 to construct up to x classifiers in this manner. Trials over numerous datasets, large and small, show that on average 10-classifier boosting reduces the error rate for test cases by about 25%.Keywords: C5.0, Error Ratio, text mining, training data, test data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2489