Search results for: particle size distribution
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3911

Search results for: particle size distribution

3071 Investigation into the Optimum Hydraulic Loading Rate for Selected Filter Media Packed in a Continuous Upflow Filter

Authors: A. Alzeyadi, E. Loffill, R. Alkhaddar

Abstract:

Continuous upflow filters can combine the nutrient (nitrogen and phosphate) and suspended solid removal in one unit process. The contaminant removal could be achieved chemically or biologically; in both processes the filter removal efficiency depends on the interaction between the packed filter media and the influent. In this paper a residence time distribution (RTD) study was carried out to understand and compare the transfer behaviour of contaminants through a selected filter media packed in a laboratory-scale continuous up flow filter; the selected filter media are limestone and white dolomite. The experimental work was conducted by injecting a tracer (red drain dye tracer –RDD) into the filtration system and then measuring the tracer concentration at the outflow as a function of time; the tracer injection was applied at hydraulic loading rates (HLRs) (3.8 to 15.2 m h-1). The results were analysed according to the cumulative distribution function F(t) to estimate the residence time of the tracer molecules inside the filter media. The mean residence time (MRT) and variance σ2 are two moments of RTD that were calculated to compare the RTD characteristics of limestone with white dolomite. The results showed that the exit-age distribution of the tracer looks better at HLRs (3.8 to 7.6 m h-1) and (3.8 m h-1) for limestone and white dolomite respectively. At these HLRs the cumulative distribution function F(t) revealed that the residence time of the tracer inside the limestone was longer than in the white dolomite; whereas all the tracer took 8 minutes to leave the white dolomite at 3.8 m h-1. On the other hand, the same amount of the tracer took 10 minutes to leave the limestone at the same HLR. In conclusion, the determination of the optimal level of hydraulic loading rate, which achieved the better influent distribution over the filtration system, helps to identify the applicability of the material as filter media. Further work will be applied to examine the efficiency of the limestone and white dolomite for phosphate removal by pumping a phosphate solution into the filter at HLRs (3.8 to 7.6 m h-1).

Keywords: Filter media, hydraulic loading rate, residence time distribution, tracer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1871
3070 Determining the Best Fitting Distributions for Minimum Flows of Streams in Gediz Basin

Authors: Naci Büyükkaracığan

Abstract:

Today, the need for water sources is swiftly increasing due to population growth. At the same time, it is known that some regions will face with shortage of water and drought because of the global warming and climate change. In this context, evaluation and analysis of hydrological data such as the observed trends, drought and flood prediction of short term flow has great deal of importance. The most accurate selection probability distribution is important to describe the low flow statistics for the studies related to drought analysis. As in many basins In Turkey, Gediz River basin will be affected enough by the drought and will decrease the amount of used water. The aim of this study is to derive appropriate probability distributions for frequency analysis of annual minimum flows at 6 gauging stations of the Gediz Basin. After applying 10 different probability distributions, six different parameter estimation methods and 3 fitness test, the Pearson 3 distribution and general extreme values distributions were found to give optimal results.

Keywords: Gediz Basin, goodness-of-fit tests, Minimum flows, probability distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2505
3069 Unconditionally Secure Quantum Payment System

Authors: Essam Al-Daoud

Abstract:

A potentially serious problem with current payment systems is that their underlying hard problems from number theory may be solved by either a quantum computer or unanticipated future advances in algorithms and hardware. A new quantum payment system is proposed in this paper. The suggested system makes use of fundamental principles of quantum mechanics to ensure the unconditional security without prior arrangements between customers and vendors. More specifically, the new system uses Greenberger-Home-Zeilinger (GHZ) states and Quantum Key Distribution to authenticate the vendors and guarantee the transaction integrity.

Keywords: Bell state, GHZ state, Quantum key distribution, Quantum payment system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1550
3068 Surfactant-Free O/W-Emulsion as Drug Delivery System

Authors: M. Kumpugdee-Vollrath, J.-P. Krause, S. Bürk

Abstract:

Most of the drugs used for pharmaceutical purposes are poorly water-soluble drugs. About 40% of all newly discovered drugs are lipophilic and the numbers of lipophilic drugs seem to increase more and more. Drug delivery systems such as nanoparticles, micelles or liposomes are applied to improve their solubility and thus their bioavailability. Besides various techniques of solubilization, oil-in-water emulsions are often used to incorporate lipophilic drugs into the oil phase. To stabilize emulsions surface active substances (surfactants) are generally used. An alternative method to avoid the application of surfactants was of great interest. One possibility is to develop O/W-emulsion without any addition of surface active agents or the so called “surfactant-free emulsion or SFE”. The aim of this study was to develop and characterize SFE as a drug carrier by varying the production conditions. Lidocaine base was used as a model drug. The injection method was developed. Effects of ultrasound as well as of temperature on the properties of the emulsion were studied. Particle sizes and release were determined. The long-term stability up to 30 days was performed. The results showed that the surfactant-free O/W emulsions with pharmaceutical oil as drug carrier can be produced.

Keywords: Emulsion, lidocaine, Miglyol, size, surfactant, light scattering, release, injection, ultrasound, stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3311
3067 A Novel Probablistic Strategy for Modeling Photovoltaic Based Distributed Generators

Authors: Engy A. Mohamed, Yasser G. Hegazy

Abstract:

This paper presents a novel algorithm for modeling photovoltaic based distributed generators for the purpose of optimal planning of distribution networks. The proposed algorithm utilizes sequential Monte Carlo method in order to accurately consider the stochastic nature of photovoltaic based distributed generators. The proposed algorithm is implemented in MATLAB environment and the results obtained are presented and discussed.

Keywords: Comulative distribution function, distributed generation, Monte Carlo.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2483
3066 Application of Single Tuned Passive Filters in Distribution Networks at the Point of Common Coupling

Authors: M. Almutairi, S. Hadjiloucas

Abstract:

The harmonic distortion of voltage is important in relation to power quality due to the interaction between the large diffusion of non-linear and time-varying single-phase and three-phase loads with power supply systems. However, harmonic distortion levels can be reduced by improving the design of polluting loads or by applying arrangements and adding filters. The application of passive filters is an effective solution that can be used to achieve harmonic mitigation mainly because filters offer high efficiency, simplicity, and are economical. Additionally, possible different frequency response characteristics can work to achieve certain required harmonic filtering targets. With these ideas in mind, the objective of this paper is to determine what size single tuned passive filters work in distribution networks best, in order to economically limit violations caused at a given point of common coupling (PCC). This article suggests that a single tuned passive filter could be employed in typical industrial power systems. Furthermore, constrained optimization can be used to find the optimal sizing of the passive filter in order to reduce both harmonic voltage and harmonic currents in the power system to an acceptable level, and, thus, improve the load power factor. The optimization technique works to minimize voltage total harmonic distortions (VTHD) and current total harmonic distortions (ITHD), where maintaining a given power factor at a specified range is desired. According to the IEEE Standard 519, both indices are viewed as constraints for the optimal passive filter design problem. The performance of this technique will be discussed using numerical examples taken from previous publications.

Keywords: Harmonics, passive filter, power factor, power quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2191
3065 Split-Pipe Design of Water Distribution Network Using Simulated Annealing

Authors: J. Tospornsampan, I. Kita, M. Ishii, Y. Kitamura

Abstract:

In this paper a procedure for the split-pipe design of looped water distribution network based on the use of simulated annealing is proposed. Simulated annealing is a heuristic-based search algorithm, motivated by an analogy of physical annealing in solids. It is capable for solving the combinatorial optimization problem. In contrast to the split-pipe design that is derived from a continuous diameter design that has been implemented in conventional optimization techniques, the split-pipe design proposed in this paper is derived from a discrete diameter design where a set of pipe diameters is chosen directly from a specified set of commercial pipes. The optimality and feasibility of the solutions are found to be guaranteed by using the proposed method. The performance of the proposed procedure is demonstrated through solving the three well-known problems of water distribution network taken from the literature. Simulated annealing provides very promising solutions and the lowest-cost solutions are found for all of these test problems. The results obtained from these applications show that simulated annealing is able to handle a combinatorial optimization problem of the least cost design of water distribution network. The technique can be considered as an alternative tool for similar areas of research. Further applications and improvements of the technique are expected as well.

Keywords: Combinatorial problem, Heuristics, Least-cost design, Looped network, Pipe network, Optimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2678
3064 The Estimation Method of Stress Distribution for Beam Structures Using the Terrestrial Laser Scanning

Authors: Sang Wook Park, Jun Su Park, Byung Kwan Oh, Yousok Kim, Hyo Seon Park

Abstract:

This study suggests the estimation method of stress distribution for the beam structures based on TLS (Terrestrial Laser Scanning). The main components of method are the creation of the lattices of raw data from TLS to satisfy the suitable condition and application of CSSI (Cubic Smoothing Spline Interpolation) for estimating stress distribution. Estimation of stress distribution for the structural member or the whole structure is one of the important factors for safety evaluation of the structure. Existing sensors which include ESG (Electric strain gauge) and LVDT (Linear Variable Differential Transformer) can be categorized as contact type sensor which should be installed on the structural members and also there are various limitations such as the need of separate space where the network cables are installed and the difficulty of access for sensor installation in real buildings. To overcome these problems inherent in the contact type sensors, TLS system of LiDAR (light detection and ranging), which can measure the displacement of a target in a long range without the influence of surrounding environment and also get the whole shape of the structure, has been applied to the field of structural health monitoring. The important characteristic of TLS measuring is a formation of point clouds which has many points including the local coordinate. Point clouds are not linear distribution but dispersed shape. Thus, to analyze point clouds, the interpolation is needed vitally. Through formation of averaged lattices and CSSI for the raw data, the method which can estimate the displacement of simple beam was developed. Also, the developed method can be extended to calculate the strain and finally applicable to estimate a stress distribution of a structural member. To verify the validity of the method, the loading test on a simple beam was conducted and TLS measured it. Through a comparison of the estimated stress and reference stress, the validity of the method is confirmed.

Keywords: Structural health monitoring, terrestrial laser scanning, estimation of stress distribution, coordinate transformation, cubic smoothing spline interpolation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2743
3063 Minimizing Risk Costs through Optimal Responses in NPD Projects

Authors: Chan-Sik Kim, Jong-Seong Kim, Se Won Lee, Hoo-Gon Choi

Abstract:

In rapidly changing market environment, firms are investing a lot of time and resources into new product development (NPD) projects to make profit and to obtain competitive advantage. However, failure rate of NPD projects is becoming high due to various internal and external risks which hinder successful NPD projects. To reduce the failure rate, it is critical that risks have to be managed effectively and efficiently through good strategy, and treated by optimal responses to minimize risk cost. Four strategies are adopted to handle the risks in this study. The optimal responses are characterized by high reduction of risk costs with high efficiency. This study suggests a framework to decide the optimal responses considering the core risks, risk costs, response efficiency and response costs for successful NPD projects. Both binary particles warm optimization (BPSO) and multi-objective particle swarm optimization (MOPSO) methods are mainly used in the framework. Although several limitations exist in use for real industries, the frame work shows good strength for handling the risks with highly scientific ways through an example.

Keywords: NPD projects, risk cost, strategy, optimal responses, Particle Swarm Optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1957
3062 Using Artificial Neural Network to Forecast Groundwater Depth in Union County Well

Authors: Zahra Ghadampour, Gholamreza Rakhshandehroo

Abstract:

A concern that researchers usually face in different applications of Artificial Neural Network (ANN) is determination of the size of effective domain in time series. In this paper, trial and error method was used on groundwater depth time series to determine the size of effective domain in the series in an observation well in Union County, New Jersey, U.S. different domains of 20, 40, 60, 80, 100, and 120 preceding day were examined and the 80 days was considered as effective length of the domain. Data sets in different domains were fed to a Feed Forward Back Propagation ANN with one hidden layer and the groundwater depths were forecasted. Root Mean Square Error (RMSE) and the correlation factor (R2) of estimated and observed groundwater depths for all domains were determined. In general, groundwater depth forecast improved, as evidenced by lower RMSEs and higher R2s, when the domain length increased from 20 to 120. However, 80 days was selected as the effective domain because the improvement was less than 1% beyond that. Forecasted ground water depths utilizing measured daily data (set #1) and data averaged over the effective domain (set #2) were compared. It was postulated that more accurate nature of measured daily data was the reason for a better forecast with lower RMSE (0.1027 m compared to 0.255 m) in set #1. However, the size of input data in this set was 80 times the size of input data in set #2; a factor that may increase the computational effort unpredictably. It was concluded that 80 daily data may be successfully utilized to lower the size of input data sets considerably, while maintaining the effective information in the data set.

Keywords: Neural networks, groundwater depth, forecast.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2516
3061 Isotropic Stress Distribution in Cu/(001) Fe Two Sheets

Authors: A. Derardja, L. Baroura, M. Brioua

Abstract:

The nanotechnology based on epitaxial systems includes single or arranged misfit dislocations. In general, whatever is the type of dislocation or the geometry of the array formed by the dislocations; it is important for experimental studies to know exactly the stress distribution for which there is no analytical expression [1, 2]. This work, using a numerical analysis, deals with relaxation of epitaxial layers having at their interface a periodic network of edge misfit dislocations. The stress distribution is estimated by using isotropic elasticity. The results show that the thickness of the two sheets is a crucial parameter in the stress distributions and then in the profile of the two sheets. A comparative study between the case of single dislocation and the case of parallel network shows that the layers relaxed better when the interface is covered by a parallel arrangement of misfit. Consequently, a single dislocation at the interface produces an important stress field which can be reduced by inserting a parallel network of dislocations with suitable periodicity.

Keywords: Parallel array of misfit, interface, isotropic elasticity, single crystalline substrates, coherent interface

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1571
3060 High-Speed Particle Image Velocimetry of the Flow around a Moving Train Model with Boundary Layer Control Elements

Authors: Alexander Buhr, Klaus Ehrenfried

Abstract:

Trackside induced airflow velocities, also known as slipstream velocities, are an important criterion for the design of high-speed trains. The maximum permitted values are given by the Technical Specifications for Interoperability (TSI) and have to be checked in the approval process. For train manufactures it is of great interest to know in advance, how new train geometries would perform in TSI tests. The Reynolds number in moving model experiments is lower compared to full-scale. Especially the limited model length leads to a thinner boundary layer at the rear end. The hypothesis is that the boundary layer rolls up to characteristic flow structures in the train wake, in which the maximum flow velocities can be observed. The idea is to enlarge the boundary layer using roughness elements at the train model head so that the ratio between the boundary layer thickness and the car width at the rear end is comparable to a full-scale train. This may lead to similar flow structures in the wake and better prediction accuracy for TSI tests. In this case, the design of the roughness elements is limited by the moving model rig. Small rectangular roughness shapes are used to get a sufficient effect on the boundary layer, while the elements are robust enough to withstand the high accelerating and decelerating forces during the test runs. For this investigation, High-Speed Particle Image Velocimetry (HS-PIV) measurements on an ICE3 train model have been realized in the moving model rig of the DLR in Göttingen, the so called tunnel simulation facility Göttingen (TSG). The flow velocities within the boundary layer are analysed in a plain parallel to the ground. The height of the plane corresponds to a test position in the EN standard (TSI). Three different shapes of roughness elements are tested. The boundary layer thickness and displacement thickness as well as the momentum thickness and the form factor are calculated along the train model. Conditional sampling is used to analyse the size and dynamics of the flow structures at the time of maximum velocity in the train wake behind the train. As expected, larger roughness elements increase the boundary layer thickness and lead to larger flow velocities in the boundary layer and in the wake flow structures. The boundary layer thickness, displacement thickness and momentum thickness are increased by using larger roughness especially when applied in the height close to the measuring plane. The roughness elements also cause high fluctuations in the form factors of the boundary layer. Behind the roughness elements, the form factors rapidly are approaching toward constant values. This indicates that the boundary layer, while growing slowly along the second half of the train model, has reached a state of equilibrium.

Keywords: Boundary layer, high-speed PIV, ICE3, moving train model, roughness elements.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1528
3059 Determination of the Optimum Size of Building Stone Blocks: Case Study of Delichai Travertine Mine

Authors: Hesam Sedaghat Nejad, Navid Hosseini, Arash Nikvar Hassani

Abstract:

Determination of the optimum block size with high profitability is one of the significant parameters in designation of the building stone mines. The aim of this study was to determine the optimum dimensions of building stone blocks in Delichai travertine mine of Damavand in Tehran province through combining the effective parameters proven in determination of the optimum dimensions in building stones such as the spacing of joints and gaps, extraction tools constraints with the help of modeling by Gemcom software. To this end, following simulation of the topography of the mine, the block model was prepared and then in order to use spacing joints and discontinuities as a limiting factor, the existing joints set was added to the model. Since only one almost horizontal joint set with a slope of 5 degrees was available, this factor was effective only in determining the optimum height of the block, and thus to determine the longitudinal and transverse optimum dimensions of the extracted block, the power of available loader in the mine was considered as the secondary limiting factor. According to the aforementioned factors, the optimal block size in this mine was measured as 3.4×4×7 meter.

Keywords: Building stone, optimum block size, Delichai Travertine Mine, loader power.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1244
3058 Tumble Flow Analysis in an Unfired Engine Using Particle Image Velocimetry

Authors: B. Murali Krishna, J. M. Mallikarjuna

Abstract:

This paper deals with the experimental investigations of the in-cylinder tumble flows in an unfired internal combustion engine with a flat piston at the engine speeds ranging from 400 to 1000 rev/min., and also with the dome and dome-cavity pistons at an engine speed of 1000 rev/min., using particle image velocimetry. From the two-dimensional in-cylinder flow measurements, tumble flow analysis is carried out in the combustion space on a vertical plane passing through cylinder axis. To analyze the tumble flows, ensemble average velocity vectors are used and to characterize it, tumble ratio is estimated. From the results, generally, we have found that tumble ratio varies mainly with crank angle position. Also, at the end of compression stroke, average turbulent kinetic energy is more at higher engine speeds. We have also found that, at 330 crank angle position, flat piston shows an improvement of about 85 and 23% in tumble ratio, and about 24 and 2.5% in average turbulent kinetic energy compared to dome and dome-cavity pistons respectively

Keywords: In-cylinder flow, Dome piston, Cavity, Tumble, PIV

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2281
3057 Inference of Stress-Strength Model for a Lomax Distribution

Authors: H. Panahi, S. Asadi

Abstract:

In this paper, the estimation of the stress-strength parameter R = P(Y < X), when X and Y are independent and both are Lomax distributions with the common scale parameters but different shape parameters is studied. The maximum likelihood estimator of R is derived. Assuming that the common scale parameter is known, the bayes estimator and exact confidence interval of R are discussed. Simulation study to investigate performance of the different proposed methods has been carried out.

Keywords: Stress-Strength model; maximum likelihoodestimator; Bayes estimator; Lomax distribution

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793
3056 Impact of Loading Conditions on the Emission- Economic Dispatch

Authors: M. R. Alrashidi, M. E. El-Hawary

Abstract:

Environmental awareness and the recent environmental policies have forced many electric utilities to restructure their operational practices to account for their emission impacts. One way to accomplish this is by reformulating the traditional economic dispatch problem such that emission effects are included in the mathematical model. This paper presents a Particle Swarm Optimization (PSO) algorithm to solve the Economic- Emission Dispatch problem (EED) which gained recent attention due to the deregulation of the power industry and strict environmental regulations. The problem is formulated as a multi-objective one with two competing functions, namely economic cost and emission functions, subject to different constraints. The inequality constraints considered are the generating unit capacity limits while the equality constraint is generation-demand balance. A novel equality constraint handling mechanism is proposed in this paper. PSO algorithm is tested on a 30-bus standard test system. Results obtained show that PSO algorithm has a great potential in handling multi-objective optimization problems and is capable of capturing Pareto optimal solution set under different loading conditions.

Keywords: Economic emission dispatch, economic cost dispatch, particle swarm, multi-objective optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1897
3055 Facial Emotion Recognition with Convolutional Neural Network Based Architecture

Authors: Koray U. Erbas

Abstract:

Neural networks are appealing for many applications since they are able to learn complex non-linear relationships between input and output data. As the number of neurons and layers in a neural network increase, it is possible to represent more complex relationships with automatically extracted features. Nowadays Deep Neural Networks (DNNs) are widely used in Computer Vision problems such as; classification, object detection, segmentation image editing etc. In this work, Facial Emotion Recognition task is performed by proposed Convolutional Neural Network (CNN)-based DNN architecture using FER2013 Dataset. Moreover, the effects of different hyperparameters (activation function, kernel size, initializer, batch size and network size) are investigated and ablation study results for Pooling Layer, Dropout and Batch Normalization are presented.

Keywords: Convolutional Neural Network, Deep Learning, Deep Learning Based FER, Facial Emotion Recognition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371
3054 Mining Correlated Bicluster from Web Usage Data Using Discrete Firefly Algorithm Based Biclustering Approach

Authors: K. Thangavel, R. Rathipriya

Abstract:

For the past one decade, biclustering has become popular data mining technique not only in the field of biological data analysis but also in other applications like text mining, market data analysis with high-dimensional two-way datasets. Biclustering clusters both rows and columns of a dataset simultaneously, as opposed to traditional clustering which clusters either rows or columns of a dataset. It retrieves subgroups of objects that are similar in one subgroup of variables and different in the remaining variables. Firefly Algorithm (FA) is a recently-proposed metaheuristic inspired by the collective behavior of fireflies. This paper provides a preliminary assessment of discrete version of FA (DFA) while coping with the task of mining coherent and large volume bicluster from web usage dataset. The experiments were conducted on two web usage datasets from public dataset repository whereby the performance of FA was compared with that exhibited by other population-based metaheuristic called binary Particle Swarm Optimization (PSO). The results achieved demonstrate the usefulness of DFA while tackling the biclustering problem.

Keywords: Biclustering, Binary Particle Swarm Optimization, Discrete Firefly Algorithm, Firefly Algorithm, Usage profile Web usage mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2133
3053 Capacitor Placement in Radial Distribution System for Loss Reduction Using Artificial Bee Colony Algorithm

Authors: R. Srinivasa Rao

Abstract:

This paper presents a new method which applies an artificial bee colony algorithm (ABC) for capacitor placement in distribution systems with an objective of improving the voltage profile and reduction of power loss. The ABC algorithm is a new population based meta heuristic approach inspired by intelligent foraging behavior of honeybee swarm. The advantage of ABC algorithm is that it does not require external parameters such as cross over rate and mutation rate as in case of genetic algorithm and differential evolution and it is hard to determine these parameters in prior. The other advantage is that the global search ability in the algorithm is implemented by introducing neighborhood source production mechanism which is a similar to mutation process. To demonstrate the validity of the proposed algorithm, computer simulations are carried out on 69-bus system and compared the results with the other approach available in the literature. The proposed method has outperformed the other methods in terms of the quality of solution and computational efficiency.

Keywords: Distribution system, Capacitor Placement, Loss reduction, Artificial Bee Colony Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2817
3052 Geospatial Network Analysis Using Particle Swarm Optimization

Authors: Varun Singh, Mainak Bandyopadhyay, Maharana Pratap Singh

Abstract:

The shortest path (SP) problem concerns with finding the shortest path from a specific origin to a specified destination in a given network while minimizing the total cost associated with the path. This problem has widespread applications. Important applications of the SP problem include vehicle routing in transportation systems particularly in the field of in-vehicle Route Guidance System (RGS) and traffic assignment problem (in transportation planning). Well known applications of evolutionary methods like Genetic Algorithms (GA), Ant Colony Optimization, Particle Swarm Optimization (PSO) have come up to solve complex optimization problems to overcome the shortcomings of existing shortest path analysis methods. It has been reported by various researchers that PSO performs better than other evolutionary optimization algorithms in terms of success rate and solution quality. Further Geographic Information Systems (GIS) have emerged as key information systems for geospatial data analysis and visualization. This research paper is focused towards the application of PSO for solving the shortest path problem between multiple points of interest (POI) based on spatial data of Allahabad City and traffic speed data collected using GPS. Geovisualization of results of analysis is carried out in GIS.

Keywords: GIS, Outliers, PSO, Traffic Data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2892
3051 Image Mapping with Cumulative Distribution Function for Quick Convergence of Counter Propagation Neural Networks in Image Compression

Authors: S. Anna Durai, E. Anna Saro

Abstract:

In general the images used for compression are of different types like dark image, high intensity image etc. When these images are compressed using Counter Propagation Neural Network, it takes longer time to converge. The reason for this is that the given image may contain a number of distinct gray levels with narrow difference with their neighborhood pixels. If the gray levels of the pixels in an image and their neighbors are mapped in such a way that the difference in the gray levels of the neighbor with the pixel is minimum, then compression ratio as well as the convergence of the network can be improved. To achieve this, a Cumulative Distribution Function is estimated for the image and it is used to map the image pixels. When the mapped image pixels are used the Counter Propagation Neural Network yield high compression ratio as well as it converges quickly.

Keywords: Correlation, Counter Propagation Neural Networks, Cummulative Distribution Function, Image compression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1670
3050 Comparison between Lift and Drag-Driven VAWT Concepts on Low-Wind Site AEO

Authors: Marco Raciti Castelli, Ernesto Benini

Abstract:

This work presents a comparison between the Annual Energy Output (AEO) of two commercial vertical-axis wind turbines (VAWTs) for a low-wind urban site: both a drag-driven and a liftdriven concepts are examined in order to be installed on top of the new Via dei Giustinelli building, Trieste (Italy). The power-curves, taken from the product specification sheets, have been matched to the wind characteristics of the selected installation site. The influence of rotor swept area and rated power on the performance of the two proposed wind turbines have been examined in detail, achieving a correlation between rotor swept area, electrical generator size and wind distribution, to be used as a guideline for the calculation of the AEO.

Keywords: Annual Energy Output, micro-generationtechnology, urban environment, Vertical-Axis Wind Turbine

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6028
3049 Nodal Load Profiles Estimation for Time Series Load Flow Using Independent Component Analysis

Authors: Mashitah Mohd Hussain, Salleh Serwan, Zuhaina Hj Zakaria

Abstract:

This paper presents a method to estimate load profile in a multiple power flow solutions for every minutes in 24 hours per day. A method to calculate multiple solutions of non linear profile is introduced. The Power System Simulation/Engineering (PSS®E) and python has been used to solve the load power flow. The result of this power flow solutions has been used to estimate the load profiles for each load at buses using Independent Component Analysis (ICA) without any knowledge of parameter and network topology of the systems. The proposed algorithm is tested with IEEE 69 test bus system represents for distribution part and the method of ICA has been programmed in MATLAB R2012b version. Simulation results and errors of estimations are discussed in this paper.

Keywords: Electrical Distribution System, Power Flow Solution, Distribution Network, Independent Component Analysis, Newton Raphson, Power System Simulation for Engineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2916
3048 Understanding the Discharge Activities in Transformer Oil under AC and DC Voltage Adopting UHF Technique

Authors: R. Sarathi, G. Koperundevi

Abstract:

Design of Converter transformer insulation is a major challenge. The insulation of these transformers is stressed by both AC and DC voltages. Particle contamination is one of the major problems in insulation structures, as they generate partial discharges leading it to major failure of insulation. Similarly corona discharges occur in transformer insulation. This partial discharge due to particle movement / corona formation in insulation structure under different voltage wave shapes, are different. In the present study, UHF technique is adopted to understand the discharge activity and could be realized that the characteristics of UHF signal generated under low and high fields are different. In the case of corona generated signal, the frequency content of the UHF sensor output lies in the range 0.3-1.2 GHz and is not much varied except for its increase in magnitude of discharge with the increase in applied voltage. It is realized that the current signal injected due to partial discharges/corona is about 4ns duration measured for first one half cycle. Wavelet technique is adopted in the present study. It allows one to identify the frequency content present in the signal at different instant of time. The STD-MRA analysis helps one to identify the frequency band in which the energy content of the UHF signal is maximum.

Keywords: Contamination, Insulation, Partial Discharges, Transformer oil, UHF sensors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3840
3047 Diversity Analysis of a Quinoa (Chenopodium quinoa Willd.) Germplasm during Two Seasons

Authors: M. Mhada, E. N. Jellen, S. E. Jacobsen, O. Benlhabib

Abstract:

The present work has been carried out to evaluate the diversity of a collection of 78 quinoa accessions developed through recurrent selection from Andean germplasm introduced to Morocco in the winter of 2000. Twenty-three quantitative and qualitative characters were used for the evaluation of genetic diversity and the relationship between the accessions, and also for the establishment of a core collection in Morocco. Important variation was found among the accessions in terms of plant morphology and growth behavior. Data analysis showed positive correlation of the plant height, the plant fresh and the dry weight with the grain yield, while days to flowering was found to be negatively correlated with grain yield. The first four PCs contributed 74.76% of the variability; the first PC showed significant variation with 42.86% of the total variation, PC2 with 15.37%, PC3 with 9.05% and PC4 contributed 7.49% of the total variation. Plant size, days to grain filling and days to maturity are correlated to the PC1; and seed size, inflorescence density and mildew resistance are correlated to the PC2. Hierarchical cluster analysis rearranged the 78 quinoa accessions into four main groups and ten sub-clusters. Clustering was found in associations with days to maturity and also with plant size and seed-size traits.

Keywords: Character association, Chenopodium quinoa, Diversity analysis, Morphotypic cluster, Multivariate analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2586
3046 Theory of Nanowire Radial p-n-Junction

Authors: Stepan Petrosyan, Ashkhen Yesayan, Suren Nersesyan

Abstract:

We have developed an analytic model for the radial pn-junction in a nanowire (NW) core-shell structure utilizing as a new building block in different semiconductor devices. The potential distribution through the p-n-junction is calculated and the analytical expressions are derived to compute the depletion region widths. We show that the widths of space charge layers, surrounding the core, are the functions of core radius, which is the manifestation of so called classical size effect. The relationship between the depletion layer width and the built-in potential in the asymptotes of infinitely large core radius transforms to square-root dependence specific for conventional planar p-n-junctions. The explicit equation is derived to compute the capacitance of radial p-n-junction. The current-voltage behavior is also carefully determined taking into account the “short base" effects.

Keywords: Snanowire, p-n- junction, barrier capacitance, high injection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2787
3045 Resilient Modulus and Deformation Responses of Waste Glass in Flexible Pavement System

Authors: M. Al-Saedi, A. Chegenizadeh, H. Nikraz

Abstract:

Experimental investigations are conducted to assess a layered structure of glass (G) - rock (R) blends under the impact of repeated loading. Laboratory tests included sieve analyses, modified compaction test and repeated load triaxial test (RLTT) is conducted on different structures of stratified GR samples to reach the objectives of this study. Waste materials are such essential components in the climate system, and also commonly used in minimising the need for natural materials in many countries. Glass is one of the most widely used groups of waste materials which have been extensively using in road applications. Full range particle size and colours of glass are collected and mixed at different ratios with natural rock material trying to use the blends in pavement layers. Whole subsurface specimen sequentially consists of a single layer of R and a layer of G-R blend. 12G/88R and 45G/55R mix ratios are employed in this research, the thickness of G-R layer was changed, and the results were compared between the pure rock and the layered specimens. The relations between resilient module (Mr) and permanent deformation with sequence number are presented. During the earlier stages of RLTT, the results indicated that the 45G/55R specimen shows higher moduli than R specimen.

Keywords: Rock base course, layered structure, glass, resilient modulus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 625
3044 UAV Position Estimation Using Remote Radio Head With Adaptive Power Control

Authors: Hyeon-Cheol Lee

Abstract:

The adaptive power control of Code Division Multiple Access (CDMA) communications using Remote Radio Head (RRH) between multiple Unmanned Aerial Vehicles (UAVs) with a link-budget based Signal-to-Interference Ratio (SIR) estimate is applied to four inner loop power control algorithms. It is concluded that Base Station (BS) can calculate not only UAV distance using linearity between speed and Consecutive Transmit-Power-Control Ratio (CTR) of Adaptive Step-size Closed Loop Power Control (ASCLPC), Consecutive TPC Ratio Step-size Closed Loop Power Control (CS-CLPC), Fixed Step-size Power Control (FSPC), but also UAV position with Received Signal Strength Indicator (RSSI) ratio of RRHs.

Keywords: speed estimation, adaptive power control, link-budget, SIR, multi-bit quantizer, RRH

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2141
3043 A New Approach for Network Reconfiguration Problem in Order to Deviation Bus Voltage Minimization with Regard to Probabilistic Load Model and DGs

Authors: Mahmood Reza Shakarami, Reza Sedaghati

Abstract:

Recently, distributed generation technologies have received much attention for the potential energy savings and reliability assurances that might be achieved as a result of their widespread adoption. The distribution feeder reconfiguration (DFR) is one of the most important control schemes in the distribution networks, which can be affected by DGs. This paper presents a new approach to DFR at the distribution networks considering wind turbines. The main objective of the DFR is to minimize the deviation of the bus voltage. Since the DFR is a nonlinear optimization problem, we apply the Adaptive Modified Firefly Optimization (AMFO) approach to solve it. As a result of the conflicting behavior of the single- objective function, a fuzzy based clustering technique is employed to reach the set of optimal solutions called Pareto solutions. The approach is tested on the IEEE 32-bus standard test system.

Keywords: Adaptive Modified Firefly Optimization (AMFO), Pareto solutions, feeder reconfiguration, wind turbines, bus voltage.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2017
3042 An Examination of the Factors Influencing Software Development Effort

Authors: Zhizhong Jiang, Peter Naudé

Abstract:

Effective evaluation of software development effort is an important aspect of successful project management. Based on a large database with 4106 projects ever developed, this study statistically examines the factors that influence development effort. The factors found to be significant for effort are project size, average number of developers that worked on the project, type of development, development language, development platform, and the use of rapid application development. Among these factors, project size is the most critical cost driver. Unsurprisingly, this study found that the use of CASE tools does not necessarily reduce development effort, which adds support to the claim that the use of tools is subtle. As many of the current estimation models are rarely or unsuccessfully used, this study proposes a parsimonious parametric model for the prediction of effort which is both simple and more accurate than previous models.

Keywords: Development effort, function points, team size, development language, CASE tool, rapid application development.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2508