Search results for: parallel particle swarm optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5579

Search results for: parallel particle swarm optimization

5339 Two-Stage Approach for Solving the Multi-Objective Optimization Problem on Combinatorial Configurations

Authors: Liudmyla Koliechkina, Olena Dvirna

Abstract:

The statement of the multi-objective optimization problem on combinatorial configurations is formulated, and the approach to its solution is proposed. The problem is of interest as a combinatorial optimization one with many criteria, which is a model of many applied tasks. The approach to solving the multi-objective optimization problem on combinatorial configurations consists of two stages; the first is the reduction of the multi-objective problem to the single criterion based on existing multi-objective optimization methods, the second stage solves the directly replaced single criterion combinatorial optimization problem by the horizontal combinatorial method. This approach provides the optimal solution to the multi-objective optimization problem on combinatorial configurations, taking into account additional restrictions for a finite number of steps.

Keywords: discrete set, linear combinatorial optimization, multi-objective optimization, Pareto solutions, partial permutation set, structural graph

Procedia PDF Downloads 131
5338 Simulation of the Reactive Rotational Molding Using Smoothed Particle Hydrodynamics

Authors: A. Hamidi, S. Khelladi, L. Illoul, A. Tcharkhtchi

Abstract:

Reactive rotational molding (RRM) is a process to manufacture hollow plastic parts with reactive material has several advantages compared to conventional roto molding of thermoplastic powders: process cycle time is shorter; raw material is less expensive because polymerization occurs during processing and high-performance polymers may be used such as thermosets, thermoplastics or blends. However, several phenomena occur during this process which makes the optimization of the process quite complex. In this study, we have used a mixture of isocyanate and polyol as a reactive system. The chemical transformation of this system to polyurethane has been studied by thermal analysis and rheology tests. Thanks to these results of the curing process and rheological measurements, the kinetic and rheokinetik of polyurethane was identified. Smoothed Particle Hydrodynamics, a Lagrangian meshless method, was chosen to simulate reactive fluid flow in 2 and 3D configurations of the polyurethane during the process taking into account the chemical, and chemiorehological results obtained experimentally in this study.

Keywords: reactive rotational molding, simulation, smoothed particle hydrodynamics, surface tension, rheology, free surface flows, viscoelastic, interpolation

Procedia PDF Downloads 252
5337 Cold Model Experimental Research on Particle Velocity Distribution in Gas-Solid Circulating Fluidized Bed for Methanol-To-Olefins Process

Authors: Yongzheng Li, Hongfang Ma, Qiwen Sun, Haitao Zhang, Weiyong Ying

Abstract:

Radial profiles of particle velocities were investigated in a 6.1 m tall methanol-to-olefins cold model experimental device using a TSI laser Doppler velocimeter. The measurement of axial levels was conducted in the full developed region. The effect of axial level on flow development was not obvious under the same operating condition. Superficial gas velocity and solid circulating rate had significant influence on particle velocity in the center region of the riser. Besides, comparisons between upward, downward and average particle velocity were conducted. The average particle velocity was close to upward velocity and higher than downward velocity in radial locations except the wall region of riser.

Keywords: circulating fluidized bed, laser doppler velocimeter, particle velocity, radial profile

Procedia PDF Downloads 339
5336 Single Pole-To-Earth Fault Detection and Location on the Tehran Railway System Using ICA and PSO Trained Neural Network

Authors: Masoud Safarishaal

Abstract:

Detecting the location of pole-to-earth faults is essential for the safe operation of the electrical system of the railroad. This paper aims to use a combination of evolutionary algorithms and neural networks to increase the accuracy of single pole-to-earth fault detection and location on the Tehran railroad power supply system. As a result, the Imperialist Competitive Algorithm (ICA) and Particle Swarm Optimization (PSO) are used to train the neural network to improve the accuracy and convergence of the learning process. Due to the system's nonlinearity, fault detection is an ideal application for the proposed method, where the 600 Hz harmonic ripple method is used in this paper for fault detection. The substations were simulated by considering various situations in feeding the circuit, the transformer, and typical Tehran metro parameters that have developed the silicon rectifier. Required data for the network learning process has been gathered from simulation results. The 600Hz component value will change with the change of the location of a single pole to the earth's fault. Therefore, 600Hz components are used as inputs of the neural network when fault location is the output of the network system. The simulation results show that the proposed methods can accurately predict the fault location.

Keywords: single pole-to-pole fault, Tehran railway, ICA, PSO, artificial neural network

Procedia PDF Downloads 75
5335 Investigation of Single Particle Breakage inside an Impact Mill

Authors: E. Ghasemi Ardi, K. J. Dong, A. B. Yu, R. Y. Yang

Abstract:

In current work, a numerical model based on the discrete element method (DEM) was developed which provided information about particle dynamic and impact event condition inside a laboratory scale impact mill (Fritsch). It showed that each particle mostly experiences three impacts inside the mill. While the first impact frequently happens at front surface of the rotor’s rib, the frequent location of the second impact is side surfaces of the rotor’s rib. It was also showed that while the first impact happens at small impact angle mostly varying around 35º, the second impact happens at around 70º which is close to normal impact condition. Also analyzing impact energy revealed that varying mill speed from 6000 to 14000 rpm, the ratio of first impact’s average impact energy and minimum required energy to break particle (Wₘᵢₙ) increased from 0.30 to 0.85. Moreover, it was seen that second impact poses intense impact energy on particle which can be considered as the main cause of particle splitting. Finally, obtained information from DEM simulation along with obtained data from conducted experiments was implemented in semi-empirical equations in order to find selection and breakage functions. Then, using a back-calculation approach, those parameters were used to predict the PSDs of ground particles under different impact energies. Results were compared with experiment results and showed reasonable accuracy and prediction ability.

Keywords: single particle breakage, particle dynamic, population balance model, particle size distribution, discrete element method

Procedia PDF Downloads 259
5334 Effect of Copper Particle on the PD Characteristics in a Coaxial Duct with Mixture of SF6 (10%) and N2 (90%) Gases

Authors: B. Rajesh Kamath, J. Sundara Rajan, M. K. Veeraiah, M. Z. Kurian

Abstract:

Insulation performance of a gas insulated system is severely affected by particle contaminants. These metallic particles adversely affect the characteristics of insulating system. These particles can produce surface charges due to partial discharge activities. These particles which are free to move enhance the local electric fields. This paper deals with the influence of conducting particle placed in a co-axial duct on the discharge characteristics of gas mixtures. Co-axial duct placed in a high pressure chamber is used for the purpose. A gas pressure of 0.1, 0.2 and 0.3 MPa have been considered with a 10:90 SF6 and N2 gas mixtures. The 2D and 3D histograms of clean duct and duct with copper particle are discussed in this paper.

Keywords: coaxial duct, gas insulated system, gas mixtures, metallic particle, partial discharges, histograms

Procedia PDF Downloads 352
5333 Critical Velocities for Particle Transport from Experiments and CFD Simulations

Authors: Sajith Sajeev, Brenton McLaury, Siamack Shirazi

Abstract:

In the petroleum industry, solid particles are often present along with the produced fluids. It is imperative to keep particles from accumulating in flow lines. In this study, various experiments are conducted to study sand particle transport, where critical velocity is defined as the average fluid velocity to keep particles continuously moving. Many parameters related to the fluid, particles and pipe affect the transport process. Experimental results are presented varying the particle concentration. Additionally, CFD simulations using a discrete element modeling (DEM) approach are presented to compare with experimental result.

Keywords: particle transport, critical velocity, CFD, DEM

Procedia PDF Downloads 279
5332 Pushing the Boundary of Parallel Tractability for Ontology Materialization via Boolean Circuits

Authors: Zhangquan Zhou, Guilin Qi

Abstract:

Materialization is an important reasoning service for applications built on the Web Ontology Language (OWL). To make materialization efficient in practice, current research focuses on deciding tractability of an ontology language and designing parallel reasoning algorithms. However, some well-known large-scale ontologies, such as YAGO, have been shown to have good performance for parallel reasoning, but they are expressed in ontology languages that are not parallelly tractable, i.e., the reasoning is inherently sequential in the worst case. This motivates us to study the problem of parallel tractability of ontology materialization from a theoretical perspective. That is we aim to identify the ontologies for which materialization is parallelly tractable, i.e., in the NC complexity. Since the NC complexity is defined based on Boolean circuit that is widely used to investigate parallel computing problems, we first transform the problem of materialization to evaluation of Boolean circuits, and then study the problem of parallel tractability based on circuits. In this work, we focus on datalog rewritable ontology languages. We use Boolean circuits to identify two classes of datalog rewritable ontologies (called parallelly tractable classes) such that materialization over them is parallelly tractable. We further investigate the parallel tractability of materialization of a datalog rewritable OWL fragment DHL (Description Horn Logic). Based on the above results, we analyze real-world datasets and show that many ontologies expressed in DHL belong to the parallelly tractable classes.

Keywords: ontology materialization, parallel reasoning, datalog, Boolean circuit

Procedia PDF Downloads 235
5331 Comparative Analysis of Classical and Parallel Inpainting Algorithms Based on Affine Combinations of Projections on Convex Sets

Authors: Irina Maria Artinescu, Costin Radu Boldea, Eduard-Ionut Matei

Abstract:

The paper is a comparative study of two classical variants of parallel projection methods for solving the convex feasibility problem with their equivalents that involve variable weights in the construction of the solutions. We used a graphical representation of these methods for inpainting a convex area of an image in order to investigate their effectiveness in image reconstruction applications. We also presented a numerical analysis of the convergence of these four algorithms in terms of the average number of steps and execution time in classical CPU and, alternatively, in parallel GPU implementation.

Keywords: convex feasibility problem, convergence analysis, inpainting, parallel projection methods

Procedia PDF Downloads 134
5330 Achievable Average Secrecy Rates over Bank of Parallel Independent Fading Channels with Friendly Jamming

Authors: Munnujahan Ara

Abstract:

In this paper, we investigate the effect of friendly jamming power allocation strategies on the achievable average secrecy rate over a bank of parallel fading wiretap channels. We investigate the achievable average secrecy rate in parallel fading wiretap channels subject to Rayleigh and Rician fading. The achievable average secrecy rate, due to the presence of a line-of-sight component in the jammer channel is also evaluated. Moreover, we study the detrimental effect of correlation across the parallel sub-channels, and evaluate the corresponding decrease in the achievable average secrecy rate for the various fading configurations. We also investigate the tradeoff between the transmission power and the jamming power for a fixed total power budget. Our results, which are applicable to current orthogonal frequency division multiplexing (OFDM) communications systems, shed further light on the achievable average secrecy rates over a bank of parallel fading channels in the presence of friendly jammers.

Keywords: fading parallel channels, wire-tap channel, OFDM, secrecy capacity, power allocation

Procedia PDF Downloads 471
5329 Evaluation of Particle Settling in Flow Chamber

Authors: Abdulrahman Alenezi, B. Stefan

Abstract:

Abstract— The investigation of fluids containing particles or filaments includes a category of complex fluids and is vital in both theory and application. The forecast of particle behaviors plays a significant role in the existing technology as well as future technology. This paper focuses on the prediction of the particle behavior through the investigation of the particle disentrainment from a pipe on a horizontal air stream. This allows for examining the influence of the particle physical properties on its behavior when falling on horizontal air stream. This investigation was conducted on a device located at the University of Greenwich's Medway Campus. Two materials were selected to carry out this study: Salt and Glass Beads particles. The shape of the Slat particles is cubic where the shape of the Glass Beads is almost spherical. The outcome from the experimental work were presented in terms of distance travelled by the particles according to their diameters as After that, the particles sizes were measured using Laser Diffraction device and used to determine the drag coefficient and the settling velocity.

Keywords: flow experiment, drag coefficient, Particle Settling, Flow Chamber

Procedia PDF Downloads 96
5328 City-Wide Simulation on the Effects of Optimal Appliance Scheduling in a Time-of-Use Residential Environment

Authors: Rudolph Carl Barrientos, Juwaln Diego Descallar, Rainer James Palmiano

Abstract:

Household Appliance Scheduling Systems (HASS) coupled with a Time-of-Use (TOU) pricing scheme, a form of Demand Side Management (DSM), is not widely utilized in the Philippines’ residential electricity sector. This paper’s goal is to encourage distribution utilities (DUs) to adopt HASS and TOU by analyzing the effect of household schedulers on the electricity price and load profile in a residential environment. To establish this, a city based on an implemented survey is generated using Monte Carlo Analysis (MCA). Then, a Binary Particle Swarm Optimization (BPSO) algorithm-based HASS is developed considering user satisfaction, electricity budget, appliance prioritization, energy storage systems, solar power, and electric vehicles. The simulations were assessed under varying levels of user compliance. Results showed that the average electricity cost, peak demand, and peak-to-average ratio (PAR) of the city load profile were all reduced. Therefore, the deployment of the HASS and TOU pricing scheme is beneficial for both stakeholders.

Keywords: appliance scheduling, DSM, TOU, BPSO, city-wide simulation, electric vehicle, appliance prioritization, energy storage system, solar power

Procedia PDF Downloads 63
5327 Simulation of Complex-Shaped Particle Breakage with a Bonded Particle Model Using the Discrete Element Method

Authors: Felix Platzer, Eric Fimbinger

Abstract:

In Discrete Element Method (DEM) simulations, the breakage behavior of particles can be simulated based on different principles. In the case of large, complex-shaped particles that show various breakage patterns depending on the scenario leading to the failure and often only break locally instead of fracturing completely, some of these principles do not lead to realistic results. The reason for this is that in said cases, the methods in question, such as the Particle Replacement Method (PRM) or Voronoi Fracture, replace the initial particle (that is intended to break) into several sub-particles when certain breakage criteria are reached, such as exceeding the fracture energy. That is why those methods are commonly used for the simulation of materials that fracture completely instead of breaking locally. That being the case, when simulating local failure, it is advisable to pre-build the initial particle from sub-particles that are bonded together. The dimensions of these sub-particles consequently define the minimum size of the fracture results. This structure of bonded sub-particles enables the initial particle to break at the location of the highest local loads – due to the failure of the bonds in those areas – with several sub-particle clusters being the result of the fracture, which can again also break locally. In this project, different methods for the generation and calibration of complex-shaped particle conglomerates using bonded particle modeling (BPM) to enable the ability to depict more realistic fracture behavior were evaluated based on the example of filter cake. The method that proved suitable for this purpose and which furthermore allows efficient and realistic simulation of breakage behavior of complex-shaped particles applicable to industrial-sized simulations is presented in this paper.

Keywords: bonded particle model, DEM, filter cake, particle breakage

Procedia PDF Downloads 175
5326 Effects of Particle Size Distribution on Mechanical Strength and Physical Properties in Engineered Quartz Stone

Authors: Esra Arici, Duygu Olmez, Murat Ozkan, Nurcan Topcu, Furkan Capraz, Gokhan Deniz, Arman Altinyay

Abstract:

Engineered quartz stone is a composite material comprising approximately 90 wt.% fine quartz aggregate with a variety of particle size ranges and `10 wt.% unsaturated polyester resin (UPR). In this study, the objective is to investigate the influence of particle size distribution on mechanical strength and physical properties of the engineered stone slabs. For this purpose, granular quartz with two particle size ranges of 63-200 µm and 100-300 µm were used individually and mixed with a difference in ratios of mixing. The void volume of each granular packing was measured in order to define the amount of filler; quartz powder with the size of less than 38 µm, and UPR required filling inter-particle spaces. Test slabs were prepared using vibration-compression under vacuum. The study reports that both impact strength and flexural strength of samples increased as the mix ratio of the particle size range of 63-200 µm increased. On the other hand, the values of water absorption rate, apparent density and abrasion resistance were not affected by the particle size distribution owing to vacuum compaction. It is found that increasing the mix ratio of the particle size range of 63-200 µm caused the higher porosity. This led to increasing in the amount of the binder paste needed. It is also observed that homogeneity in the slabs was improved with the particle size range of 63-200 µm.

Keywords: engineered quartz stone, fine quartz aggregate, granular packing, mechanical strength, particle size distribution, physical properties.

Procedia PDF Downloads 111
5325 Cylindrical Spacer Shape Optimization for Enhanced Inhalation Therapy

Authors: Shahab Azimi, Siamak Arzanpour, Anahita Sayyar

Abstract:

Asthma and Chronic obstructive pulmonary disease (COPD) are common lung diseases that have a significant global impact. Pressurized metered dose inhalers (pMDIs) are widely used for treatment, but they can have limitations such as high medication release speed resulting in drug deposition in the mouth or oral cavity and difficulty achieving proper synchronization with inhalation by users. Spacers are add-on devices that improve the efficiency of pMDIs by reducing the release speed and providing space for aerosol particle breakup to have finer and medically effective medication. The aim of this study is to optimize the size and cylindrical shape of spacers to enhance their drug delivery performance. The study was based on fluid dynamics theory and employed Ansys software for simulation and optimization. Results showed that optimization of the spacer's geometry greatly influenced its performance and improved drug delivery. This study provides a foundation for future research on enhancing the efficiency of inhalation therapy for lung diseases.

Keywords: asthma, COPD, pressurized metered dose inhalers, spacers, CFD, shape optimization

Procedia PDF Downloads 49
5324 Automatic Tuning for a Systemic Model of Banking Originated Losses (SYMBOL) Tool on Multicore

Authors: Ronal Muresano, Andrea Pagano

Abstract:

Nowadays, the mathematical/statistical applications are developed with more complexity and accuracy. However, these precisions and complexities have brought as result that applications need more computational power in order to be executed faster. In this sense, the multicore environments are playing an important role to improve and to optimize the execution time of these applications. These environments allow us the inclusion of more parallelism inside the node. However, to take advantage of this parallelism is not an easy task, because we have to deal with some problems such as: cores communications, data locality, memory sizes (cache and RAM), synchronizations, data dependencies on the model, etc. These issues are becoming more important when we wish to improve the application’s performance and scalability. Hence, this paper describes an optimization method developed for Systemic Model of Banking Originated Losses (SYMBOL) tool developed by the European Commission, which is based on analyzing the application's weakness in order to exploit the advantages of the multicore. All these improvements are done in an automatic and transparent manner with the aim of improving the performance metrics of our tool. Finally, experimental evaluations show the effectiveness of our new optimized version, in which we have achieved a considerable improvement on the execution time. The time has been reduced around 96% for the best case tested, between the original serial version and the automatic parallel version.

Keywords: algorithm optimization, bank failures, OpenMP, parallel techniques, statistical tool

Procedia PDF Downloads 339
5323 The Parallelization of Algorithm Based on Partition Principle for Association Rules Discovery

Authors: Khadidja Belbachir, Hafida Belbachir

Abstract:

subsequently the expansion of the physical supports storage and the needs ceaseless to accumulate several data, the sequential algorithms of associations’ rules research proved to be ineffective. Thus the introduction of the new parallel versions is imperative. We propose in this paper, a parallel version of a sequential algorithm “Partition”. This last is fundamentally different from the other sequential algorithms, because it scans the data base only twice to generate the significant association rules. By consequence, the parallel approach does not require much communication between the sites. The proposed approach was implemented for an experimental study. The obtained results, shows a great reduction in execution time compared to the sequential version and Count Distributed algorithm.

Keywords: association rules, distributed data mining, partition, parallel algorithms

Procedia PDF Downloads 367
5322 Influence of Processing Parameters on the Reliability of Sieving as a Particle Size Distribution Measurements

Authors: Eseldin Keleb

Abstract:

In the pharmaceutical industry particle size distribution is an important parameter for the characterization of pharmaceutical powders. The powder flowability, reactivity and compatibility, which have a decisive impact on the final product, are determined by particle size and size distribution. Therefore, the aim of this study was to evaluate the influence of processing parameters on the particle size distribution measurements. Different Size fractions of α-lactose monohydrate and 5% polyvinylpyrrolidone were prepared by wet granulation and were used for the preparation of samples. The influence of sieve load (50, 100, 150, 200, 250, 300, and 350 g), processing time (5, 10, and 15 min), sample size ratios (high percentage of small and large particles), type of disturbances (vibration and shaking) and process reproducibility have been investigated. Results obtained showed that a sieve load of 50 g produce the best separation, a further increase in sample weight resulted in incomplete separation even after the extension of the processing time for 15 min. Performing sieving using vibration was rapider and more efficient than shaking. Meanwhile between day reproducibility showed that particle size distribution measurements are reproducible. However, for samples containing 70% fines or 70% large particles, which processed at optimized parameters, the incomplete separation was always observed. These results indicated that sieving reliability is highly influenced by the particle size distribution of the sample and care must be taken for samples with particle size distribution skewness.

Keywords: sieving, reliability, particle size distribution, processing parameters

Procedia PDF Downloads 575
5321 A Numerical and Experimental Study on Fast Pyrolysis of Single Wood Particle

Authors: Hamid Rezaei, Xiaotao Bi, C. Jim Lim, Anthony Lau, Shahab Sokhansanj

Abstract:

A one-dimensional heat transfer model coupled with the kinetic information has been used to predict the overall pyrolysis mass loss of a single wood particle. The kinetic parameters were determined experimentally and the regime and characteristics of the conversion were evaluated in terms of the particle size and reactor temperature. The order of overall mass loss changed from n=1 at temperatures lower than 350 °C to n=0.5 at temperatures higher that 350 °C. Conversion time analysis showed that particles larger than 0.5 mm were controlled by internal thermal resistances. The valid range of particle size to use the simplified lumped model depends on the fluid temperature around the particles. The critical particle size was 0.6-0.7 mm for the fluid temperature of 500 °C and 0.9-1.0 mm for the fluid temperature of 100 °C. Experimental pyrolysis of moist particles did not show distinct drying and pyrolysis stages. The process was divided into two hypothetical drying and pyrolysis dominated zones and empirical correlations are developed to predict the rate of mass loss in each zone.

Keywords: pyrolysis, kinetics, model, single particle

Procedia PDF Downloads 284
5320 A Simplified, Fabrication-Friendly Acoustophoretic Model for Size Sensitive Particle Sorting

Authors: V. Karamzadeh, J. Adhvaryu, A. Chandrasekaran, M. Packirisamy

Abstract:

In Bulk Acoustic Wave (BAW) microfluidics, the throughput of particle sorting is dependent on the complex interplay between the geometric configuration of the channel, the size of the particles, and the properties of the fluid medium, which therefore calls for a detailed modeling and understanding of the fluid-particle interaction dynamics under an acoustic field, prior to designing the system. In this work, we propose a simplified Bulk acoustophoretic system that can be used for size dependent particle sorting. A Finite Element Method (FEM) based analytical model has been developed to study the dependence of particle sizes on channel parameters, and the sorting efficiency in a given fluid medium. Based on the results, the microfluidic system has been designed to take into account all the variables involved with the underlying physics, and has been fabricated using an additive manufacturing technique employing a commercial 3D printer, to generate a simple, cost-effective system that can be used for size sensitive particle sorting.

Keywords: 3D printing, 3D microfluidic chip, acoustophoresis, cell separation, MEMS (Microelectromechanical Systems), microfluidics

Procedia PDF Downloads 142
5319 Effect of Alloying Elements on Particle Incorporation of Boron Carbide Reinforced Aluminum Matrix Composites

Authors: Steven Ploetz, Andreas Lohmueller, Robert F. Singer

Abstract:

The outstanding performance of aluminum matrix composites (AMCs) regarding stiffness/weight ratio makes AMCs attractive material for lightweight construction. Low-density boride compounds promise simultaneously an increase in stiffness and decrease in composite density. This is why boron carbide is chosen for composite manufacturing. The composites are fabricated with the stir casting process. To avoid gas entrapment during mixing and ensure nonporous composites, partial vacuum is adapted during particle feeding and stirring. Poor wettability of boron carbide with liquid aluminum hinders particle incorporation, but alloying elements such as magnesium and titanium could improve wettability and thus particle incorporation. Next to alloying elements, adapted stirring parameters and impeller geometries improve particle incorporation and enable homogenous particle distribution and high particle volume fractions of boron carbide. AMCs with up to 15 vol.% of boron carbide particles are produced via melt stirring, resulting in an increase in stiffness and strength.

Keywords: aluminum matrix composites, boron carbide, stiffness, stir casting

Procedia PDF Downloads 281
5318 Polarization Effects in Cosmic-Ray Acceleration by Cyclotron Auto-Resonance

Authors: Yousef I. Salamin

Abstract:

Theoretical investigations, analytical as well as numerical, have shown that electrons can be accelerated to GeV energies by the process of cyclotron auto-resonance acceleration (CARA). In CARA, the particle would be injected along the lines of a uniform magnetic field aligned parallel to the direction of propagation of a plane-wave radiation field. Unfortunately, an accelerator based on CARA would be prohibitively too long and too expensive to build and maintain. However, the process stands a better chance of success near the polar cap of a compact object (such as a neutron star, a black hole or a magnetar) or in an environment created in the wake of a binary neutron-star or blackhole merger. Dynamics of the nuclides ₁H¹, ₂He⁴, ₂₆Fe⁵⁶, and ₂₈Ni⁶², in such astrophysical conditions, have been investigated by single-particle calculations and many-particle simulations. The investigations show that these nuclides can reach ZeV energies (1 ZeV = 10²¹ eV) due to interaction with super-intense radiation of wavelengths = 1 and 10 m and = 50 pm and magnetic fields of strengths at the mega- and giga-tesla levels. Examples employing radiation intensities in the range 10³²-10⁴² W/m² have been used. Employing a two-parameter model for representing the radiation field, CARA is analytically generalized to include any state of polarization, and the basic working equations are derived rigorously and in closed analytic form.

Keywords: compact objects, cosmic-ray acceleration, cyclotron auto-resonance, polarization effects, zevatron

Procedia PDF Downloads 86
5317 Economic Optimization of Shell and Tube Heat Exchanger Using Nanofluid

Authors: Hassan Hajabdollahi

Abstract:

Economic optimization of shell and tube heat exchanger (STHE) is presented in this paper. To increase the rate of heat transfer, copper oxide (CuO) nanoparticle is added into the tube side fluid and their optimum results are compared with the case of without additive nanoparticle. Total annual cost (TAC) is selected as fitness function and nine decision variables related to the heat exchanger parameters as well as concentration of nanoparticle are considered. Optimization results reveal the noticeable improvement in the TAC and in the case of heat exchanger working with nanofluid compared with the case of base fluid (8.9%). Comparison of the results between two studied cases also reveal that the lower tube diameter, tube number, and baffle spacing are needed in the case of heat exchanger working with nanofluid compared with the case of base fluid.

Keywords: shell and tube heat exchanger, nanoparticles additive, total annual cost, particle volumetric concentration

Procedia PDF Downloads 381
5316 Performance Enhancement of Hybrid Racing Car by Design Optimization

Authors: Tarang Varmora, Krupa Shah, Karan Patel

Abstract:

Environmental pollution and shortage of conventional fuel are the main concerns in the transportation sector. Most of the vehicles use an internal combustion engine (ICE), powered by gasoline fuels. This results into emission of toxic gases. Hybrid electric vehicle (HEV) powered by electric machine and ICE is capable of reducing emission of toxic gases and fuel consumption. However to build HEV, it is required to accommodate motor and batteries in the vehicle along with engine and fuel tank. Thus, overall weight of the vehicle increases. To improve the fuel economy and acceleration, the weight of the HEV can be minimized. In this paper, the design methodology to reduce the weight of the hybrid racing car is proposed. To this end, the chassis design is optimized. Further, attempt is made to obtain the maximum strength with minimum material weight. The best configuration out of the three main configurations such as series, parallel and the dual-mode (series-parallel) is chosen. Moreover, the most suitable type of motor, battery, braking system, steering system and suspension system are identified. The racing car is designed and analyzed in the simulating software. The safety of the vehicle is assured by performing static and dynamic analysis on the chassis frame. From the results, it is observed that, the weight of the racing car is reduced by 11 % without compromising on safety and cost. It is believed that the proposed design and specifications can be implemented practically for manufacturing hybrid racing car.

Keywords: design optimization, hybrid racing car, simulation, vehicle, weight reduction

Procedia PDF Downloads 268
5315 Diagnosis of the Heart Rhythm Disorders by Using Hybrid Classifiers

Authors: Sule Yucelbas, Gulay Tezel, Cuneyt Yucelbas, Seral Ozsen

Abstract:

In this study, it was tried to identify some heart rhythm disorders by electrocardiography (ECG) data that is taken from MIT-BIH arrhythmia database by subtracting the required features, presenting to artificial neural networks (ANN), artificial immune systems (AIS), artificial neural network based on artificial immune system (AIS-ANN) and particle swarm optimization based artificial neural network (PSO-NN) classifier systems. The main purpose of this study is to evaluate the performance of hybrid AIS-ANN and PSO-ANN classifiers with regard to the ANN and AIS. For this purpose, the normal sinus rhythm (NSR), atrial premature contraction (APC), sinus arrhythmia (SA), ventricular trigeminy (VTI), ventricular tachycardia (VTK) and atrial fibrillation (AF) data for each of the RR intervals were found. Then these data in the form of pairs (NSR-APC, NSR-SA, NSR-VTI, NSR-VTK and NSR-AF) is created by combining discrete wavelet transform which is applied to each of these two groups of data and two different data sets with 9 and 27 features were obtained from each of them after data reduction. Afterwards, the data randomly was firstly mixed within themselves, and then 4-fold cross validation method was applied to create the training and testing data. The training and testing accuracy rates and training time are compared with each other. As a result, performances of the hybrid classification systems, AIS-ANN and PSO-ANN were seen to be close to the performance of the ANN system. Also, the results of the hybrid systems were much better than AIS, too. However, ANN had much shorter period of training time than other systems. In terms of training times, ANN was followed by PSO-ANN, AIS-ANN and AIS systems respectively. Also, the features that extracted from the data affected the classification results significantly.

Keywords: AIS, ANN, ECG, hybrid classifiers, PSO

Procedia PDF Downloads 406
5314 Application of the Global Optimization Techniques to the Optical Thin Film Design

Authors: D. Li

Abstract:

Optical thin films are used in a wide variety of optical components and there are many software tools programmed for advancing multilayer thin film design. The available software packages for designing the thin film structure may not provide optimum designs. Normally, almost all current software programs obtain their final designs either from optimizing a starting guess or by technique, which may or may not involve a pseudorandom process, that give different answers every time, depending upon the initial conditions. With the increasing power of personal computers, functional methods in optimization and synthesis of optical multilayer systems have been developed such as DGL Optimization, Simulated Annealing, Genetic Algorithms, Needle Optimization, Inductive Optimization and Flip-Flop Optimization. Among these, DGL Optimization has proved its efficiency in optical thin film designs. The application of the DGL optimization technique to the design of optical coating is presented. A DGL optimization technique is provided, and its main features are discussed. Guidelines on the application of the DGL optimization technique to various types of design problems are given. The innovative global optimization strategies used in a software tool, OnlyFilm, to optimize multilayer thin film designs through different filter designs are outlined. OnlyFilm is a powerful, versatile, and user-friendly thin film software on the market, which combines optimization and synthesis design capabilities with powerful analytical tools for optical thin film designers. It is also the only thin film design software that offers a true global optimization function.

Keywords: optical coatings, optimization, design software, thin film design

Procedia PDF Downloads 278
5313 Experimental Design for Formulation Optimization of Nanoparticle of Cilnidipine

Authors: Arti Bagada, Kantilal Vadalia, Mihir Raval

Abstract:

Cilnidipine is practically insoluble in water which results in its insufficient oral bioavailability. The purpose of the present investigation was to formulate cilnidipine nanoparticles by nanoprecipitation method to increase the aqueous solubility and dissolution rate and hence bioavailability by utilizing various experimental statistical design modules. Experimental design were used to investigate specific effects of independent variables during preparation cilnidipine nanoparticles and corresponding responses in optimizing the formulation. Plackett Burman design for independent variables was successfully employed for optimization of nanoparticles of cilnidipine. The influence of independent variables studied were drug concentration, solvent to antisolvent ratio, polymer concentration, stabilizer concentration and stirring speed. The dependent variables namely average particle size, polydispersity index, zeta potential value and saturation solubility of the formulated nanoparticles of cilnidipine. The experiments were carried out according to 13 runs involving 5 independent variables (higher and lower levels) employing Plackett-Burman design. The cilnidipine nanoparticles were characterized by average particle size, polydispersity index value, zeta potential value and saturation solubility and it results were 149 nm, 0.314, 43.24 and 0.0379 mg/ml, respectively. The experimental results were good correlated with predicted data analysed by Plackett-Burman statistical method.

Keywords: dissolution enhancement, nanoparticles, Plackett-Burman design, nanoprecipitation

Procedia PDF Downloads 132
5312 Optimization of Interface Radio of Universal Mobile Telecommunication System Network

Authors: O. Mohamed Amine, A. Khireddine

Abstract:

Telecoms operators are always looking to meet their share of the other customers, they try to gain optimum utilization of the deployed equipment and network optimization has become essential. This project consists of optimizing UMTS network, and the study area is an urban area situated in the center of Algiers. It was initially questions to become familiar with the different communication systems (3G) and the optimization technique, its main components, and its fundamental characteristics radios were introduced.

Keywords: UMTS, UTRAN, WCDMA, optimization

Procedia PDF Downloads 341
5311 Study of Temperature Difference and Current Distribution in Parallel-Connected Cells at Low Temperature

Authors: Sara Kamalisiahroudi, Jun Huang, Zhe Li, Jianbo Zhang

Abstract:

Two types of commercial cylindrical lithium ion batteries (Panasonic 3.4 Ah NCR-18650B and Samsung 2.9 Ah INR-18650), were investigated experimentally. The capacities of these samples were individually measured using constant current-constant voltage (CC-CV) method at different ambient temperatures (-10 ℃, 0 ℃, 25 ℃). Their internal resistance was determined by electrochemical impedance spectroscopy (EIS) and pulse discharge methods. The cells with different configurations of parallel connection NCR-NCR, INR-INR and NCR-INR were charged/discharged at the aforementioned ambient temperatures. The results showed that the difference of internal resistance between cells much more evident at low temperatures. Furthermore, the parallel connection of NCR-NCR exhibits the most uniform temperature distribution in cells at -10 ℃, this feature is quite favorable for the safety of the battery pack.

Keywords: batteries in parallel connection, internal resistance, low temperature, temperature difference, current distribution

Procedia PDF Downloads 443
5310 Advanced Particle Characterisation of Suspended Sediment in the Danube River Using Automated Imaging and Laser Diffraction

Authors: Flóra Pomázi, Sándor Baranya, Zoltán Szalai

Abstract:

A harmonized monitoring of the suspended sediment transport along such a large river as the world’s most international river, the Danube River, is a rather challenging task. The traditional monitoring method in Hungary is obsolete but using indirect measurement devices and techniques like optical backscatter sensors (OBS), laser diffraction or acoustic backscatter sensors (ABS) could provide a fast and efficient alternative option of direct methods. However, these methods are strongly sensitive to the particle characteristics (i.e. particle shape, particle size and mineral composition). The current method does not provide sufficient information about particle size distribution, mineral analysis is rarely done, and the shape of the suspended sediment particles have not been examined yet. The aims of the study are (1) to determine the particle characterisation of suspended sediment in the Danube River using advanced particle characterisation methods as laser diffraction and automated imaging, and (2) to perform a sensitivity analysis of the indirect methods in order to determine the impact of suspended particle characteristics. The particle size distribution is determined by laser diffraction. The particle shape and mineral composition analysis is done by the Morphologi G3ID image analyser. The investigated indirect measurement devices are the LISST-Portable|XR, the LISST-ABS (Sequoia Inc.) and the Rio Grande 1200 kHz ADCP (Teledyne Marine). The major findings of this study are (1) the statistical shape of the suspended sediment particle - this is the first research in this context, (2) the actualised particle size distribution – that can be compared to historical information, so that the morphological changes can be tracked, (3) the actual mineral composition of the suspended sediment in the Danube River, and (4) the reliability of the tested indirect methods has been increased – based on the results of the sensitivity analysis and the previous findings.

Keywords: advanced particle characterisation, automated imaging, indirect methods, laser diffraction, mineral composition, suspended sediment

Procedia PDF Downloads 109