Search results for: particle swarm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1691

Search results for: particle swarm

1571 Basic Properties of a Fundamental Particle: Behavioral-Physical and Visual Methods for the Study of Fundamental Particle

Authors: Shukran M. Dadayev

Abstract:

To author's best knowledge, in this paper, the Basic Properties and Research methods of a Fundamental Particle is studied for the first time. That's to say, Fundamental Particle has not been discovered in the Nature yet. Because Fundamental Particle consists of specific Physical, Geometrical and Internal bases. Geometrical and Internal characteristics that are considered significant for the elementary and fundamental particles aren’t basic properties, characteristics or criteria of a Fundamental Particle. Of course, completely new Physical and Visual experimental methods of Quantum mechanics and Behavioral-Physical investigations of Particles are needed to study and discover the Fundamental Particle. These are new Physical, Visual and Behavioral-Physical experimental methods for describing and discovering the Fundamental Particle in the Nature and Microworld. Fundamental Particle consists of the same Energy-Mass-Motion system and a symmetry of Energy-Mass-Motion. Fundamental Particle supplies each of the elementary particles with the same Energy-Mass-Motion system at the same time and regulates each of the particles. Fundamental Particle gives Energy, Mass and Motion to each particles at the same time, each of the Particles consists of acquired Energy-Mass-Motion system and symmetry. Energy, Mass, Motion given by the Fundamental Particle to the particles are Symmetrical Equivalent and they remain in their primary shapes in all cases. Fundamental Particle gives Energy-Mass-Motion system and symmetry consisting of different measures and functions to each of the particles. The Motion given by the Fundamental Particle to the particles is Gravitation, Gravitational Interaction not only gives Motion, but also cause Motion by attracting. All Substances, Fields and Cosmic objects consist of Energy-Mass-Motion. The Field also includes specific Mass. They are always Energetic, Massive and Active. Fundamental Particle establishes the bases of the Nature. Supplement and Regulating of all the particles existing in the Nature belongs to Fundamental Particle.

Keywords: basic properties of a fundamental particle, behavioral-physical and visual methods, energy-mass-motion system and symmetrical equivalence, fundamental particle

Procedia PDF Downloads 3452
1570 Desing of PSS and SVC to Improve Power System Stability

Authors: Mahmoud Samkan

Abstract:

In this paper, the design and assessment of new coordination between Power System Stabilizers (PSSs) and Static Var Compensator (SVC) in a multimachine power system via statistical method are proposed. The coordinated design problem of PSSs and SVC over a wide range of loading conditions is handled as an optimization problem. The Bacterial Swarming Optimization (BSO), which synergistically couples the Bacterial Foraging (BF) with the Particle Swarm Optimization (PSO), is employed to seek for optimal controllers parameters. By minimizing the proposed objective function, in which the speed deviations between generators are involved; stability performance of the system is enhanced. To compare the capability of PSS and SVC, both are designed independently, and then in a coordinated manner. Simultaneous tuning of the BSO based coordinated controller gives robust damping performance over wide range of operating conditions and large disturbance in compare to optimized PSS controller based on BSO (BSOPSS) and optimized SVC controller based on BSO (BSOSVC). Moreover, a statistical T test is executed to validate the robustness of coordinated controller versus uncoordinated one.

Keywords: SVC, PSSs, multimachine power system, coordinated design, bacteria swarm optimization, statistical assessment

Procedia PDF Downloads 353
1569 Effect of Bi-Dispersity on Particle Clustering in Sedimentation

Authors: Ali Abbas Zaidi

Abstract:

In free settling or sedimentation, particles form clusters at high Reynolds number and dilute suspensions. It is due to the entrapment of particles in the wakes of upstream particles. In this paper, the effect of bi-dispersity of settling particles on particle clustering is investigated using particle-resolved direct numerical simulation. Immersed boundary method is used for particle fluid interactions and discrete element method is used for particle-particle interactions. The solid volume fraction used in the simulation is 1% and the Reynolds number based on Sauter mean diameter is 350. Both solid volume fraction and Reynolds number lie in the clustering regime of sedimentation. In simulations, the particle diameter ratio (i.e. diameter of larger particle to smaller particle (d₁/d₂)) is varied from 2:1, 3:1 and 4:1. For each case of particle diameter ratio, solid volume fraction for each particle size (φ₁/φ₂) is varied from 1:1, 1:2 and 2:1. For comparison, simulations are also performed for monodisperse particles. For studying particles clustering, radial distribution function and instantaneous location of particles in the computational domain are studied. It is observed that the degree of particle clustering decreases with the increase in the bi-dispersity of settling particles. The smallest degree of particle clustering or dispersion of particles is observed for particles with d₁/d₂ equal to 4:1 and φ₁/φ₂ equal to 1:2. Simulations showed that the reduction in particle clustering by increasing bi-dispersity is due to the difference in settling velocity of particles. Particles with larger size settle faster and knockout the smaller particles from clustered regions of particles in the computational domain.

Keywords: dispersion in bi-disperse settling particles, particle microstructures in bi-disperse suspensions, particle resolved direct numerical simulations, settling of bi-disperse particles

Procedia PDF Downloads 174
1568 Towards a Robust Patch Based Multi-View Stereo Technique for Textureless and Occluded 3D Reconstruction

Authors: Ben Haines, Li Bai

Abstract:

Patch based reconstruction methods have been and still are one of the top performing approaches to 3D reconstruction to date. Their local approach to refining the position and orientation of a patch, free of global minimisation and independent of surface smoothness, make patch based methods extremely powerful in recovering fine grained detail of an objects surface. However, patch based approaches still fail to faithfully reconstruct textureless or highly occluded surface regions thus though performing well under lab conditions, deteriorate in industrial or real world situations. They are also computationally expensive. Current patch based methods generate point clouds with holes in texturesless or occluded regions that require expensive energy minimisation techniques to fill and interpolate a high fidelity reconstruction. Such shortcomings hinder the adaptation of the methods for industrial applications where object surfaces are often highly textureless and the speed of reconstruction is an important factor. This paper presents on-going work towards a multi-resolution approach to address the problems, utilizing particle swarm optimisation to reconstruct high fidelity geometry, and increasing robustness to textureless features through an adapted approach to the normalised cross correlation. The work also aims to speed up the reconstruction using advances in GPU technologies and remove the need for costly initialization and expansion. Through the combination of these enhancements, it is the intention of this work to create denser patch clouds even in textureless regions within a reasonable time. Initial results show the potential of such an approach to construct denser point clouds with a comparable accuracy to that of the current top-performing algorithms.

Keywords: 3D reconstruction, multiview stereo, particle swarm optimisation, photo consistency

Procedia PDF Downloads 179
1567 Investigating Data Normalization Techniques in Swarm Intelligence Forecasting for Energy Commodity Spot Price

Authors: Yuhanis Yusof, Zuriani Mustaffa, Siti Sakira Kamaruddin

Abstract:

Data mining is a fundamental technique in identifying patterns from large data sets. The extracted facts and patterns contribute in various domains such as marketing, forecasting, and medical. Prior to that, data are consolidated so that the resulting mining process may be more efficient. This study investigates the effect of different data normalization techniques, which are Min-max, Z-score, and decimal scaling, on Swarm-based forecasting models. Recent swarm intelligence algorithms employed includes the Grey Wolf Optimizer (GWO) and Artificial Bee Colony (ABC). Forecasting models are later developed to predict the daily spot price of crude oil and gasoline. Results showed that GWO works better with Z-score normalization technique while ABC produces better accuracy with the Min-Max. Nevertheless, the GWO is more superior that ABC as its model generates the highest accuracy for both crude oil and gasoline price. Such a result indicates that GWO is a promising competitor in the family of swarm intelligence algorithms.

Keywords: artificial bee colony, data normalization, forecasting, Grey Wolf optimizer

Procedia PDF Downloads 447
1566 A Novel Algorithm for Production Scheduling

Authors: Ali Mohammadi Bolban Abad, Fariborz Ahmadi

Abstract:

Optimization in manufacture is a method to use limited resources to obtain the best performance and reduce waste. In this paper a new algorithm based on eurygaster life is introduced to obtain a plane in which task order and completion time of resources are defined. Evaluation results show our approach has less make span when the resources are allocated with some products in comparison to genetic algorithm.

Keywords: evolutionary computation, genetic algorithm, particle swarm optimization, NP-Hard problems, production scheduling

Procedia PDF Downloads 352
1565 Swarm Optimization of Unmanned Vehicles and Object Localization

Authors: Venkataramana Sovenahalli Badigar, B. M. Suryakanth, Akshar Prasanna, Karthik Veeramalai, Vishwak Ram Vishwak Ram

Abstract:

Technological advances have led to widespread autonomy in vehicles. Empowering these autonomous with the intelligence to cooperate amongst themselves leads to a more efficient use of the resources available to them. This paper proposes a demonstration of a swarm algorithm implemented on a group of autonomous vehicles. The demonstration involves two ground bots and an aerial drone which cooperate amongst them to locate an object of interest. The object of interest is modelled using a high-intensity light source which acts as a beacon. The ground bots are light sensitive and move towards the beacon. The ground bots and the drone traverse in random paths and jointly locate the beacon. This finds application in various scenarios in where human interference is difficult such as search and rescue during natural disasters, delivering crucial packages in perilous situations, etc. Experimental results show that the modified swarm algorithm implemented in this system has better performance compared to fully random based moving algorithm for object localization and tracking.

Keywords: swarm algorithm, object localization, ground bots, drone, beacon

Procedia PDF Downloads 224
1564 Unified Gas-Kinetic Scheme for Gas-Particle Flow in Shock-Induced Fluidization of Particles Bed

Authors: Zhao Wang, Hong Yan

Abstract:

In this paper, a unified-gas kinetic scheme (UGKS) for the gas-particle flow is constructed. UGKS is a direct modeling method for both continuum and rarefied flow computations. The dynamics of particle and gas are described as rarefied and continuum flow, respectively. Therefore, we use the Bhatnagar-Gross-Krook (BGK) equation for the particle distribution function. For the gas phase, the gas kinetic scheme for Navier-Stokes equation is solved. The momentum transfer between gas and particle is achieved by the acceleration term added to the BGK equation. The new scheme is tested by a 2cm-in-thickness dense bed comprised of glass particles with 1.5mm in diameter, and reasonable agreement is achieved.

Keywords: gas-particle flow, unified gas-kinetic scheme, momentum transfer, shock-induced fluidization

Procedia PDF Downloads 234
1563 Value Index, a Novel Decision Making Approach for Waste Load Allocation

Authors: E. Feizi Ashtiani, S. Jamshidi, M.H Niksokhan, A. Feizi Ashtiani

Abstract:

Waste load allocation (WLA) policies may use multi-objective optimization methods to find the most appropriate and sustainable solutions. These usually intend to simultaneously minimize two criteria, total abatement costs (TC) and environmental violations (EV). If other criteria, such as inequity, need for minimization as well, it requires introducing more binary optimizations through different scenarios. In order to reduce the calculation steps, this study presents value index as an innovative decision making approach. Since the value index contains both the environmental violation and treatment costs, it can be maximized simultaneously with the equity index. It implies that the definition of different scenarios for environmental violations is no longer required. Furthermore, the solution is not necessarily the point with minimized total costs or environmental violations. This idea is testified for Haraz River, in north of Iran. Here, the dissolved oxygen (DO) level of river is simulated by Streeter-Phelps equation in MATLAB software. The WLA is determined for fish farms using multi-objective particle swarm optimization (MOPSO) in two scenarios. At first, the trade-off curves of TC-EV and TC-Inequity are plotted separately as the conventional approach. In the second, the Value-Equity curve is derived. The comparative results show that the solutions are in a similar range of inequity with lower total costs. This is due to the freedom of environmental violation attained in value index. As a result, the conventional approach can well be replaced by the value index particularly for problems optimizing these objectives. This reduces the process to achieve the best solutions and may find better classification for scenario definition. It is also concluded that decision makers are better to focus on value index and weighting its contents to find the most sustainable alternatives based on their requirements.

Keywords: waste load allocation (WLA), value index, multi objective particle swarm optimization (MOPSO), Haraz River, equity

Procedia PDF Downloads 399
1562 A Proposed Optimized and Efficient Intrusion Detection System for Wireless Sensor Network

Authors: Abdulaziz Alsadhan, Naveed Khan

Abstract:

In recent years intrusions on computer network are the major security threat. Hence, it is important to impede such intrusions. The hindrance of such intrusions entirely relies on its detection, which is primary concern of any security tool like Intrusion Detection System (IDS). Therefore, it is imperative to accurately detect network attack. Numerous intrusion detection techniques are available but the main issue is their performance. The performance of IDS can be improved by increasing the accurate detection rate and reducing false positive. The existing intrusion detection techniques have the limitation of usage of raw data set for classification. The classifier may get jumble due to redundancy, which results incorrect classification. To minimize this problem, Principle Component Analysis (PCA), Linear Discriminant Analysis (LDA), and Local Binary Pattern (LBP) can be applied to transform raw features into principle features space and select the features based on their sensitivity. Eigen values can be used to determine the sensitivity. To further classify, the selected features greedy search, back elimination, and Particle Swarm Optimization (PSO) can be used to obtain a subset of features with optimal sensitivity and highest discriminatory power. These optimal feature subset used to perform classification. For classification purpose, Support Vector Machine (SVM) and Multilayer Perceptron (MLP) used due to its proven ability in classification. The Knowledge Discovery and Data mining (KDD’99) cup dataset was considered as a benchmark for evaluating security detection mechanisms. The proposed approach can provide an optimal intrusion detection mechanism that outperforms the existing approaches and has the capability to minimize the number of features and maximize the detection rates.

Keywords: Particle Swarm Optimization (PSO), Principle Component Analysis (PCA), Linear Discriminant Analysis (LDA), Local Binary Pattern (LBP), Support Vector Machine (SVM), Multilayer Perceptron (MLP)

Procedia PDF Downloads 336
1561 The Influence of the Form of Grain on the Mechanical Behaviour of Sand

Authors: Mohamed Boualem Salah

Abstract:

The size and shape of soil particles reflect the formation history of the grains. In turn, the macro scale behavior of the soil mass results from particle level interactions which are affected by particle shape. Sphericity, roundness and smoothness characterize different scales associated to particle shape. New experimental data and data from previously published studies are gathered into two databases to explore the effects of particle shape on packing as well as small and large-strain properties of sandy soils. Data analysis shows that increased particle irregularity (angularity and/or eccentricity) leads to: an increase in emax and emin, a decrease in stiffness yet with increased sensitivity to the state of stress, an increase in compressibility under zero-lateral strain loading, and an increase in critical state friction angle φcs and intercept Γ with a weak effect on slope λ. Therefore, particle shape emerges as a significant soil index property that needs to be properly characterized and documented, particularly in clean sands and gravels. The systematic assessment of particle shape will lead to a better understanding of sand behavior.

Keywords: angularity, eccentricity, shape particle, behavior of soil

Procedia PDF Downloads 385
1560 Simulation of Stress in Graphite Anode of Lithium-Ion Battery: Intra and Inter-Particle

Authors: Wenxin Mei, Jinhua Sun, Qingsong Wang

Abstract:

The volume expansion of lithium-ion batteries is mainly induced by intercalation induced stress within the negative electrode, resulting in capacity degradation and even battery failure. Stress generation due to lithium intercalation into graphite particles is investigated based on an electrochemical-mechanical model in this work. The two-dimensional model presented is fully coupled, inclusive of the impacts of intercalation-induced stress, stress-induced intercalation, to evaluate the lithium concentration, stress generation, and displacement intra and inter-particle. The results show that the distribution of lithium concentration and stress exhibits an analogous pattern, which reflects the relation between lithium diffusion and stress. The results of inter-particle stress indicate that larger Von-Mises stress is displayed where the two particles are in contact with each other, and deformation at the edge of particles is also observed, predicting fracture. Additionally, the maximum inter-particle stress at the end of lithium intercalation is nearly ten times the intraparticle stress. And the maximum inter-particle displacement is increased by 24% compared to the single-particle. Finally, the effect of graphite particle arrangement on inter-particle stress is studied. It is found that inter-particle stress with tighter arrangement exhibits lower stress. This work can provide guidance for predicting the intra and inter-particle stress to take measures to avoid cracking of electrode material.

Keywords: electrochemical-mechanical model, graphite particle, lithium concentration, lithium ion battery, stress

Procedia PDF Downloads 161
1559 Screen Method of Distributed Cooperative Navigation Factors for Unmanned Aerial Vehicle Swarm

Authors: Can Zhang, Qun Li, Yonglin Lei, Zhi Zhu, Dong Guo

Abstract:

Aiming at the problem of factor screen in distributed collaborative navigation of dense UAV swarm, an efficient distributed collaborative navigation factor screen method is proposed. The method considered the balance between computing load and positioning accuracy. The proposed algorithm utilized the factor graph model to implement a distributed collaborative navigation algorithm. The GNSS information of the UAV itself and the ranging information between the UAVs are used as the positioning factors. In this distributed scheme, a local factor graph is established for each UAV. The positioning factors of nodes with good geometric position distribution and small variance are selected to participate in the navigation calculation. To demonstrate and verify the proposed methods, the simulation and experiments in different scenarios are performed in this research. Simulation results show that the proposed scheme achieves a good balance between the computing load and positioning accuracy in the distributed cooperative navigation calculation of UAV swarm. This proposed algorithm has important theoretical and practical value for both industry and academic areas.

Keywords: screen method, cooperative positioning system, UAV swarm, factor graph, cooperative navigation

Procedia PDF Downloads 48
1558 Modeling of Polyethylene Particle Size Distribution in Fluidized Bed Reactors

Authors: R. Marandi, H. Shahrir, T. Nejad Ghaffar Borhani, M. Kamaruddin

Abstract:

In the present study, a steady state population balance model was developed to predict the polymer particle size distribution (PSD) in ethylene gas phase fluidized bed olefin polymerization reactors. The multilayer polymeric flow model (MPFM) was used to calculate the growth rate of a single polymer particle under intra-heat and mass transfer resistance. The industrial plant data were used to calculate the growth rate of polymer particle and the polymer PSD. Numerical simulations carried out to describe the influence of effective monomer diffusion coefficient, polymerization rate and initial catalyst size on the catalyst particle growth and final polymer PSD. The results present that the intra-heat and mass limitation is important for the ethylene polymerization, the growth rate of particle and the polymer PSD in the fluidized bed reactor. The effect of the agglomeration on the PSD is also considered. The result presents that the polymer particle size distribution becomes broader as the agglomeration exits.

Keywords: population balance, olefin polymerization, fluidized bed reactor, particle size distribution, agglomeration

Procedia PDF Downloads 301
1557 Discrete Breeding Swarm for Cost Minimization of Parallel Job Shop Scheduling Problem

Authors: Tarek Aboueldahab, Hanan Farag

Abstract:

Parallel Job Shop Scheduling Problem (JSP) is a multi-objective and multi constrains NP- optimization problem. Traditional Artificial Intelligence techniques have been widely used; however, they could be trapped into the local minimum without reaching the optimum solution, so we propose a hybrid Artificial Intelligence model (AI) with Discrete Breeding Swarm (DBS) added to traditional Artificial Intelligence to avoid this trapping. This model is applied in the cost minimization of the Car Sequencing and Operator Allocation (CSOA) problem. The practical experiment shows that our model outperforms other techniques in cost minimization.

Keywords: parallel job shop scheduling problem, artificial intelligence, discrete breeding swarm, car sequencing and operator allocation, cost minimization

Procedia PDF Downloads 150
1556 Hierarchical Operation Strategies for Grid Connected Building Microgrid with Energy Storage and Photovoltatic Source

Authors: Seon-Ho Yoon, Jin-Young Choi, Dong-Jun Won

Abstract:

This paper presents hierarchical operation strategies which are minimizing operation error between day ahead operation plan and real time operation. Operating power systems between centralized and decentralized approaches can be represented as hierarchical control scheme, featured as primary control, secondary control and tertiary control. Primary control is known as local control, featuring fast response. Secondary control is referred to as microgrid Energy Management System (EMS). Tertiary control is responsible of coordinating the operations of multi-microgrids. In this paper, we formulated 3 stage microgrid operation strategies which are similar to hierarchical control scheme. First stage is to set a day ahead scheduled output power of Battery Energy Storage System (BESS) which is only controllable source in microgrid and it is optimized to minimize cost of exchanged power with main grid using Particle Swarm Optimization (PSO) method. Second stage is to control the active and reactive power of BESS to be operated in day ahead scheduled plan in case that State of Charge (SOC) error occurs between real time and scheduled plan. The third is rescheduling the system when the predicted error is over the limited value. The first stage can be compared with the secondary control in that it adjusts the active power. The second stage is comparable to the primary control in that it controls the error in local manner. The third stage is compared with the secondary control in that it manages power balancing. The proposed strategies will be applied to one of the buildings in Electronics and Telecommunication Research Institute (ETRI). The building microgrid is composed of Photovoltaic (PV) generation, BESS and load and it will be interconnected with the main grid. Main purpose of that is minimizing operation cost and to be operated in scheduled plan. Simulation results support validation of proposed strategies.

Keywords: Battery Energy Storage System (BESS), Energy Management System (EMS), Microgrid (MG), Particle Swarm Optimization (PSO)

Procedia PDF Downloads 229
1555 Particle Deflection in a PDMS Microchannel Caused by a Plane Travelling Surface Acoustic Wave

Authors: Florian Keipert, Hagen Schmitd

Abstract:

The size selective separation of different species in a microfluidic system is an actual task in biological or medical research. Former works dealt with the utilisation of the acoustic radiation force (ARF) caused by a plane travelling Surface Acoustic Wave (tSAW). In literature the ARF is described by a dimensionless parameter κ, depending on the wavelength and the particle diameter. To our knowledge research was done for values 0.2 < κ < 5.8 showing that the ARF is dominating the acoustic streaming force (ASF) for κ > 1.2. As a consequence the particle separation is limited by κ. In addition the dependence on the electrical power level was examined but only for κ > 1 pointing out an increased particle deflection for higher electrical power levels. Nevertheless a detailed study on the ASF and ARF especially for κ < 1 is still missing. In our setup we used a tSAW with a wavelength λ = 90 µm and 3 µm PS particles corresponding to κ = 0.3. Herewith the influence of the applied electrical power level on the particle deflection in a polydimethylsiloxan micro channel was investigated. Our results show an increased particle deflection for an increased electrical power level, which coincides with the reported results for κ > 1. Therefore particle separation is in contrast to literature also possible for lower κ values. Thereby the experimental setup can be generally simplified by a coordinated electrical power level for the specific particle size. Furthermore this raises the question of whether this particle deflection is caused only by the ARF as adopted so far or by the ASF or the sum of both forces. To investigate this fact a 0% - 24% saline solution was used and thus the mismatch between the compressibility of the PS particle and the working fluid could be changed. Therefore it is possible to change the relative strength between ARF and ASF and consequently the particle deflection. We observed a decreasing in the particle deflection for an increased NaCl content up to a 12% saline solution and subsequently an increasing of the particle deflection. Our observation could be explained by the acoustic contrast factor Φ, which depends on the compressibility mismatch. The compressibility of water is increased by the NaCl and the range of a 0% - 24% saline solution covers the PS particle compressibility. Hence the particle deflection reaches a minimum value for the accordance between compressibility of PS particle and saline solution. This minimum value can be estimated as the particle deflection only caused by the ASF. Knowing the particle deflection due to the ASF the particle deflection caused by the ARF can be calculated and thus finally the relation between both forces. Concluding, the particle deflection and therefore the size selective particle separation generated by a tSAW can be achieved for values κ < 1, simplifying actual setups by adjusting the electrical power level. Beyond we studied for the first time the relative strength between ARF and ASF to characterise the particle deflection in a microchannel.

Keywords: ARF, ASF, particle separation, saline solution, tSAW

Procedia PDF Downloads 233
1554 Gaussian Particle Flow Bernoulli Filter for Single Target Tracking

Authors: Hyeongbok Kim, Lingling Zhao, Xiaohong Su, Junjie Wang

Abstract:

The Bernoulli filter is a precise Bayesian filter for single target tracking based on the random finite set theory. The standard Bernoulli filter often underestimates the number of targets. This study proposes a Gaussian particle flow (GPF) Bernoulli filter employing particle flow to migrate particles from prior to posterior positions to improve the performance of the standard Bernoulli filter. By employing the particle flow filter, the computational speed of the Bernoulli filters is significantly improved. In addition, the GPF Bernoulli filter provides a more accurate estimation compared with that of the standard Bernoulli filter. Simulation results confirm the improved tracking performance and computational speed in two- and three-dimensional scenarios compared with other algorithms.

Keywords: Bernoulli filter, particle filter, particle flow filter, random finite sets, target tracking

Procedia PDF Downloads 56
1553 Simulation Study on Particle Fluidization and Drying in a Spray Fluidized Bed

Authors: Jinnan Guo, Daoyin Liu

Abstract:

The quality of final products in the coating process significantly depends on particle fluidization and drying in the spray-fluidized bed. In this study, fluidizing gas temperature and velocity are changed, and their effects on particle flow, moisture content, and heat transfer in a spray fluidized bed are investigated by the CFD – Discrete Element Model (DEM). The gas flow velocity distribution of the fluidized bed is symmetrical, with high velocity in the middle and low velocity on both sides. During the heating process, the particles inside the central tube and at the bottom of the bed are rapidly heated. The particle circulation in the annular area is heated slowly and the temperature is low. The inconsistency of particle circulation results in two peaks in the probability density distribution of the particle temperature during the heating process, and the overall temperature of the particles increases uniformly. During the drying process, the distribution of particle moisture transitions from initial uniform moisture to two peaks, and then the number of completely dried (moisture content of 0) particles gradually increases. Increasing the fluidizing gas temperature and velocity improves particle circulation, drying and heat transfer in the bed. The current study provides an effective method for studying the hydrodynamics of spray fluidized beds with simultaneous processes of heating and particle fluidization.

Keywords: heat transfer, CFD-DEM, spray fluidized bed, drying

Procedia PDF Downloads 23
1552 An Efficient Process Analysis and Control Method for Tire Mixing Operation

Authors: Hwang Ho Kim, Do Gyun Kim, Jin Young Choi, Sang Chul Park

Abstract:

Since tire production process is very complicated, company-wide management of it is very difficult, necessitating considerable amounts of capital and labors. Thus, productivity should be enhanced and maintained competitive by developing and applying effective production plans. Among major processes for tire manufacturing, consisting of mixing component preparation, building and curing, the mixing process is an essential and important step because the main component of tire, called compound, is formed at this step. Compound as a rubber synthesis with various characteristics plays its own role required for a tire as a finished product. Meanwhile, scheduling tire mixing process is similar to flexible job shop scheduling problem (FJSSP) because various kinds of compounds have their unique orders of operations, and a set of alternative machines can be used to process each operation. In addition, setup time required for different operations may differ due to alteration of additives. In other words, each operation of mixing processes requires different setup time depending on the previous one, and this kind of feature, called sequence dependent setup time (SDST), is a very important issue in traditional scheduling problems such as flexible job shop scheduling problems. However, despite of its importance, there exist few research works dealing with the tire mixing process. Thus, in this paper, we consider the scheduling problem for tire mixing process and suggest an efficient particle swarm optimization (PSO) algorithm to minimize the makespan for completing all the required jobs belonging to the process. Specifically, we design a particle encoding scheme for the considered scheduling problem, including a processing sequence for compounds and machine allocation information for each job operation, and a method for generating a tire mixing schedule from a given particle. At each iteration, the coordination and velocity of particles are updated, and the current solution is compared with new solution. This procedure is repeated until a stopping condition is satisfied. The performance of the proposed algorithm is validated through a numerical experiment by using some small-sized problem instances expressing the tire mixing process. Furthermore, we compare the solution of the proposed algorithm with it obtained by solving a mixed integer linear programming (MILP) model developed in previous research work. As for performance measure, we define an error rate which can evaluate the difference between two solutions. As a result, we show that PSO algorithm proposed in this paper outperforms MILP model with respect to the effectiveness and efficiency. As the direction for future work, we plan to consider scheduling problems in other processes such as building, curing. We can also extend our current work by considering other performance measures such as weighted makespan or processing times affected by aging or learning effects.

Keywords: compound, error rate, flexible job shop scheduling problem, makespan, particle encoding scheme, particle swarm optimization, sequence dependent setup time, tire mixing process

Procedia PDF Downloads 235
1551 Cold Model Experimental Research on Particle Velocity Distribution in Gas-Solid Circulating Fluidized Bed for Methanol-To-Olefins Process

Authors: Yongzheng Li, Hongfang Ma, Qiwen Sun, Haitao Zhang, Weiyong Ying

Abstract:

Radial profiles of particle velocities were investigated in a 6.1 m tall methanol-to-olefins cold model experimental device using a TSI laser Doppler velocimeter. The measurement of axial levels was conducted in the full developed region. The effect of axial level on flow development was not obvious under the same operating condition. Superficial gas velocity and solid circulating rate had significant influence on particle velocity in the center region of the riser. Besides, comparisons between upward, downward and average particle velocity were conducted. The average particle velocity was close to upward velocity and higher than downward velocity in radial locations except the wall region of riser.

Keywords: circulating fluidized bed, laser doppler velocimeter, particle velocity, radial profile

Procedia PDF Downloads 343
1550 Investigation of Single Particle Breakage inside an Impact Mill

Authors: E. Ghasemi Ardi, K. J. Dong, A. B. Yu, R. Y. Yang

Abstract:

In current work, a numerical model based on the discrete element method (DEM) was developed which provided information about particle dynamic and impact event condition inside a laboratory scale impact mill (Fritsch). It showed that each particle mostly experiences three impacts inside the mill. While the first impact frequently happens at front surface of the rotor’s rib, the frequent location of the second impact is side surfaces of the rotor’s rib. It was also showed that while the first impact happens at small impact angle mostly varying around 35º, the second impact happens at around 70º which is close to normal impact condition. Also analyzing impact energy revealed that varying mill speed from 6000 to 14000 rpm, the ratio of first impact’s average impact energy and minimum required energy to break particle (Wₘᵢₙ) increased from 0.30 to 0.85. Moreover, it was seen that second impact poses intense impact energy on particle which can be considered as the main cause of particle splitting. Finally, obtained information from DEM simulation along with obtained data from conducted experiments was implemented in semi-empirical equations in order to find selection and breakage functions. Then, using a back-calculation approach, those parameters were used to predict the PSDs of ground particles under different impact energies. Results were compared with experiment results and showed reasonable accuracy and prediction ability.

Keywords: single particle breakage, particle dynamic, population balance model, particle size distribution, discrete element method

Procedia PDF Downloads 262
1549 Effect of Copper Particle on the PD Characteristics in a Coaxial Duct with Mixture of SF6 (10%) and N2 (90%) Gases

Authors: B. Rajesh Kamath, J. Sundara Rajan, M. K. Veeraiah, M. Z. Kurian

Abstract:

Insulation performance of a gas insulated system is severely affected by particle contaminants. These metallic particles adversely affect the characteristics of insulating system. These particles can produce surface charges due to partial discharge activities. These particles which are free to move enhance the local electric fields. This paper deals with the influence of conducting particle placed in a co-axial duct on the discharge characteristics of gas mixtures. Co-axial duct placed in a high pressure chamber is used for the purpose. A gas pressure of 0.1, 0.2 and 0.3 MPa have been considered with a 10:90 SF6 and N2 gas mixtures. The 2D and 3D histograms of clean duct and duct with copper particle are discussed in this paper.

Keywords: coaxial duct, gas insulated system, gas mixtures, metallic particle, partial discharges, histograms

Procedia PDF Downloads 356
1548 Critical Velocities for Particle Transport from Experiments and CFD Simulations

Authors: Sajith Sajeev, Brenton McLaury, Siamack Shirazi

Abstract:

In the petroleum industry, solid particles are often present along with the produced fluids. It is imperative to keep particles from accumulating in flow lines. In this study, various experiments are conducted to study sand particle transport, where critical velocity is defined as the average fluid velocity to keep particles continuously moving. Many parameters related to the fluid, particles and pipe affect the transport process. Experimental results are presented varying the particle concentration. Additionally, CFD simulations using a discrete element modeling (DEM) approach are presented to compare with experimental result.

Keywords: particle transport, critical velocity, CFD, DEM

Procedia PDF Downloads 283
1547 Evaluation of Particle Settling in Flow Chamber

Authors: Abdulrahman Alenezi, B. Stefan

Abstract:

Abstract— The investigation of fluids containing particles or filaments includes a category of complex fluids and is vital in both theory and application. The forecast of particle behaviors plays a significant role in the existing technology as well as future technology. This paper focuses on the prediction of the particle behavior through the investigation of the particle disentrainment from a pipe on a horizontal air stream. This allows for examining the influence of the particle physical properties on its behavior when falling on horizontal air stream. This investigation was conducted on a device located at the University of Greenwich's Medway Campus. Two materials were selected to carry out this study: Salt and Glass Beads particles. The shape of the Slat particles is cubic where the shape of the Glass Beads is almost spherical. The outcome from the experimental work were presented in terms of distance travelled by the particles according to their diameters as After that, the particles sizes were measured using Laser Diffraction device and used to determine the drag coefficient and the settling velocity.

Keywords: flow experiment, drag coefficient, Particle Settling, Flow Chamber

Procedia PDF Downloads 100
1546 Simulation of Complex-Shaped Particle Breakage with a Bonded Particle Model Using the Discrete Element Method

Authors: Felix Platzer, Eric Fimbinger

Abstract:

In Discrete Element Method (DEM) simulations, the breakage behavior of particles can be simulated based on different principles. In the case of large, complex-shaped particles that show various breakage patterns depending on the scenario leading to the failure and often only break locally instead of fracturing completely, some of these principles do not lead to realistic results. The reason for this is that in said cases, the methods in question, such as the Particle Replacement Method (PRM) or Voronoi Fracture, replace the initial particle (that is intended to break) into several sub-particles when certain breakage criteria are reached, such as exceeding the fracture energy. That is why those methods are commonly used for the simulation of materials that fracture completely instead of breaking locally. That being the case, when simulating local failure, it is advisable to pre-build the initial particle from sub-particles that are bonded together. The dimensions of these sub-particles consequently define the minimum size of the fracture results. This structure of bonded sub-particles enables the initial particle to break at the location of the highest local loads – due to the failure of the bonds in those areas – with several sub-particle clusters being the result of the fracture, which can again also break locally. In this project, different methods for the generation and calibration of complex-shaped particle conglomerates using bonded particle modeling (BPM) to enable the ability to depict more realistic fracture behavior were evaluated based on the example of filter cake. The method that proved suitable for this purpose and which furthermore allows efficient and realistic simulation of breakage behavior of complex-shaped particles applicable to industrial-sized simulations is presented in this paper.

Keywords: bonded particle model, DEM, filter cake, particle breakage

Procedia PDF Downloads 180
1545 Software Verification of Systematic Resampling for Optimization of Particle Filters

Authors: Osiris Terry, Kenneth Hopkinson, Laura Humphrey

Abstract:

Systematic resampling is the most popularly used resampling method in particle filters. This paper seeks to further the understanding of systematic resampling by defining a formula made up of variables from the sampling equation and the particle weights. The formula is then verified via SPARK, a software verification language. The verified systematic resampling formula states that the minimum/maximum number of possible samples taken of a particle is equal to the floor/ceiling value of particle weight divided by the sampling interval, respectively. This allows for the creation of a randomness spectrum that each resampling method can fall within. Methods on the lower end, e.g., systematic resampling, have less randomness and, thus, are quicker to reach an estimate. Although lower randomness allows for error by having a larger bias towards the size of the weight, having this bias creates vulnerabilities to the noise in the environment, e.g., jamming. Conclusively, this is the first step in characterizing each resampling method. This will allow target-tracking engineers to pick the best resampling method for their environment instead of choosing the most popularly used one.

Keywords: SPARK, software verification, resampling, systematic resampling, particle filter, tracking

Procedia PDF Downloads 51
1544 Effects of Particle Size Distribution on Mechanical Strength and Physical Properties in Engineered Quartz Stone

Authors: Esra Arici, Duygu Olmez, Murat Ozkan, Nurcan Topcu, Furkan Capraz, Gokhan Deniz, Arman Altinyay

Abstract:

Engineered quartz stone is a composite material comprising approximately 90 wt.% fine quartz aggregate with a variety of particle size ranges and `10 wt.% unsaturated polyester resin (UPR). In this study, the objective is to investigate the influence of particle size distribution on mechanical strength and physical properties of the engineered stone slabs. For this purpose, granular quartz with two particle size ranges of 63-200 µm and 100-300 µm were used individually and mixed with a difference in ratios of mixing. The void volume of each granular packing was measured in order to define the amount of filler; quartz powder with the size of less than 38 µm, and UPR required filling inter-particle spaces. Test slabs were prepared using vibration-compression under vacuum. The study reports that both impact strength and flexural strength of samples increased as the mix ratio of the particle size range of 63-200 µm increased. On the other hand, the values of water absorption rate, apparent density and abrasion resistance were not affected by the particle size distribution owing to vacuum compaction. It is found that increasing the mix ratio of the particle size range of 63-200 µm caused the higher porosity. This led to increasing in the amount of the binder paste needed. It is also observed that homogeneity in the slabs was improved with the particle size range of 63-200 µm.

Keywords: engineered quartz stone, fine quartz aggregate, granular packing, mechanical strength, particle size distribution, physical properties.

Procedia PDF Downloads 114
1543 Influence of Processing Parameters on the Reliability of Sieving as a Particle Size Distribution Measurements

Authors: Eseldin Keleb

Abstract:

In the pharmaceutical industry particle size distribution is an important parameter for the characterization of pharmaceutical powders. The powder flowability, reactivity and compatibility, which have a decisive impact on the final product, are determined by particle size and size distribution. Therefore, the aim of this study was to evaluate the influence of processing parameters on the particle size distribution measurements. Different Size fractions of α-lactose monohydrate and 5% polyvinylpyrrolidone were prepared by wet granulation and were used for the preparation of samples. The influence of sieve load (50, 100, 150, 200, 250, 300, and 350 g), processing time (5, 10, and 15 min), sample size ratios (high percentage of small and large particles), type of disturbances (vibration and shaking) and process reproducibility have been investigated. Results obtained showed that a sieve load of 50 g produce the best separation, a further increase in sample weight resulted in incomplete separation even after the extension of the processing time for 15 min. Performing sieving using vibration was rapider and more efficient than shaking. Meanwhile between day reproducibility showed that particle size distribution measurements are reproducible. However, for samples containing 70% fines or 70% large particles, which processed at optimized parameters, the incomplete separation was always observed. These results indicated that sieving reliability is highly influenced by the particle size distribution of the sample and care must be taken for samples with particle size distribution skewness.

Keywords: sieving, reliability, particle size distribution, processing parameters

Procedia PDF Downloads 579
1542 A Numerical and Experimental Study on Fast Pyrolysis of Single Wood Particle

Authors: Hamid Rezaei, Xiaotao Bi, C. Jim Lim, Anthony Lau, Shahab Sokhansanj

Abstract:

A one-dimensional heat transfer model coupled with the kinetic information has been used to predict the overall pyrolysis mass loss of a single wood particle. The kinetic parameters were determined experimentally and the regime and characteristics of the conversion were evaluated in terms of the particle size and reactor temperature. The order of overall mass loss changed from n=1 at temperatures lower than 350 °C to n=0.5 at temperatures higher that 350 °C. Conversion time analysis showed that particles larger than 0.5 mm were controlled by internal thermal resistances. The valid range of particle size to use the simplified lumped model depends on the fluid temperature around the particles. The critical particle size was 0.6-0.7 mm for the fluid temperature of 500 °C and 0.9-1.0 mm for the fluid temperature of 100 °C. Experimental pyrolysis of moist particles did not show distinct drying and pyrolysis stages. The process was divided into two hypothetical drying and pyrolysis dominated zones and empirical correlations are developed to predict the rate of mass loss in each zone.

Keywords: pyrolysis, kinetics, model, single particle

Procedia PDF Downloads 286