Search results for: dispersed particle swarm optimization (DPSO)
3696 Coupled Analysis with Fluid and Flexible Multibody Dynamics of 6-DOF Platform with Liquid Sloshing Tank
Authors: Sung-Pill Kim, Dae-Gyu Sung, Hee-Sung Shin, Jong-Chun Park
Abstract:
When a sloshing tank filled partially with liquid is excited with the motion of platform, it can be observed that the center of mass inside the tank is changed and impact loads is instantaneously applied to the wall, which causes dynamic loads additionally to the supporting links of platform. In this case, therefore, the dynamic behavior of platform associated with fluid motion should be considered in the early stage of design for safety and economics of the system. In this paper, the dynamic loads due to liquid sloshing motion in a rectangular tank which is loaded up on the upper deck of a Stewart platform are simulated using a coupled analysis of Moving Particle Simulation (MPS) and Flexible Multi-Body Dynamics (FMBD). The co-simulation is performed using two commercial softwares, Recurdyn for solving FMBD and Particleworks for analyzing fluid motion based on MPS method. For validating the present coupled system, a rectangular sloshing tank being enforced with inline sway motion by 1-DOF motion platform is assumed, and time-varied free-surface elevation and reaction force at a fixed joint are compared with experiments.Keywords: dynamic loads, liquid sloshing tank, Stewart platform, moving particle semi-implicit (MPS) method, flexible multi-body dynamics (FMBD)
Procedia PDF Downloads 7083695 Foregrounding Events in Modern Sundanese: The Pragmatics of Particle-to-Active Voice Marking Shift
Authors: Rama Munajat
Abstract:
Discourse information levels may be viewed from either a background-foreground distinction or a multi-level perspective, and cross-linguistic studies on this area suggest that each information level is marked by a specific linguistic device. In this sense, Sundanese, spoken in Indonesia’s West Javanese Province, further differentiates the background and foreground information into ordinary and significant types. This paper will report an ongoing shift from particle-to-active voice marking in the way Sundanese signals foregrounding events. The shift relates to decades of contact with Bahasa Indonesia (Indonesia’s official language) and linguistic compatibility between the two surface marking strategies. Representative data analyzed include three groups of short stories in both Sundanese and Bahasa Indonesia (Indonesian) published in three periods: before 1945, 1965-2006, and 2016-2019. In the first group of Sundanese data, forward-moving events dominantly appear in particle KA (Kecap Anteuran, word-accompanying) constructions, where the KA represents different particles that co-occur with a special group of verbs. The second group, however, shows that the foregrounded events are more frequently described in active-voice forms with a subject-predicate (SP) order. Subsequently, the third offers stronger evidence for the use of the SP structure. As for the Indonesian data, the foregrounding events in the first group occur in verb-initial and passive-voice constructions, while in the second and third, the events more frequently appear in active-voice structures (subject-predicate sequence). The marking shift above suggests a structural influence from Indonesian, stemmed from generational differences among authors of the Sundanese short stories, particularly related to their education and language backgrounds. The first group of short stories – published before 1945 or before Indonesia's independence from Dutch – were written by native speakers of Sundanese who spoke Indonesian as a foreign language and went through the Dutch education system. The second group of authors, on the other hand, represents a generation of Sundanese native speakers who spoke Indonesian as a second language. Finally, the third group consists of authors who are bilingual speakers of both Sundanese and Indonesian. The data suggest that the last two groups of authors completed the Indonesian education system. With these, the use of subject-predicate sequences to denote foregrounding events began to appear more frequently in the second group and then became more dominant in those of the third. The coded data also signify that cohesion, coherence, and pragmatic purposes in Particle KA constructions are intact in their respective active-voice structure counterparts. For instance, the foregrounding events in Particle KA constructions occur in Sentence-initial KA and Pre-verbal KA forms, whereas those in the active-voice are described in Subject-Predicate (SP) and Zero-Subject active-voice patterns. Cross-language data further demonstrate that the Sentence-initial KA and the SP active-voice structures each contain an overt noun phrase (NP) co-referential with one of the entities introduced in a preceding context. Similarly, the pre-verbal KA and Zero-Subject active-voice patterns have a deleted noun phrase unambiguously referable to the only one entity previously mentioned. The presence and absence of an NP inform a pragmatic strategy to place prominence on topic/given and comment/new information, respectively.Keywords: discourse analysis, foregrounding marking, pragmatics, language contact
Procedia PDF Downloads 1403694 Bioinformatics High Performance Computation and Big Data
Authors: Javed Mohammed
Abstract:
Right now, bio-medical infrastructure lags well behind the curve. Our healthcare system is dispersed and disjointed; medical records are a bit of a mess; and we do not yet have the capacity to store and process the crazy amounts of data coming our way from widespread whole-genome sequencing. And then there are privacy issues. Despite these infrastructure challenges, some researchers are plunging into bio medical Big Data now, in hopes of extracting new and actionable knowledge. They are doing delving into molecular-level data to discover bio markers that help classify patients based on their response to existing treatments; and pushing their results out to physicians in novel and creative ways. Computer scientists and bio medical researchers are able to transform data into models and simulations that will enable scientists for the first time to gain a profound under-standing of the deepest biological functions. Solving biological problems may require High-Performance Computing HPC due either to the massive parallel computation required to solve a particular problem or to algorithmic complexity that may range from difficult to intractable. Many problems involve seemingly well-behaved polynomial time algorithms (such as all-to-all comparisons) but have massive computational requirements due to the large data sets that must be analyzed. High-throughput techniques for DNA sequencing and analysis of gene expression have led to exponential growth in the amount of publicly available genomic data. With the increased availability of genomic data traditional database approaches are no longer sufficient for rapidly performing life science queries involving the fusion of data types. Computing systems are now so powerful it is possible for researchers to consider modeling the folding of a protein or even the simulation of an entire human body. This research paper emphasizes the computational biology's growing need for high-performance computing and Big Data. It illustrates this article’s indispensability in meeting the scientific and engineering challenges of the twenty-first century, and how Protein Folding (the structure and function of proteins) and Phylogeny Reconstruction (evolutionary history of a group of genes) can use HPC that provides sufficient capability for evaluating or solving more limited but meaningful instances. This article also indicates solutions to optimization problems, and benefits Big Data and Computational Biology. The article illustrates the Current State-of-the-Art and Future-Generation Biology of HPC Computing with Big Data.Keywords: high performance, big data, parallel computation, molecular data, computational biology
Procedia PDF Downloads 3653693 Optimal Production and Maintenance Policy for a Partially Observable Production System with Stochastic Demand
Authors: Leila Jafari, Viliam Makis
Abstract:
In this paper, the joint optimization of the economic manufacturing quantity (EMQ), safety stock level, and condition-based maintenance (CBM) is presented for a partially observable, deteriorating system subject to random failure. The demand is stochastic and it is described by a Poisson process. The stochastic model is developed and the optimization problem is formulated in the semi-Markov decision process framework. A modification of the policy iteration algorithm is developed to find the optimal policy. A numerical example is presented to compare the optimal policy with the policy considering zero safety stock.Keywords: condition-based maintenance, economic manufacturing quantity, safety stock, stochastic demand
Procedia PDF Downloads 4653692 Sensitivity Analysis Optimization of a Horizontal Axis Wind Turbine from Its Aerodynamic Profiles
Authors: Kevin Molina, Daniel Ortega, Manuel Martinez, Andres Gonzalez-Estrada, William Pinto
Abstract:
Due to the increasing environmental impact, the wind energy is getting strong. This research studied the relationship between the power produced by a horizontal axis wind turbine (HAWT) and the aerodynamic profiles used for its construction. The analysis is studied using the Computational Fluid Dynamic (CFD), presenting the parallel between the energy generated by a turbine designed with selected profiles and another one optimized. For the study, a selection process was carried out from profile NACA 6 digits recommended by the National Renewable Energy Laboratory (NREL) for the construction of this type of turbines. The selection was taken into account different characteristics of the wind (speed and density) and the profiles (aerodynamic coefficients Cl and Cd to different Reynolds and incidence angles). From the selected profiles, was carried out a sensitivity analysis optimization process between its geometry and the aerodynamic forces that are induced on it. The 3D model of the turbines was realized using the Blade Element Momentum method (BEM) and both profiles. The flow fields on the turbines were simulated, obtaining the forces induced on the blade, the torques produced and an increase of 3% in power due to the optimized profiles. Therefore, the results show that the sensitivity analysis optimization process can assist to increment the wind turbine power.Keywords: blade element momentum, blade, fluid structure interaction, horizontal axis wind turbine, profile design
Procedia PDF Downloads 2603691 Effect of Particles Size and Volume Fraction Concentration on the Thermal Conductivity and Thermal Diffusivity of Al2O3 Nanofluids Measured Using Transient Hot–Wire Laser Beam Deflection Technique
Authors: W. Mahmood Mat Yunus, Faris Mohammed Ali, Zainal Abidin Talib
Abstract:
In this study we present new data for the thermal conductivity enhancement in four nanofluids containing 11, 25, 50, 63 nm diameter aluminum oxide (Al2O3) nanoparticles in distilled water. The nanofluids were prepared using single step method (i.e. by dispersing nanoparticle directly in base fluid) which was gathered in ultrasonic device for approximately 7 hours. The transient hot-wire laser beam displacement technique was used to measure the thermal conductivity and thermal diffusivity of the prepared nanofluids. The thermal conductivity and thermal diffusivity were obtained by fitting the experimental data to the numerical data simulated for aluminum oxide in distilled water. The results show that the thermal conductivity and thermal diffusivity of nanofluids increases in non-linear behavior as the particle size increases. While, the thermal conductivity and thermal diffusivity of Al2O3 nanofluids was observed increasing linearly with concentration as the volume fraction concentration increases. We believe that the interfacial layer between solid/fluid is the main factor for the enhancement of thermal conductivity and thermal diffusivity of Al2O3 nanofluids in the present work.Keywords: transient hot wire-laser beam technique, Al2O3 nanofluid, particle size, volume fraction concentration
Procedia PDF Downloads 5533690 Separation of Composites for Recycling: Measurement of Electrostatic Charge of Carbon and Glass Fiber Particles
Authors: J. Thirunavukkarasu, M. Poulet, T. Turner, S. Pickering
Abstract:
Composite waste from manufacturing can consist of different fiber materials, including blends of different fiber. Commercially, the recycling of composite waste is currently limited to carbon fiber waste and recycling glass fiber waste is currently not economically viable due to the low cost of virgin glass fiber and the reduced mechanical properties of the recovered fibers. For this reason, the recycling of hybrid fiber materials, where carbon fiber is combined with a proportion of glass fiber, cannot be processed economically. Therefore, a separation method is required to remove the glass fiber materials during the recycling process. An electrostatic separation method is chosen for this work because of the significant difference between carbon and glass fiber electrical properties. In this study, an experimental rig has been developed to measure the electrostatic charge achievable as the materials are passed through a tube. A range of particle lengths (80-100 µm, 6 mm and 12 mm), surface state conditions (0%SA, 2%SA and 6%SA), and several tube wall materials have been studied. A polytetrafluoroethylene (PTFE) tube and recycled without sizing agent was identified as the most suitable parameters for the electrical separation method. It was also found that shorter fiber lengths helped to encourage particle flow and attain higher charge values. These findings can be used to develop a separation process to enable the cost-effective recycling of hybrid fiber composite waste.Keywords: electrostatic charging, hybrid fiber composites, recycling, short fiber composites
Procedia PDF Downloads 1323689 Irradion: Portable Small Animal Imaging and Irradiation Unit
Authors: Josef Uher, Jana Boháčová, Richard Kadeřábek
Abstract:
In this paper, we present a multi-robot imaging and irradiation research platform referred to as Irradion, with full capabilities of portable arbitrary path computed tomography (CT). Irradion is an imaging and irradiation unit entirely based on robotic arms for research on cancer treatment with ion beams on small animals (mice or rats). The platform comprises two subsystems that combine several imaging modalities, such as 2D X-ray imaging, CT, and particle tracking, with precise positioning of a small animal for imaging and irradiation. Computed Tomography: The CT subsystem of the Irradion platform is equipped with two 6-joint robotic arms that position a photon counting detector and an X-ray tube independently and freely around the scanned specimen and allow image acquisition utilizing computed tomography. Irradiation measures nearly all conventional 2D and 3D trajectories of X-ray imaging with precisely calibrated and repeatable geometrical accuracy leading to a spatial resolution of up to 50 µm. In addition, the photon counting detectors allow X-ray photon energy discrimination, which can suppress scattered radiation, thus improving image contrast. It can also measure absorption spectra and recognize different materials (tissue) types. X-ray video recording and real-time imaging options can be applied for studies of dynamic processes, including in vivo specimens. Moreover, Irradion opens the door to exploring new 2D and 3D X-ray imaging approaches. We demonstrate in this publication various novel scan trajectories and their benefits. Proton Imaging and Particle Tracking: The Irradion platform allows combining several imaging modules with any required number of robots. The proton tracking module comprises another two robots, each holding particle tracking detectors with position, energy, and time-sensitive sensors Timepix3. Timepix3 detectors can track particles entering and exiting the specimen and allow accurate guiding of photon/ion beams for irradiation. In addition, quantifying the energy losses before and after the specimen brings essential information for precise irradiation planning and verification. Work on the small animal research platform Irradion involved advanced software and hardware development that will offer researchers a novel way to investigate new approaches in (i) radiotherapy, (ii) spectral CT, (iii) arbitrary path CT, (iv) particle tracking. The robotic platform for imaging and radiation research developed for the project is an entirely new product on the market. Preclinical research systems with precision robotic irradiation with photon/ion beams combined with multimodality high-resolution imaging do not exist currently. The researched technology can potentially cause a significant leap forward compared to the current, first-generation primary devices.Keywords: arbitrary path CT, robotic CT, modular, multi-robot, small animal imaging
Procedia PDF Downloads 913688 Revalidation and Hormonization of Existing IFCC Standardized Hepatic, Cardiac, and Thyroid Function Tests by Precison Optimization and External Quality Assurance Programs
Authors: Junaid Mahmood Alam
Abstract:
Revalidating and harmonizing clinical chemistry analytical principles and optimizing methods through quality control programs and assessments is the preeminent means to attain optimal outcome within the clinical laboratory services. Present study reports revalidation of our existing IFCC regularized analytical methods, particularly hepatic and thyroid function tests, by optimization of precision analyses and processing through external and internal quality assessments and regression determination. Parametric components of hepatic (Bilirubin ALT, γGT, ALP), cardiac (LDH, AST, Trop I) and thyroid/pituitary (T3, T4, TSH, FT3, FT4) function tests were used to validate analytical techniques on automated chemistry and immunological analyzers namely Hitachi 912, Cobas 6000 e601, Cobas c501, Cobas e411 with UV kinetic, colorimetric dry chemistry principles and Electro-Chemiluminescence immunoassay (ECLi) techniques. Process of validation and revalidation was completed with evaluating and assessing the precision analyzed Preci-control data of various instruments plotting against each other with regression analyses R2. Results showed that: Revalidation and optimization of respective parameters that were accredited through CAP, CLSI and NEQAPP assessments depicted 99.0% to 99.8% optimization, in addition to the methodology and instruments used for analyses. Regression R2 analysis of BilT was 0.996, whereas that of ALT, ALP, γGT, LDH, AST, Trop I, T3, T4, TSH, FT3, and FT4 exhibited R2 0.998, 0.997, 0.993, 0.967, 0.970, 0.980, 0.976, 0.996, 0.997, 0.997, and R2 0.990, respectively. This confirmed marked harmonization of analytical methods and instrumentations thus revalidating optimized precision standardization as per IFCC recommended guidelines. It is concluded that practices of revalidating and harmonizing the existing or any new services should be followed by all clinical laboratories, especially those associated with tertiary care hospital. This is will ensure deliverance of standardized, proficiency tested, optimized services for prompt and better patient care that will guarantee maximum patients’ confidence.Keywords: revalidation, standardized, IFCC, CAP, harmonized
Procedia PDF Downloads 2693687 Optimization of Air Pollution Control Model for Mining
Authors: Zunaira Asif, Zhi Chen
Abstract:
The sustainable measures on air quality management are recognized as one of the most serious environmental concerns in the mining region. The mining operations emit various types of pollutants which have significant impacts on the environment. This study presents a stochastic control strategy by developing the air pollution control model to achieve a cost-effective solution. The optimization method is formulated to predict the cost of treatment using linear programming with an objective function and multi-constraints. The constraints mainly focus on two factors which are: production of metal should not exceed the available resources, and air quality should meet the standard criteria of the pollutant. The applicability of this model is explored through a case study of an open pit metal mine, Utah, USA. This method simultaneously uses meteorological data as a dispersion transfer function to support the practical local conditions. The probabilistic analysis and the uncertainties in the meteorological conditions are accomplished by Monte Carlo simulation. Reasonable results have been obtained to select the optimized treatment technology for PM2.5, PM10, NOx, and SO2. Additional comparison analysis shows that baghouse is the least cost option as compared to electrostatic precipitator and wet scrubbers for particulate matter, whereas non-selective catalytical reduction and dry-flue gas desulfurization are suitable for NOx and SO2 reduction respectively. Thus, this model can aid planners to reduce these pollutants at a marginal cost by suggesting control pollution devices, while accounting for dynamic meteorological conditions and mining activities.Keywords: air pollution, linear programming, mining, optimization, treatment technologies
Procedia PDF Downloads 2083686 Investigation and Optimization of DNA Isolation Efficiency Using Ferrite-Based Magnetic Nanoparticles
Authors: Tímea Gerzsenyi, Ágnes M. Ilosvai, László Vanyorek, Emma Szőri-Dorogházi
Abstract:
DNA isolation is a crucial step in many molecular biological applications for diagnostic and research purposes. However, traditional extraction requires toxic reagents, and commercially available kits are expensive, this leading to the recently wide-spread method, the magnetic nanoparticle (MNP)-based DNA isolation. Different ferrite containing MNPs were examined and compared in their plasmid DNA isolation efficiency. Among the tested MNPs, one has never been used for the extraction of plasmid molecules, marking a distinct application. pDNA isolation process was optimized for each type of nanoparticle and the best protocol was selected based on different criteria: DNA quantity, quality and integrity. With the best-performing magnetic nanoparticle, which excelled in all aspects, further tests were performed to recover genomic DNA from bacterial cells and a protocol was developed.Keywords: DNA isolation, nanobiotechnology, magnetic nanoparticles, protocol optimization, pDNA, gDNA
Procedia PDF Downloads 163685 Mathematical Modeling and Optimization of Burnishing Parameters for 15NiCr6 Steel
Authors: Tarek Litim, Ouahiba Taamallah
Abstract:
The present paper is an investigation of the effect of burnishing on the surface integrity of a component made of 15NiCr6 steel. This work shows a statistical study based on regression, and Taguchi's design has allowed the development of mathematical models to predict the output responses as a function of the technological parameters studied. The response surface methodology (RSM) showed a simultaneous influence of the burnishing parameters and observe the optimal processing parameters. ANOVA analysis of the results resulted in the validation of the prediction model with a determination coefficient R=90.60% and 92.41% for roughness and hardness, respectively. Furthermore, a multi-objective optimization allowed to identify a regime characterized by P=10kgf, i=3passes, and f=0.074mm/rev, which favours minimum roughness and maximum hardness. The result was validated by the desirability of D= (0.99 and 0.95) for roughness and hardness, respectively.Keywords: 15NiCr6 steel, burnishing, surface integrity, Taguchi, RSM, ANOVA
Procedia PDF Downloads 1943684 Design Channel Non Persistent CSMA MAC Protocol Model for Complex Wireless Systems Based on SoC
Authors: Ibrahim A. Aref, Tarek El-Mihoub, Khadiga Ben Musa
Abstract:
This paper presents Carrier Sense Multiple Access (CSMA) communication model based on SoC design methodology. Such model can be used to support the modelling of the complex wireless communication systems, therefore use of such communication model is an important technique in the construction of high performance communication. SystemC has been chosen because it provides a homogeneous design flow for complex designs (i.e. SoC and IP based design). We use a swarm system to validate CSMA designed model and to show how advantages of incorporating communication early in the design process. The wireless communication created through the modeling of CSMA protocol that can be used to achieve communication between all the agents and to coordinate access to the shared medium (channel).Keywords: systemC, modelling, simulation, CSMA
Procedia PDF Downloads 4283683 Design and Optimization of a Mini High Altitude Long Endurance (HALE) Multi-Role Unmanned Aerial Vehicle
Authors: Vishaal Subramanian, Annuatha Vinod Kumar, Santosh Kumar Budankayala, M. Senthil Kumar
Abstract:
This paper discusses the aerodynamic and structural design, simulation and optimization of a mini-High Altitude Long Endurance (HALE) UAV. The applications of this mini HALE UAV vary from aerial topological surveys, quick first aid supply, emergency medical blood transport, search and relief activates to border patrol, surveillance and estimation of forest fire progression. Although classified as a mini UAV according to UVS International, our design is an amalgamation of the features of ‘mini’ and ‘HALE’ categories, combining the light weight of the ‘mini’ and the high altitude ceiling and endurance of the HALE. Designed with the idea of implementation in India, it is in strict compliance with the UAS rules proposed by the office of the Director General of Civil Aviation. The plane can be completely automated or have partial override control and is equipped with an Infra-Red camera and a multi coloured camera with on-board storage or live telemetry, GPS system with Geo Fencing and fail safe measures. An additional of 1.5 kg payload can be attached to three major hard points on the aircraft and can comprise of delicate equipment or releasable payloads. The paper details the design, optimization process and the simulations performed using various software such as Design Foil, XFLR5, Solidworks and Ansys.Keywords: aircraft, endurance, HALE, high altitude, long range, UAV, unmanned aerial vehicle
Procedia PDF Downloads 4003682 Physicochemical Properties of Soy Protein Isolate (SPI): Starch Conjugates Treated by Sonication
Authors: Gulcin Yildiz, Hao Feng
Abstract:
In recent years there is growing interested in using soy protein because of several advantages compared to other protein sources, such as high nutritional value, steady supply, and low cost. Soy protein isolate (SPI) is the most refined soy protein product. It contains 90% protein in a moisture-free form and has some desirable functionalities. Creating a protein-polysaccharide conjugate to be the emulsifying agent rather than the protein alone can markedly enhance its stability. This study was undertaken to examine the effects of ultrasound treatments on the physicochemical properties of SPI-starch conjugates. The soy protein isolate (SPI, Pro-Fam® 955) samples were obtained from the Archer Daniels Midland Company. Protein concentrations were analyzed by the Bardford method using BSA as the standard. The volume-weighted mean diameters D [4,3] of protein–polysaccharide conjugates were measured by dynamic light scattering (DLS). Surface hydrophobicity of the conjugates was measured by using 1-anilino-8-naphthalenesulfonate (ANS) (Sigma-Aldrich, St. Louis, MO, USA). Increasing the pH from 2 to 12 resulted in increased protein solubility. The highest solubility was 69.2% for the sample treated with ultrasonication at pH 12, while the lowest (9.13%) was observed in the Control. For the other pH conditions, the protein solubility values ranged from 40.53 to 49.65%. The ultrasound treatment significantly decreased the particle sizes of the SPI-modified starch conjugates. While the D [4,3] for the Control was 731.6 nm, it was 293.7 nm for the samples treated by sonication at pH 12. The surface hydrophobicity (H0) of SPI-starch at all pH conditions were significantly higher than those in the Control. Ultrasonication was proven to be effective in improving the solubility and emulsifying properties of soy protein isolate-starch conjugates.Keywords: particle size, solubility, soy protein isolate, ultrasonication
Procedia PDF Downloads 4223681 Natural Mexican Zeolite Modified with Iron to Remove Arsenic Ions from Water Sources
Authors: Maritza Estela Garay-Rodriguez, Mirella Gutierrez-Arzaluz, Miguel Torres-Rodriguez, Violeta Mugica-Alvarez
Abstract:
Arsenic is an element present in the earth's crust and is dispersed in the environment through natural processes and some anthropogenic activities. Naturally released into the environment through the weathering and erosion of sulphides mineral, some activities such as mining, the use of pesticides or wood preservatives potentially increase the concentration of arsenic in air, water, and soil. The natural arsenic release of a geological material is a threat to the world's drinking water sources. In aqueous phase is found in inorganic form, as arsenate and arsenite mainly, the contamination of groundwater by salts of this element originates what is known as endemic regional hydroarsenicism. The International Agency for Research on Cancer (IARC) categorizes the inorganic As within group I, as a substance with proven carcinogenic action for humans. It has been found the presence of As in groundwater in several countries such as Argentina, Mexico, Bangladesh, Canada and the United States. Regarding the concentration of arsenic in drinking water according to the World Health Organization (WHO) and the Environmental Protection Agency (EPA) establish maximum concentrations of 10 μg L⁻¹. In Mexico, in some states as Hidalgo, Morelos and Michoacán concentrations of arsenic have been found in bodies of water around 1000 μg L⁻¹, a concentration that is well above what is allowed by Mexican regulations with the NOM-127- SSA1-1994 that establishes a limit of 25 μg L⁻¹. Given this problem in Mexico, this research proposes the use of a natural Mexican zeolite (clinoptilolite type) native to the district of Etla in the central valley region of Oaxaca, as an adsorbent for the removal of arsenic. The zeolite was subjected to a conditioning with iron oxide by the precipitation-impregnation method with 0.5 M iron nitrate solution, in order to increase the natural adsorption capacity of this material. The removal of arsenic was carried out in a column with a fixed bed of conditioned zeolite, since it combines the advantages of a conventional filter with those of a natural adsorbent medium, providing a continuous treatment, of low cost and relatively easy to operate, for its implementation in marginalized areas. The zeolite was characterized by XRD, SEM/EDS, and FTIR before and after the arsenic adsorption tests, the results showed that the modification methods used are adequate to prepare adsorbent materials since it does not modify its structure, the results showed that with a particle size of 1.18 mm, an initial concentration of As (V) ions of 1 ppm, a pH of 7 and at room temperature, a removal of 98.7% was obtained with an adsorption capacity of 260 μg As g⁻¹ zeolite. The results obtained indicated that the conditioned zeolite is favorable for the elimination of arsenate in water containing up to 1000 μg As L⁻¹ and could be suitable for removing arsenate from pits of water.Keywords: adsorption, arsenic, iron conditioning, natural zeolite
Procedia PDF Downloads 1733680 Urban Planning Compilation Problems in China and the Corresponding Optimization Ideas under the Vision of the Hyper-Cycle Theory
Authors: Hong Dongchen, Chen Qiuxiao, Wu Shuang
Abstract:
Systematic science reveals the complex nonlinear mechanisms of behaviour in urban system. However, in China, when the current city planners face with the system, most of them are still taking simple linear thinking to consider the open complex giant system. This paper introduces the hyper-cycle theory, which is one of the basis theories of systematic science, based on the analysis of the reasons why the current urban planning failed, and proposals for optimization ideas that urban planning compilation should change, from controlling quantitative to the changes of relationship, from blueprint planning to progressive planning based on the nonlinear characteristics and from management control to dynamically monitor feedback.Keywords: systematic science, hyper-cycle theory, urban planning, urban management
Procedia PDF Downloads 4093679 Reverse Supply Chain Analysis of Lithium-Ion Batteries Considering Economic and Environmental Aspects
Authors: Aravind G., Arshinder Kaur, Pushpavanam S.
Abstract:
There is a strong emphasis on shifting to electric vehicles (EVs) throughout the globe for reducing the impact on global warming following the Paris climate accord. Lithium-ion batteries (LIBs) are predominantly used in EVs, and these can be a significant threat to the environment if not disposed of safely. Lithium is also a valuable resource not widely available. There are several research groups working on developing an efficient recycling process for LIBs. Two routes - pyrometallurgical and hydrometallurgical processes have been proposed for recycling LIBs. In this paper, we focus on life cycle assessment (LCA) as a tool to quantify the environmental impact of these recycling processes. We have defined the boundary of the LCA to include only the recycling phase of the end-of-life (EoL) of the battery life cycle. The analysis is done assuming ideal conditions for the hydrometallurgical and a combined hydrometallurgical and pyrometallurgical process in the inventory analysis. CML-IA method is used for quantifying the impact assessment across eleven indicators. Our results show that cathode, anode, and foil contribute significantly to the impact. The environmental impacts of both hydrometallurgical and combined recycling processes are similar across all the indicators. Further, the results of LCA are used in developing a multi-objective optimization model for the design of lithium-ion battery recycling network. Greenhouse gas emissions and cost are the two parameters minimized for the optimization study.Keywords: life cycle assessment, lithium-ion battery recycling, multi-objective optimization, network design, reverse supply chain
Procedia PDF Downloads 1573678 Decision Support System for Optimal Placement of Wind Turbines in Electric Distribution Grid
Authors: Ahmed Ouammi
Abstract:
This paper presents an integrated decision framework to support decision makers in the selection and optimal allocation of wind power plants in the electric grid. The developed approach intends to maximize the benefice related to the project investment during the planning period. The proposed decision model considers the main cost components, meteorological data, environmental impacts, operation and regulation constraints, and territorial information. The decision framework is expressed as a stochastic constrained optimization problem with the aim to identify the suitable locations and related optimal wind turbine technology considering the operational constraints and maximizing the benefice. The developed decision support system is applied to a case study to demonstrate and validate its performance.Keywords: decision support systems, electric power grid, optimization, wind energy
Procedia PDF Downloads 1533677 Optimization of Photocatalytic Degradation of Para-Nitrophenol in Visible Light by Nitrogen and Phosphorus Co-Doped Zinc Oxide Using Factorial Design of Experimental
Authors: Friday Godwin Okibe, Elaoyi David Paul, Oladayo Thomas Ojekunle
Abstract:
In this study, Nitrogen and Phosphorous co-doped Zinc Oxide (NPZ) was prepared through a solvent-free reaction. The NPZ was characterized by Scanning Electron Microscopy (SEM) and Fourier Transform Infrared (FTIR) spectroscopy. The photocatalytic activity of the catalyst was investigated by monitoring the degradation of para-nitrophenol (PNP) under visible light irradiation and the process was optimized using factorial design of experiment. The factors investigated were initial concentration of para-nitrophenol, catalyst loading, pH and irradiation time. The characterization results revealed a successful doping of ZnO by nitrogen and phosphorus and an improvement in the surface morphology of the catalyst. The photo-catalyst exhibited improved photocatalytic activity under visible light by 73.8%. The statistical analysis of the optimization result showed that the model terms were significant at 95% confidence level. Interactions plots revealed that irradiation time was the most significant factor affecting the degradation process. The cube plots of the interactions of the variables showed that an optimum degradation efficiency of 66.9% was achieved at 10mg/L initial PNP concentration, 0.5g catalyst loading, pH 7 and 150 minutes irradiation time.Keywords: nitrogen and phosphorous co-doped Zno, p-nitrophenol, photocatalytic degradation, optimization, factorial design of experimental
Procedia PDF Downloads 5273676 Study the Effect of Rubbery Phase on Morphology Development of PP/PA6/(EPDM:EPDM-g-MA) Ternary Blends
Authors: B. Afsari, M. Hassanpour, M. Shabani
Abstract:
This study aimed to investigate the phase morphology of ternary blends comprising PP, PA6, and a blend of EPDM and EPDM-g-MA in a 70/15/15 ratio. Varying ratios of EPDM to EPDM-g-MA were examined. As the proportion of EPDM-g-MA increased, an interlayer phase formed between the dispersed PA6 domains and the PP matrix. This resulted in the development of a core-shell encapsulation morphology within the blends. The concentration of the EPDM-g-MA component is inversely correlated with the average size of PA6 particles. Additionally, blends containing higher proportions of the EPDM-g-MA rubbery phase exhibited an aggregated structure of the modifier particles. Notably, as the concentration of EPDM-g-MA increased from 0% to 15% in the blend, there was a consistent monotonic reduction in the size of PA6 particles.Keywords: phase morphology, rubbery phase, rubber functionality, ternary blends
Procedia PDF Downloads 913675 Artificial Intelligent-Based Approaches for Task Offloading, Resource Allocation and Service Placement of Internet of Things Applications: State of the Art
Authors: Fatima Z. Cherhabil, Mammar Sedrati, Sonia-Sabrina Bendib
Abstract:
In order to support the continued growth, critical latency of IoT applications, and various obstacles of traditional data centers, mobile edge computing (MEC) has emerged as a promising solution that extends cloud data-processing and decision-making to edge devices. By adopting a MEC structure, IoT applications could be executed locally, on an edge server, different fog nodes, or distant cloud data centers. However, we are often faced with wanting to optimize conflicting criteria such as minimizing energy consumption of limited local capabilities (in terms of CPU, RAM, storage, bandwidth) of mobile edge devices and trying to keep high performance (reducing response time, increasing throughput and service availability) at the same time. Achieving one goal may affect the other, making task offloading (TO), resource allocation (RA), and service placement (SP) complex processes. It is a nontrivial multi-objective optimization problem to study the trade-off between conflicting criteria. The paper provides a survey on different TO, SP, and RA recent multi-objective optimization (MOO) approaches used in edge computing environments, particularly artificial intelligent (AI) ones, to satisfy various objectives, constraints, and dynamic conditions related to IoT applications.Keywords: mobile edge computing, multi-objective optimization, artificial intelligence approaches, task offloading, resource allocation, service placement
Procedia PDF Downloads 1173674 Use of Nanosensors in Detection and Treatment of HIV
Authors: Sayed Obeidullah Abrar
Abstract:
Nanosensor is the combination of two terms nanoparticles and sensors. These are chemical or physical sensor constructed using nanoscale components, usually microscopic or submicroscopic in size. These sensors are very sensitive and can detect single virus particle or even very low concentrations of substances that could be potentially harmful. Nanosensors have a large scope of research especially in the field of medical sciences, military applications, pharmaceuticals etc.Keywords: HIV/AIDS, nanosensors, DNA, RNA
Procedia PDF Downloads 2993673 Curriculum System Optimization under Outstanding Engineers Training Mode of Mechanical and Electronic Engineering
Authors: El Miloudi Djelloul
Abstract:
Teaching program of `A plan for educating and training outstanding engineers' is divided into intramural teaching program and enterprise practice teaching program. Based on analyzing the basic principles of teaching plans which teaching plan follows for undergraduate mechanical and electrical engineering, major contents of specialty teaching project are studied amply. The study contents include the system optimization and reform of common curriculum, specialty curriculum and practice curriculum. The practice indicated that under outstanding engineers training mode, the optimized curriculum system have practicability, and achieve the training objectives.Keywords: curriculum system, mechanical and electronic engineering, outstanding engineers, teaching program
Procedia PDF Downloads 5303672 Regularizing Software for Aerosol Particles
Authors: Christine Böckmann, Julia Rosemann
Abstract:
We present an inversion algorithm that is used in the European Aerosol Lidar Network for the inversion of data collected with multi-wavelength Raman lidar. These instruments measure backscatter coefficients at 355, 532, and 1064 nm, and extinction coefficients at 355 and 532 nm. The algorithm is based on manually controlled inversion of optical data which allows for detailed sensitivity studies and thus provides us with comparably high quality of the derived data products. The algorithm allows us to derive particle effective radius, volume, surface-area concentration with comparably high confidence. The retrieval of the real and imaginary parts of the complex refractive index still is a challenge in view of the accuracy required for these parameters in climate change studies in which light-absorption needs to be known with high accuracy. Single-scattering albedo (SSA) can be computed from the retrieve microphysical parameters and allows us to categorize aerosols into high and low absorbing aerosols. From mathematical point of view the algorithm is based on the concept of using truncated singular value decomposition as regularization method. This method was adapted to work for the retrieval of the particle size distribution function (PSD) and is called hybrid regularization technique since it is using a triple of regularization parameters. The inversion of an ill-posed problem, such as the retrieval of the PSD, is always a challenging task because very small measurement errors will be amplified most often hugely during the solution process unless an appropriate regularization method is used. Even using a regularization method is difficult since appropriate regularization parameters have to be determined. Therefore, in a next stage of our work we decided to use two regularization techniques in parallel for comparison purpose. The second method is an iterative regularization method based on Pade iteration. Here, the number of iteration steps serves as the regularization parameter. We successfully developed a semi-automated software for spherical particles which is able to run even on a parallel processor machine. From a mathematical point of view, it is also very important (as selection criteria for an appropriate regularization method) to investigate the degree of ill-posedness of the problem which we found is a moderate ill-posedness. We computed the optical data from mono-modal logarithmic PSD and investigated particles of spherical shape in our simulations. We considered particle radii as large as 6 nm which does not only cover the size range of particles in the fine-mode fraction of naturally occurring PSD but also covers a part of the coarse-mode fraction of PSD. We considered errors of 15% in the simulation studies. For the SSA, 100% of all cases achieve relative errors below 12%. In more detail, 87% of all cases for 355 nm and 88% of all cases for 532 nm are well below 6%. With respect to the absolute error for non- and weak-absorbing particles with real parts 1.5 and 1.6 in all modes the accuracy limit +/- 0.03 is achieved. In sum, 70% of all cases stay below +/-0.03 which is sufficient for climate change studies.Keywords: aerosol particles, inverse problem, microphysical particle properties, regularization
Procedia PDF Downloads 3433671 Design and Optimization of Sustainable Buildings by Combined Cooling, Heating and Power System (CCHP) Based on Exergy Analysis
Authors: Saeed Karimi, Ali Behbahaninia
Abstract:
In this study, the design and optimization of combined cooling, heating, and power system (CCHP) for a sustainable building are dealt with. Sustainable buildings are environmentally responsible and help us to save energy also reducing waste, pollution and environmental degradation. CCHP systems are widely used to save energy sources. In these systems, electricity, cooling, and heating are generating using just one primary energy source. The selection of the size of components based on the maximum demand of users will lead to an increase in the total cost of energy and equipment for the building complex. For this purpose, a system was designed in which the prime mover (gas turbine), heat recovery boiler, and absorption chiller are lower than the needed maximum. The difference in months with peak consumption is supplied with the help of electrical absorption chiller and auxiliary boiler (and the national electricity network). In this study, the optimum capacities of each of the equipment are determined based on Thermo economic method, in a way that the annual capital cost and energy consumption will be the lowest. The design was done for a gas turbine prime mover, and finally, the optimum designs were investigated using exergy analysis and were compared with a traditional energy supply system.Keywords: sustainable building, CCHP, energy optimization, gas turbine, exergy, thermo-economic
Procedia PDF Downloads 943670 Optimization of the Conditions of Oligomerization and Polymerization Processes of Selected Olefins with the Use of Complex Compounds of Transition Metal Ions
Authors: Joanna Drzeżdżon, Marzena Białek
Abstract:
Polyolefins are a group of materials used today in all areas of life. They are used in the food, domestic and other industries. In particular, polyethylene and polypropylene have found application in the production of packaging materials, pipes, containers, car parts as well as elements of medical equipment, e.g. syringes. Optimization of the polymerization and oligomerization processes of selected olefins is a very important stage before the technological implementation of polyolefin production. The purpose of the studies is to determine the conditions for ethylene polymerization as well as 3-buten-2-ol and 2-chloro-2-propen-1-ol oligomerization with the use of oxovanadium(IV) dipicolinate complexes with N-heterocyclic ligands. Additionally, the studies aims to determine the catalytic activities of the dipicolinate oxovanadium(IV) complexes with N-heterocyclic ligands in the studied polymerization and oligomerization processes.Keywords: buten-2-ol, dipicolinate, ethylene, polymerization, oligomerization, vanadium
Procedia PDF Downloads 1963669 Multi-Objective Optimization of Wear Parameters of Tube Like Clay Mineral Filled Thermoplastic Polymer Using Response Surface Methodology
Authors: Vasu Velagapudi, G. Suresh
Abstract:
PTFE/HNTs nanocomposites are fabricated with 4%, 6%, and 8% by weight fraction, and the optimization study of wear parameters are performed using response surface methodology (RSM). The experiments are carried out on a pin on disc (POD) wear tester under different operating parameters planned according to Taguchi L27 orthogonal array. The input factors considered are wt% HNTs addition, sliding velocity, load, and distance with three levels for each factor. From ANOVA: The factors load, speed and distance and their interactions have a significant effect on COF. Also for SWR, composition factor and interaction of load and speed are observed to be significant ( < 0.05) Optimum input parameters corresponding to desirability 1 are found to be: COF (0.11) and SWR (17.5)×10⁻⁶ (mm3/N-m) at 6.34 wt% of composition, 5N of load, 2 km of distance and 1 m/sec of velocity.Keywords: PTFE/HNT, nanocomposites, response surface methodology (RSM), specific wear rate
Procedia PDF Downloads 3963668 Mapping Tunnelling Parameters for Global Optimization in Big Data via Dye Laser Simulation
Authors: Sahil Imtiyaz
Abstract:
One of the biggest challenges has emerged from the ever-expanding, dynamic, and instantaneously changing space-Big Data; and to find a data point and inherit wisdom to this space is a hard task. In this paper, we reduce the space of big data in Hamiltonian formalism that is in concordance with Ising Model. For this formulation, we simulate the system using dye laser in FORTRAN and analyse the dynamics of the data point in energy well of rhodium atom. After mapping the photon intensity and pulse width with energy and potential we concluded that as we increase the energy there is also increase in probability of tunnelling up to some point and then it starts decreasing and then shows a randomizing behaviour. It is due to decoherence with the environment and hence there is a loss of ‘quantumness’. This interprets the efficiency parameter and the extent of quantum evolution. The results are strongly encouraging in favour of the use of ‘Topological Property’ as a source of information instead of the qubit.Keywords: big data, optimization, quantum evolution, hamiltonian, dye laser, fermionic computations
Procedia PDF Downloads 1953667 A Comparison of Biosorption of Radionuclides Tl-201 on Different Biosorbents and Their Empirical Modelling
Authors: Sinan Yapici, Hayrettin Eroglu
Abstract:
The discharge of the aqueous radionuclides wastes used for the diagnoses of diseases and treatments of patients in nuclear medicine can cause fatal health problems when the radionuclides and its stable daughter component mix with underground water. Tl-201, which is one of the radionuclides commonly used in the nuclear medicine, is a toxic substance and is converted to its stable daughter component Hg-201, which is also a poisonous heavy metal: Tl201 → Hg201 + Gamma Ray [135-167 Kev (12%)] + X Ray [69-83 Kev (88%)]; t1/2 = 73,1 h. The purpose of the present work was to remove Tl-201 radionuclides from aqueous solution by biosorption on the solid bio wastes of food and cosmetic industry as bio sorbents of prina from an olive oil plant, rose residue from a rose oil plant and tea residue from a tea plant, and to make a comparison of the biosorption efficiencies. The effects of the biosorption temperature, initial pH of the aqueous solution, bio sorbent dose, particle size and stirring speed on the biosorption yield were investigated in a batch process. It was observed that the biosorption is a rapid process with an equilibrium time less than 10 minutes for all the bio sorbents. The efficiencies were found to be close to each other and measured maximum efficiencies were 93,30 percent for rose residue, 94,1 for prina and 98,4 for tea residue. In a temperature range of 283 and 313 K, the adsorption decreased with increasing temperature almost in a similar way. In a pH range of 2-10, increasing pH enhanced biosorption efficiency up to pH=7 and then the efficiency remained constant in a similar path for all the biosorbents. Increasing stirring speed from 360 to 720 rpm enhanced slightly the biosorption efficiency almost at the same ratio for all bio sorbents. Increasing particle size decreased the efficiency for all biosorbent; however the most negatively effected biosorbent was prina with a decrease in biosorption efficiency from about 84 percent to 40 with an increase in the nominal particle size 0,181 mm to 1,05 while the least effected one, tea residue, went down from about 97 percent to 87,5. The biosorption efficiencies of all the bio sorbents increased with increasing biosorbent dose in the range of 1,5 to 15,0 g/L in a similar manner. The fit of the experimental results to the adsorption isotherms proved that the biosorption process for all the bio sorbents can be represented best by Freundlich model. The kinetic analysis showed that all the processes fit very well to pseudo second order rate model. The thermodynamics calculations gave ∆G values between -8636 J mol-1 and -5378 for tea residue, -5313 and -3343 for rose residue, and -5701 and -3642 for prina with a ∆H values of -39516 J mol-1, -23660 and -26190, and ∆S values of -108.8 J mol-1 K-1, -64,0, -72,0 respectively, showing spontaneous and exothermic character of the processes. An empirical biosorption model in the following form was derived for each biosorbent as function of the parameters and time, taking into account the form of kinetic model, with regression coefficients over 0.9990 where At is biosorbtion efficiency at any time and Ae is the equilibrium efficiency, t is adsorption period as s, ko a constant, pH the initial acidity of biosorption medium, w the stirring speed as s-1, S the biosorbent dose as g L-1, D the particle size as m, and a, b, c, and e are the powers of the parameters, respectively, E a constant containing activation energy and T the temperature as K.Keywords: radiation, diosorption, thallium, empirical modelling
Procedia PDF Downloads 265