Search results for: optimized closed polygonal segment method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20765

Search results for: optimized closed polygonal segment method

17615 Analysis of Heat Exchanger Area of Two Stage Cascade Refrigeration System Using Taguchi Methodology

Authors: A. D. Parekh

Abstract:

The present work describes relative contributions of operating parameters on required heat transfer area of three heat exchangers viz. evaporator, condenser and cascade condenser of two stage R404A-R508B cascade refrigeration system using Taguchi method. The operating parameters considered in present study includes (1) condensing temperature of high temperature cycle and low temperature cycle (2) evaporating temperature of low temperature cycle (3) degree of superheating in low temperature cycle (4) refrigerating effect. Heat transfer areas of three heat exchangers are studied with variation of above operating parameters and also optimum working levels of each operating parameter has been obtained for minimum heat transfer area of each heat exchanger using Taguchi method. The analysis using Taguchi method reveals that evaporating temperature of low temperature cycle and refrigerating effect contribute relatively largely on the area of evaporator. Condenser area is mainly influenced by both condensing temperature of high temperature cycle and refrigerating effect. Area of cascade condenser is mainly affected by refrigerating effect and the effects of other operating parameters are minimal.

Keywords: cascade refrigeration system, Taguchi method, heat transfer area, ANOVA, optimal solution

Procedia PDF Downloads 377
17614 Computational Analysis of the Scaling Effects on the Performance of an Axial Compressor

Authors: Junting Xiang, Jörg Uwe Schlüter, Fei Duan

Abstract:

The miniaturization of gas turbines promises many advantages. Miniature gas turbines can be used for local power generation or the propulsion of small aircraft, such as UAV and MAV. However, experience shows that the miniaturization of conventional gas turbines, which are optimized at their current large size, leads to a substantial loss of efficiency and performance at smaller scales. This may be due to a number of factors, such as the Reynolds-number effect, the increased heat transfer, and manufacturing tolerances. In the present work, we focus on computational investigations of the Reynolds number effect and the wall heat transfer on the performance of axial compressor during its size change. The NASA stage 35 compressors are selected as the configuration in this study and Computational Fluid Dynamics (CFD) is used to carry out the miniaturization process and simulations. We perform parameter studies on the effect of Reynolds number and wall thermal conditions. Our results indicate a decrease of efficiency, if the compressor is miniaturized based on its original geometry due to the increase of viscous effects. The increased heat transfer through wall has only a small effect and will actually benefit compressor performance based on our study.

Keywords: axial compressor, CFD, heat transfer, miniature gas turbines, Reynolds number

Procedia PDF Downloads 409
17613 Optimal Scheduling of Load and Operational Strategy of a Load Aggregator to Maximize Profit with PEVs

Authors: Md. Shafiullah, Ali T. Al-Awami

Abstract:

This project proposes optimal scheduling of imported power of a load aggregator with the utilization of EVs to maximize its profit. As with the increase of renewable energy resources, electricity price in competitive market becomes more uncertain and, on the other hand, with the penetration of renewable distributed generators in the distribution network the predicted load of a load aggregator also becomes uncertain in real time. Though there is uncertainties in both load and price, the use of EVs storage capacity can make the operation of load aggregator flexible. LA submits its offer to day-ahead market based on predicted loads and optimized use of its EVs to maximize its profit, as well as in real time operation it uses its energy storage capacity in such a way that it can maximize its profit. In this project, load aggregators profit maximization algorithm is formulated and the optimization problem is solved with the help of CVX. As in real time operation the forecasted loads differ from actual load, the mismatches are settled in real time balancing market. Simulation results compare the profit of a load aggregator with a hypothetical group of 1000 EVs and without EVs.

Keywords: CVX, electricity market, load aggregator, load and price uncertainties, profit maximization, real time balancing operation

Procedia PDF Downloads 409
17612 Proposing an Algorithm to Cluster Ad Hoc Networks, Modulating Two Levels of Learning Automaton and Nodes Additive Weighting

Authors: Mohammad Rostami, Mohammad Reza Forghani, Elahe Neshat, Fatemeh Yaghoobi

Abstract:

An Ad Hoc network consists of wireless mobile equipment which connects to each other without any infrastructure, using connection equipment. The best way to form a hierarchical structure is clustering. Various methods of clustering can form more stable clusters according to nodes' mobility. In this research we propose an algorithm, which allocates some weight to nodes based on factors, i.e. link stability and power reduction rate. According to the allocated weight in the previous phase, the cellular learning automaton picks out in the second phase nodes which are candidates for being cluster head. In the third phase, learning automaton selects cluster head nodes, member nodes and forms the cluster. Thus, this automaton does the learning from the setting and can form optimized clusters in terms of power consumption and link stability. To simulate the proposed algorithm we have used omnet++4.2.2. Simulation results indicate that newly formed clusters have a longer lifetime than previous algorithms and decrease strongly network overload by reducing update rate.

Keywords: mobile Ad Hoc networks, clustering, learning automaton, cellular automaton, battery power

Procedia PDF Downloads 404
17611 Modeling of Crack Propagation Path in Concrete with Coarse Trapezoidal Aggregates by Boundary Element Method

Authors: Chong Wang, Alexandre Urbano Hoffmann

Abstract:

Interaction between a crack and a trapezoidal aggregate in a single edge notched concrete beam is simulated using boundary element method with an automatic crack extension program. The stress intensity factors of the growing crack are obtained from the J-integral. Three crack extension paths: deflecting around the particulate, growing along the interface and penetrating into the particulate are achieved in terms of the mismatch state of mechanical characteristics of matrix and the particulate. The toughening is also given by the ratio of stress intensity factors. The results reveal that as stress shielding occurs, toughening is obtained when the crack is approaching to a stiff and strong aggregate weakly bonded to a relatively soft matrix. The present work intends to help for the design of aggregate reinforced concretes.

Keywords: aggregate concrete, boundary element method, two-phase composite, crack extension path, crack/particulate interaction

Procedia PDF Downloads 422
17610 Parametric Analysis and Optimal Design of Functionally Graded Plates Using Particle Swarm Optimization Algorithm and a Hybrid Meshless Method

Authors: Foad Nazari, Seyed Mahmood Hosseini, Mohammad Hossein Abolbashari, Mohammad Hassan Abolbashari

Abstract:

The present study is concerned with the optimal design of functionally graded plates using particle swarm optimization (PSO) algorithm. In this study, meshless local Petrov-Galerkin (MLPG) method is employed to obtain the functionally graded (FG) plate’s natural frequencies. Effects of two parameters including thickness to height ratio and volume fraction index on the natural frequencies and total mass of plate are studied by using the MLPG results. Then the first natural frequency of the plate, for different conditions where MLPG data are not available, is predicted by an artificial neural network (ANN) approach which is trained by back-error propagation (BEP) technique. The ANN results show that the predicted data are in good agreement with the actual one. To maximize the first natural frequency and minimize the mass of FG plate simultaneously, the weighted sum optimization approach and PSO algorithm are used. However, the proposed optimization process of this study can provide the designers of FG plates with useful data.

Keywords: optimal design, natural frequency, FG plate, hybrid meshless method, MLPG method, ANN approach, particle swarm optimization

Procedia PDF Downloads 361
17609 Analysis of Ancient and Present Lightning Protection Systems of Large Heritage Stupas in Sri Lanka

Authors: J.R.S.S. Kumara, M.A.R.M. Fernando, S.Venkatesh, D.K. Jayaratne

Abstract:

Protection of heritage monuments against lightning has become extremely important as far as their historical values are concerned. When such structures are large and tall, the risk of lightning initiated from both cloud and ground can be high. This paper presents a lightning risk analysis of three giant stupas in Anuradhapura era (fourth century BC onwards) in Sri Lanka. The three stupas are Jethawaaramaya (269-296 AD), Abayagiriya (88-76 BC) and Ruwanweliseya (161-137 BC), the third, fifth and seventh largest ancient structures in the world. These stupas are solid brick structures consisting of a base, a near hemispherical dome and a conical spire on the top. The ancient stupas constructed with a dielectric crystal on the top and connected to the ground through a conducting material, was considered as the hypothesis for their original lightning protection technique. However, at present, all three stupas are protected with Franklin rod type air termination systems located on top of the spire. First, a risk analysis was carried out according to IEC 62305 by considering the isokeraunic level of the area and the height of the stupas. Then the standard protective angle method and rolling sphere method were used to locate the possible touching points on the surface of the stupas. The study was extended to estimate the critical current which could strike on the unprotected areas of the stupas. The equations proposed by (Uman 2001) and (Cooray2007) were used to find the striking distances. A modified version of rolling sphere method was also applied to see the effects of upward leaders. All these studies were carried out for two scenarios: with original (i.e. ancient) lightning protection system and with present (i.e. new) air termination system. The field distribution on the surface of the stupa in the presence of a downward leader was obtained using finite element based commercial software COMSOL Multiphysics for further investigations of lightning risks. The obtained results were analyzed and compared each other to evaluate the performance of ancient and new lightning protection methods and identify suitable methods to design lightning protection systems for stupas. According to IEC standards, all three stupas with new and ancient lightning protection system has Level IV protection as per protection angle method. However according to rolling sphere method applied with Uman’s equation protection level is III. The same method applied with Cooray’s equation always shows a high risk with respect to Uman’s equation. It was found that there is a risk of lightning strikes on the dome and square chamber of the stupa, and the corresponding critical current values were different with respect to the equations used in the rolling sphere method and modified rolling sphere method.

Keywords: Stupa, heritage, lightning protection, rolling sphere method, protection level

Procedia PDF Downloads 236
17608 Improving Detection of Illegitimate Scores and Assessment in Most Advantageous Tenders

Authors: Hao-Hsi Tseng, Hsin-Yun Lee

Abstract:

The Most Advantageous Tender (MAT) has been criticized for its susceptibility to dictatorial situations and for its processing of same score, same rank issues. This study applies the four criteria from Arrow's Impossibility Theorem to construct a mechanism for revealing illegitimate scores in scoring methods. While commonly be used to improve on problems resulting from extreme scores, ranking methods hide significant defects, adversely affecting selection fairness. To address these shortcomings, this study relies mainly on the overall evaluated score method, using standardized scores plus normal cumulative distribution function conversion to calculate the evaluation of vender preference. This allows for free score evaluations, which reduces the influence of dictatorial behavior and avoiding same score, same rank issues. Large-scale simulations confirm that this method outperforms currently used methods using the Impossibility Theorem.

Keywords: Arrow’s impossibility theorem, cumulative normal distribution function, most advantageous tender, scoring method

Procedia PDF Downloads 458
17607 Study of Electron Cyclotron Resonance Acceleration by Cylindrical TE₀₁₁ Mode

Authors: Oswaldo Otero, Eduardo A. Orozco, Ana M. Herrera

Abstract:

In this work, we present results from analytical and numerical studies of the electron acceleration by a TE₀₁₁ cylindrical microwave mode in a static homogeneous magnetic field under electron cyclotron resonance (ECR) condition. The stability of the orbits is analyzed using the particle orbit theory. In order to get a better understanding of the interaction wave-particle, we decompose the azimuthally electric field component as the superposition of right and left-hand circular polarization standing waves. The trajectory, energy and phase-shift of the electron are found through a numerical solution of the relativistic Newton-Lorentz equation in a finite difference method by the Boris method. It is shown that an electron longitudinally injected with an energy of 7 keV in a radial position r=Rc/2, being Rc the cavity radius, is accelerated up to energy of 90 keV by an electric field strength of 14 kV/cm and frequency of 2.45 GHz. This energy can be used to produce X-ray for medical imaging. These results can be used as a starting point for study the acceleration of electrons in a magnetic field changing slowly in time (GYRAC), which has some important applications as the electron cyclotron resonance ion proton accelerator (ECR-IPAC) for cancer therapy and to control plasma bunches with relativistic electrons.

Keywords: Boris method, electron cyclotron resonance, finite difference method, particle orbit theory, X-ray

Procedia PDF Downloads 154
17606 The Analysis of Own Signals of PM Electrical Machines – Example of Eccentricity

Authors: Marcin Baranski

Abstract:

This article presents a vibration diagnostic method designed for permanent magnets (PM) traction motors. Those machines are commonly used in traction drives of electrical vehicles. Specific structural properties of machines excited by permanent magnets are used in this method - electromotive force (EMF) generated due to vibrations. This work presents: field-circuit model, results of static tests, results of calculations and simulations.

Keywords: electrical vehicle, permanent magnet, traction drive, vibrations, electrical machine, eccentricity

Procedia PDF Downloads 619
17605 Cheiloscopy: A Study on Predominant Lip Print Patterns among the Gujarati Population

Authors: Pooja Ahuja, Tejal Bhutani, M. S. Dahiya

Abstract:

Cheiloscopy, the study of lip prints, is a tool in forensic investigation technique that deals with identification of individuals based on lips patterns. The objective of this study is to determine predominant lip print pattern found among the Gujarati population, to evaluate whether any sex difference exists and to study the permanence of the pattern over six months duration. The study comprised of 100 healthy individuals (50 males and 50 females), in the age group of 18 to 25 years of Gujarati population of the Gandhinagar region of the Gujarat state, India. By using Suzuki and Tsuchihashi classification, Lip prints were then divided into four quadrants and also classified on the basis of peripheral shape of the lips. Materials used to record the lip prints were dark brown colored lipstick, cellophane tape, and white bond paper. Lipstick was applied uniformly, and lip prints were taken on the glued portion of cellophane tape and then stuck on to a white bond paper. These lip prints were analyzed with magnifying lens and virtually with stereo microscope. On the analysis of the subject population, results showed Branched pattern Type II (29.57 percentage) to be most predominant in the Gujarati population. Branched pattern Type II (35.60 percentage) and long vertical Type I (28.28 percentage) were most prevalent in males and females respectively and large full lips were most predominantly present in both the sexes. The study concludes that lip prints in any form can be an effective tool for identification of an individual in a closed or open group forms.

Keywords: cheiloscopy, lip pattern, predomianant, Gujarati population

Procedia PDF Downloads 292
17604 Mooring Analysis of Duct-Type Tidal Current Power System in Shallow Water

Authors: Chul H. Jo, Do Y. Kim, Bong K. Cho, Myeong J. Kim

Abstract:

The exhaustion of oil and the environmental pollution from the use of fossil fuel are increasing. Tidal current power (TCP) has been proposed as an alternative energy source because of its predictability and reliability. By applying a duct and single point mooring (SPM) system, a TCP device can amplify the generating power and keep its position properly. Because the generating power is proportional to cube of the current stream velocity, amplifying the current speed by applying a duct to a TCP system is an effective way to improve the efficiency of the power device. An SPM system can be applied at any water depth and is highly cost effective. Simple installation and maintenance procedures are also merits of an SPM system. In this study, we designed an SPM system for a duct-type TCP device for use in shallow water. Motions of the duct are investigated to obtain the response amplitude operator (RAO) as the magnitude of the transfer function. Parameters affecting the stability of the SPM system such as the fairlead departure angle, current velocity, and the number of clamp weights are analyzed and/or optimized. Wadam and OrcaFlex commercial software is used to design the mooring line.

Keywords: mooring design, parametric analysis, RAO (Response Amplitude Operator), SPM (Single Point Mooring)

Procedia PDF Downloads 283
17603 Integrating Artificial Neural Network and Taguchi Method on Constructing the Real Estate Appraisal Model

Authors: Mu-Yen Chen, Min-Hsuan Fan, Chia-Chen Chen, Siang-Yu Jhong

Abstract:

In recent years, real estate prediction or valuation has been a topic of discussion in many developed countries. Improper hype created by investors leads to fluctuating prices of real estate, affecting many consumers to purchase their own homes. Therefore, scholars from various countries have conducted research in real estate valuation and prediction. With the back-propagation neural network that has been popular in recent years and the orthogonal array in the Taguchi method, this study aimed to find the optimal parameter combination at different levels of orthogonal array after the system presented different parameter combinations, so that the artificial neural network obtained the most accurate results. The experimental results also demonstrated that the method presented in the study had a better result than traditional machine learning. Finally, it also showed that the model proposed in this study had the optimal predictive effect, and could significantly reduce the cost of time in simulation operation. The best predictive results could be found with a fewer number of experiments more efficiently. Thus users could predict a real estate transaction price that is not far from the current actual prices.

Keywords: artificial neural network, Taguchi method, real estate valuation model, investors

Procedia PDF Downloads 480
17602 Comparison of Filamentous Fungus (Monascus purpureus)Growth in Submerged and Solid State Culture

Authors: Shafieeh Mansoori, Fatemeh Yazdian, Ashrafsadat Hatamian, Majid Azizi

Abstract:

Monascus purpureus, which has a special metabolite with many therapeutic and medicinal properties including antioxidant, antibiotic, anti-hypercholesterolemia, and immunosuppressive properties, is a traditional Chinese fermentation fungus and is used as a natural dietary supplement. Production of desired metabolites actually determined by optimized growth which is supported by some factors such as substrates and Monascus strains type, moisture content of the fermentation mixture, aeration, and control of contamination issues. In this experiment, M. purpureus PTCC5305 was cultured in both the liquid and solid culture medium. The former medium contain YMP (yeast extract, maltose and peptone), PGC (peptone, glucose complex), and GYP (glucose, yeast extract and peptone) medium. After 8 days, the best medium for the cell production was PGC agar medium on solid culture with 0.28 g dry weight of cell mass whereas the best liquid culture was GYP medium with 3.5 g/l dry weight of cell mass. The lowest cell production was on YMP agar with 0.1 g dry weight of cell mass and then YMP medium with 2.5 g/l dry cell weight.

Keywords: Monascus purpureus, solid state fermentation, submerged culture, Chinese fermentation fungus

Procedia PDF Downloads 402
17601 Efficient Principal Components Estimation of Large Factor Models

Authors: Rachida Ouysse

Abstract:

This paper proposes a constrained principal components (CnPC) estimator for efficient estimation of large-dimensional factor models when errors are cross sectionally correlated and the number of cross-sections (N) may be larger than the number of observations (T). Although principal components (PC) method is consistent for any path of the panel dimensions, it is inefficient as the errors are treated to be homoskedastic and uncorrelated. The new CnPC exploits the assumption of bounded cross-sectional dependence, which defines Chamberlain and Rothschild’s (1983) approximate factor structure, as an explicit constraint and solves a constrained PC problem. The CnPC method is computationally equivalent to the PC method applied to a regularized form of the data covariance matrix. Unlike maximum likelihood type methods, the CnPC method does not require inverting a large covariance matrix and thus is valid for panels with N ≥ T. The paper derives a convergence rate and an asymptotic normality result for the CnPC estimators of the common factors. We provide feasible estimators and show in a simulation study that they are more accurate than the PC estimator, especially for panels with N larger than T, and the generalized PC type estimators, especially for panels with N almost as large as T.

Keywords: high dimensionality, unknown factors, principal components, cross-sectional correlation, shrinkage regression, regularization, pseudo-out-of-sample forecasting

Procedia PDF Downloads 145
17600 Spatial Data Science for Data Driven Urban Planning: The Youth Economic Discomfort Index for Rome

Authors: Iacopo Testi, Diego Pajarito, Nicoletta Roberto, Carmen Greco

Abstract:

Today, a consistent segment of the world’s population lives in urban areas, and this proportion will vastly increase in the next decades. Therefore, understanding the key trends in urbanization, likely to unfold over the coming years, is crucial to the implementation of sustainable urban strategies. In parallel, the daily amount of digital data produced will be expanding at an exponential rate during the following years. The analysis of various types of data sets and its derived applications have incredible potential across different crucial sectors such as healthcare, housing, transportation, energy, and education. Nevertheless, in city development, architects and urban planners appear to rely mostly on traditional and analogical techniques of data collection. This paper investigates the prospective of the data science field, appearing to be a formidable resource to assist city managers in identifying strategies to enhance the social, economic, and environmental sustainability of our urban areas. The collection of different new layers of information would definitely enhance planners' capabilities to comprehend more in-depth urban phenomena such as gentrification, land use definition, mobility, or critical infrastructural issues. Specifically, the research results correlate economic, commercial, demographic, and housing data with the purpose of defining the youth economic discomfort index. The statistical composite index provides insights regarding the economic disadvantage of citizens aged between 18 years and 29 years, and results clearly display that central urban zones and more disadvantaged than peripheral ones. The experimental set up selected the city of Rome as the testing ground of the whole investigation. The methodology aims at applying statistical and spatial analysis to construct a composite index supporting informed data-driven decisions for urban planning.

Keywords: data science, spatial analysis, composite index, Rome, urban planning, youth economic discomfort index

Procedia PDF Downloads 126
17599 Recirculated Sedimentation Method to Control Contamination for Algal Biomass Production

Authors: Ismail S. Bostanci, Ebru Akkaya

Abstract:

Microalgae-derived biodiesel, fertilizer or industrial chemicals' production with wastewater has great potential. Especially water from a municipal wastewater treatment plant is a very important nutrient source for biofuel production. Microalgae biomass production in open ponds system is lower cost culture systems. There are many hurdles for commercial algal biomass production in large scale. One of the important technical bottlenecks for microalgae production in open system is culture contamination. The algae culture contaminants can generally be described as invading organisms which could cause pond crash. These invading organisms can be competitors, parasites, and predators. Contamination is unavoidable in open systems. Potential contaminant organisms are already inoculated if wastewater is utilized for algal biomass cultivation. Especially, it is important to control contaminants to retain in acceptable level in order to reach true potential of algal biofuel production. There are several contamination management methods in algae industry, ranging from mechanical, chemical, biological and growth condition change applications. However, none of them are accepted as a suitable contamination control method. This experiment describes an innovative contamination control method, 'Recirculated Sedimentation Method', to manage contamination to avoid pond cash. The method can be used for the production of algal biofuel, fertilizer etc. and algal wastewater treatment. To evaluate the performance of the method on algal culture, an experiment was conducted for 90 days at a lab-scale raceway (60 L) reactor with the use of non-sterilized and non-filtered wastewater (secondary effluent and centrate of anaerobic digestion). The application of the method provided the following; removing contaminants (predators and diatoms) and other debris from reactor without discharging the culture (with microscopic evidence), increasing raceway tank’s suspended solids holding capacity (770 mg L-1), increasing ammonium removal rate (29.83 mg L-1 d-1), decreasing algal and microbial biofilm formation on inner walls of reactor, washing out generated nitrifier from reactor to prevent ammonium consumption.

Keywords: contamination control, microalgae culture contamination, pond crash, predator control

Procedia PDF Downloads 199
17598 Improved Multi-Channel Separation Algorithm for Satellite-Based Automatic Identification System Signals Based on Artificial Bee Colony and Adaptive Moment Estimation

Authors: Peng Li, Luan Wang, Haifeng Fei, Renhong Xie, Yibin Rui, Shanhong Guo

Abstract:

The applications of satellite-based automatic identification system (S-AIS) pave the road for wide-range maritime traffic monitoring and management. But the coverage of satellite’s view includes multiple AIS self-organizing networks, which leads to the collision of AIS signals from different cells. The contribution of this work is to propose an improved multi-channel blind source separation algorithm based on Artificial Bee Colony (ABC) and advanced stochastic optimization to perform separation of the mixed AIS signals. The proposed approach adopts modified ABC algorithm to get an optimized initial separating matrix, which can expedite the initialization bias correction, and utilizes the Adaptive Moment Estimation (Adam) to update the separating matrix by adjusting the learning rate for each parameter dynamically. Simulation results show that the algorithm can speed up convergence and lead to better performance in separation accuracy.

Keywords: satellite-based automatic identification system, blind source separation, artificial bee colony, adaptive moment estimation

Procedia PDF Downloads 180
17597 Fructooligosaccharide Prebiotics: Optimization of Different Cultivation Parameters on Their Microbial Production

Authors: Elsayed Ahmed Elsayed, Azza Noor El-Deen, Mohamed A. Farid, Mohamed A. Wadaan

Abstract:

Recently, a great attention has been paid to the use of dietary carbohydrates as prebiotic functional foods. Among the new commercially available products, fructooligosaccharides (FOS), which are microbial produced from sucrose, have attracted special interest due to their valuable properties and, thus, have a great economic potential for the sugar industrial branch. They are non-cariogenic sweeteners of low caloric value, as they are not hydrolyzed by the gastro-intestinal enzymes, promoting selectively the growth of the bifidobacteria in the colon, helping to eliminate the harmful microbial species to human and animal health and preventing colon cancer. FOS has been also found to reduce cholesterol, phospholipids and triglyceride levels in blood. FOS has been mainly produced by microbial fructosyltransferase (FTase) enzymes. The present work outlines bioprocess optimization for different cultivation parameters affecting the production of FTase by Penicillium aurantiogriseum AUMC 5605. The optimization involves both traditional as well as fractional factorial design approaches. Additionally, the production process will be compared under batch and fed-batch conditions. Finally, the optimized process conditions will be applied to 5-L stirred tank bioreactor cultivations.

Keywords: prebiotics, fructooligosaccharides, optimization, cultivation

Procedia PDF Downloads 378
17596 Efficient Numerical Simulation for LDC

Authors: Badr Alkahtani

Abstract:

In this poster, numerical solutions of two-dimensional and three-dimensional lid driven cavity are presented by solving the steady Navier-Stokes equations at high Reynolds numbers where it becomes difficult. Lid driven cavity is where the a fluid contained in a cube and the upper wall is moving. In two dimensions, we use the streamfunction-vorticity formulation to solve the problem in a square domain. A numerical method is employed to discretize the problem in the x and y directions with a spectral collocation method. The problem is coded in the MATLAB programming environment. Solutions at high Reynolds numbers are obtained up to Re=20000 on a fine grid of 131 * 131. Also in this presentation, the numerical solutions for the three-dimensional lid-driven cavity problem are obtained by solving the velocity-vorticity formulation of the Navier-Stokes equations (which is the first time that this has been simulated with special boundary conditions) for various Reynolds numbers. A spectral collocation method is employed to discretize the y and z directions and a finite difference method is used to discretize the x direction. Numerical solutions are obtained for Reynolds number up to 200. , The work prepared here is to show the efficiency of methods used to simulate the physical problem where accurate simulations of lid driven cavity are obtained at high Reynolds number as mentioned above. The result for the two dimensional problem is far from the previous researcher result.

Keywords: lid driven cavity, navier-stokes, simulation, Reynolds number

Procedia PDF Downloads 705
17595 Adsorption and Transformation of Lead in Coimbatore Urban Soils

Authors: K. Sivasubramanin, S. Mahimairaja, S. Pavithrapriya

Abstract:

Heavy metal pollution originating from industrial wastes is becoming a serious problem in many urban environments. These heavy metals, if not properly managed, could enter into the food chain and cause a serious health hazards in animals and humans. Industrial wastes, sewage sludge, and automobile emissions also contribute to heavy metal like Pb pollution in the urban environment. However, information is scarce on the heavy metal pollution in Coimbatore urban environment. Therefore, the current study was carried out to examine the extent of lead pollution in Coimbatore urban environment the maximum Pb concentration in Coimbatore urban environment was found in ukkadam, whose concentration in soils 352 mg kg-1. In many places, the Pb concentration was found exceeded the permissible limit of 100 mg kg-1. In laboratory, closed incubation experiment showed that the concentration of different species of Pb viz., water soluble Pb(H2O-Pb), exchangeable Pb(KNO3-Pb), organic-Pb(NaOH-Pb), and organic plus metal (Fe & Al) oxides bound-Pb(Na2 EDTA-Pb) was found significantly increased during the 15 days incubation, mainly due to biotransformation processes. Both the moisture content of soil and ambient temperature exerted a profound influence on the transformation of Pb. The results of a batch experiment has shown that the sorption data was adequately described by the Freundlich equation as indicated by the high correlation coefficients (R2= 0.64) than the Langmuir equation (R2 = 0.33). A maximum of 86 mg of Pb was found adsorbed per kilogram of soil. Consistently, a soil column experiment result had shown that only a small amount of Pb( < 1.0 µg g-1 soil) alone was found leached from the soil. This might be due to greater potential of the soil towards Pb adsorption.

Keywords: lead pollution, adsorption, transformation, heavy metal pollution

Procedia PDF Downloads 319
17594 Self-Calibration of Fish-Eye Camera for Advanced Driver Assistance Systems

Authors: Atef Alaaeddine Sarraj, Brendan Jackman, Frank Walsh

Abstract:

Tomorrow’s car will be more automated and increasingly connected. Innovative and intuitive interfaces are essential to accompany this functional enrichment. For that, today the automotive companies are competing to offer an advanced driver assistance system (ADAS) which will be able to provide enhanced navigation, collision avoidance, intersection support and lane keeping. These vision-based functions require an accurately calibrated camera. To achieve such differentiation in ADAS requires sophisticated sensors and efficient algorithms. This paper explores the different calibration methods applicable to vehicle-mounted fish-eye cameras with arbitrary fields of view and defines the first steps towards a self-calibration method that adequately addresses ADAS requirements. In particular, we present a self-calibration method after comparing different camera calibration algorithms in the context of ADAS requirements. Our method gathers data from unknown scenes while the car is moving, estimates the camera intrinsic and extrinsic parameters and corrects the wide-angle distortion. Our solution enables continuous and real-time detection of objects, pedestrians, road markings and other cars. In contrast, other camera calibration algorithms for ADAS need pre-calibration, while the presented method calibrates the camera without prior knowledge of the scene and in real-time.

Keywords: advanced driver assistance system (ADAS), fish-eye, real-time, self-calibration

Procedia PDF Downloads 243
17593 Coupling Heat Transfer by Natural Convection and Thermal Radiation in a Storage Tank of LNG

Authors: R. Hariti, M. Saighi, H. Saidani-Scott

Abstract:

A numerical simulation of natural convection double diffusion, coupled with thermal radiation in unsteady laminar regime in a storage tank is carried out. The storage tank contains a liquefied natural gas (LNG) in its gaseous phase. Fluent, a commercial CFD package, based on the numerical finite volume method, is used to simulate the flow. The radiative transfer equation is solved using the discrete coordinate method. This numerical simulation is used to determine the temperature profiles, stream function, velocity vectors and variation of the heat flux density for unsteady laminar natural convection. Furthermore, the influence of thermal radiation on the heat transfer has been investigated and the results obtained were compared to those found in the literature. Good agreement between the results obtained by the numerical method and those taken on site for the temperature values.

Keywords: tank, storage, liquefied natural gas, natural convection, thermal radiation, numerical simulation

Procedia PDF Downloads 533
17592 A Simple Finite Element Method for Glioma Tumor Growth Model with Density Dependent Diffusion

Authors: Shangerganesh Lingeshwaran

Abstract:

In this presentation, we have performed numerical simulations for a reaction-diffusion equation with various nonlinear density-dependent diffusion operators and proliferation functions. The mathematical model represented by parabolic partial differential equation is considered to study the invasion of gliomas (the most common type of brain tumors) and to describe the growth of cancer cells and response to their treatment. The unknown quantity of the given reaction-diffusion equation is the density of cancer cells and the mathematical model based on the proliferation and migration of glioma cells. A standard Galerkin finite element method is used to perform the numerical simulations of the given model. Finally, important observations on the each of nonlinear diffusion functions and proliferation functions are presented with the help of computational results.

Keywords: glioma invasion, nonlinear diffusion, reaction-diffusion, finite eleament method

Procedia PDF Downloads 223
17591 The Control of Wall Thickness Tolerance during Pipe Purchase Stage Based on Reliability Approach

Authors: Weichao Yu, Kai Wen, Weihe Huang, Yang Yang, Jing Gong

Abstract:

Metal-loss corrosion is a major threat to the safety and integrity of gas pipelines as it may result in the burst failures which can cause severe consequences that may include enormous economic losses as well as the personnel casualties. Therefore, it is important to ensure the corroding pipeline integrity and efficiency, considering the value of wall thickness, which plays an important role in the failure probability of corroding pipeline. Actually, the wall thickness is controlled during pipe purchase stage. For example, the API_SPEC_5L standard regulates the allowable tolerance of the wall thickness from the specified value during the pipe purchase. The allowable wall thickness tolerance will be used to determine the wall thickness distribution characteristic such as the mean value, standard deviation and distribution. Taking the uncertainties of the input variables in the burst limit-state function into account, the reliability approach rather than the deterministic approach will be used to evaluate the failure probability. Moreover, the cost of pipe purchase will be influenced by the allowable wall thickness tolerance. More strict control of the wall thickness usually corresponds to a higher pipe purchase cost. Therefore changing the wall thickness tolerance will vary both the probability of a burst failure and the cost of the pipe. This paper describes an approach to optimize the wall thickness tolerance considering both the safety and economy of corroding pipelines. In this paper, the corrosion burst limit-state function in Annex O of CSAZ662-7 is employed to evaluate the failure probability using the Monte Carlo simulation technique. By changing the allowable wall thickness tolerance, the parameters of the wall thickness distribution in the limit-state function will be changed. Using the reliability approach, the corresponding variations in the burst failure probability will be shown. On the other hand, changing the wall thickness tolerance will lead to a change in cost in pipe purchase. Using the variation of the failure probability and pipe cost caused by changing wall thickness tolerance specification, the optimal allowable tolerance can be obtained, and used to define pipe purchase specifications.

Keywords: allowable tolerance, corroding pipeline segment, operation cost, production cost, reliability approach

Procedia PDF Downloads 390
17590 Intelligent Fault Diagnosis for the Connection Elements of Modular Offshore Platforms

Authors: Jixiang Lei, Alexander Fuchs, Franz Pernkopf, Katrin Ellermann

Abstract:

Within the Space@Sea project, funded by the Horizon 2020 program, an island consisting of multiple platforms was designed. The platforms are connected by ropes and fenders. The connection is critical with respect to the safety of the whole system. Therefore, fault detection systems are investigated, which could detect early warning signs for a possible failure in the connection elements. Previously, a model-based method called Extended Kalman Filter was developed to detect the reduction of rope stiffness. This method detected several types of faults reliably, but some types of faults were much more difficult to detect. Furthermore, the model-based method is sensitive to environmental noise. When the wave height is low, a long time is needed to detect a fault and the accuracy is not always satisfactory. In this sense, it is necessary to develop a more accurate and robust technique that can detect all rope faults under a wide range of operational conditions. Inspired by this work on the Space at Sea design, we introduce a fault diagnosis method based on deep neural networks. Our method cannot only detect rope degradation by using the acceleration data from each platform but also estimate the contributions of the specific acceleration sensors using methods from explainable AI. In order to adapt to different operational conditions, the domain adaptation technique DANN is applied. The proposed model can accurately estimate rope degradation under a wide range of environmental conditions and help users understand the relationship between the output and the contributions of each acceleration sensor.

Keywords: fault diagnosis, deep learning, domain adaptation, explainable AI

Procedia PDF Downloads 171
17589 A Modified QuEChERS Method Using Activated Carbon Fibers as r-DSPE Sorbent for Sample Cleanup: Application to Pesticides Residues Analysis in Food Commodities Using GC-MS/MS

Authors: Anshuman Srivastava, Shiv Singh, Sheelendra Pratap Singh

Abstract:

A simple, sensitive and effective gas chromatography tandem mass spectrometry (GC-MS/MS) method was developed for simultaneous analysis of multi pesticide residues (organophosphate, organochlorines, synthetic pyrethroids and herbicides) in food commodities using phenolic resin based activated carbon fibers (ACFs) as reversed-dispersive solid phase extraction (r-DSPE) sorbent in modified QuEChERS (Quick Easy Cheap Effective Rugged Safe) method. The acetonitrile-based QuEChERS technique was used for the extraction of the analytes from food matrices followed by sample cleanup with ACFs instead of traditionally used primary secondary amine (PSA). Different physico-chemical characterization techniques such as Fourier transform infrared spectroscopy, scanning electron microscopy, X-ray diffraction and Brunauer-Emmet-Teller surface area analysis were employed to investigate the engineering and structural properties of ACFs. The recovery of pesticides and herbicides was tested at concentration levels of 0.02 and 0.2 mg/kg in different commodities such as cauliflower, cucumber, banana, apple, wheat and black gram. The recoveries of all twenty-six pesticides and herbicides were found in acceptable limit (70-120%) according to SANCO guideline with relative standard deviation value < 15%. The limit of detection and limit of quantification of the method was in the range of 0.38-3.69 ng/mL and 1.26 -12.19 ng/mL, respectively. In traditional QuEChERS method, PSA used as r-DSPE sorbent plays a vital role in sample clean-up process and demonstrates good recoveries for multiclass pesticides. This study reports that ACFs are better in terms of removal of co-extractives in comparison of PSA without compromising the recoveries of multi pesticides from food matrices. Further, ACF replaces the need of charcoal in addition to the PSA from traditional QuEChERS method which is used to remove pigments. The developed method will be cost effective because the ACFs are significantly cheaper than the PSA. So the proposed modified QuEChERS method is more robust, effective and has better sample cleanup efficiency for multiclass multi pesticide residues analysis in different food matrices such as vegetables, grains and fruits.

Keywords: QuEChERS, activated carbon fibers, primary secondary amine, pesticides, sample preparation, carbon nanomaterials

Procedia PDF Downloads 262
17588 Numerical Simulations on Feasibility of Stochastic Model Predictive Control for Linear Discrete-Time Systems with Random Dither Quantization

Authors: Taiki Baba, Tomoaki Hashimoto

Abstract:

The random dither quantization method enables us to achieve much better performance than the simple uniform quantization method for the design of quantized control systems. Motivated by this fact, the stochastic model predictive control method in which a performance index is minimized subject to probabilistic constraints imposed on the state variables of systems has been proposed for linear feedback control systems with random dither quantization. In other words, a method for solving optimal control problems subject to probabilistic state constraints for linear discrete-time control systems with random dither quantization has been already established. To our best knowledge, however, the feasibility of such a kind of optimal control problems has not yet been studied. Our objective in this paper is to investigate the feasibility of stochastic model predictive control problems for linear discrete-time control systems with random dither quantization. To this end, we provide the results of numerical simulations that verify the feasibility of stochastic model predictive control problems for linear discrete-time control systems with random dither quantization.

Keywords: model predictive control, stochastic systems, probabilistic constraints, random dither quantization

Procedia PDF Downloads 274
17587 Dual Reconfigurable Antenna Using Capacitive Coupling Slot and Parasitic Square Ring

Authors: M. Abou Al-alaa, H. A. Elsadek, E. A. Abdallah, E. A. Hashish

Abstract:

A square patch antenna with both frequency and polarization reconfigurability is presented. The antenna consists of a square patch with coplanar feed on the ground plane. On the patch side, there is a parasitic square ring that is responsible for changing the antenna polarization. On the ground plane, there is a rectangular slot. By changing of length of this slot, the antenna resonance frequency can be changed. The antenna operates at 1.57 and 2.45 GHz that used in GPS and Bluetooth applications, respectively. The length of the slot in the proposed antenna is 40 mm, and the antenna operates at the lower frequency (1.57 GHz). By using switches in the ground plane the slot length can be adjust to 24 mm, so the antenna operates at upper frequency (2.45 GHz). Two switches are mounted on the parasitic ring at optimized positions. By switching between the different states of these two switches, the proposed antenna operates with linear polarization (LP) and circular polarization (CP) at each operating frequency. The antenna gain at 1.57 and 2.45 GHz are 5.9 and 7.64 dBi, respectively. The antenna is analyzed using the CST Microwave Studio. The proposed antenna was fabricated and measured. Results comparison shows good agreement. The antenna has applications in several wireless communication systems.

Keywords: microstrip patch antenna, reconfigurable antenna, frequency reconfigurability, polarization reconfigurability, parasitic square ring, linear polarization, circular polarization

Procedia PDF Downloads 519
17586 Computer-Integrated Surgery of the Human Brain, New Possibilities

Authors: Ugo Galvanetto, Pirto G. Pavan, Mirco Zaccariotto

Abstract:

The discipline of Computer-integrated surgery (CIS) will provide equipment able to improve the efficiency of healthcare systems and, which is more important, clinical results. Surgeons and machines will cooperate in new ways that will extend surgeons’ ability to train, plan and carry out surgery. Patient specific CIS of the brain requires several steps: 1 - Fast generation of brain models. Based on image recognition of MR images and equipped with artificial intelligence, image recognition techniques should differentiate among all brain tissues and segment them. After that, automatic mesh generation should create the mathematical model of the brain in which the various tissues (white matter, grey matter, cerebrospinal fluid …) are clearly located in the correct positions. 2 – Reliable and fast simulation of the surgical process. Computational mechanics will be the crucial aspect of the entire procedure. New algorithms will be used to simulate the mechanical behaviour of cutting through cerebral tissues. 3 – Real time provision of visual and haptic feedback A sophisticated human-machine interface based on ergonomics and psychology will provide the feedback to the surgeon. The present work will address in particular point 2. Modelling the cutting of soft tissue in a structure as complex as the human brain is an extremely challenging problem in computational mechanics. The finite element method (FEM), that accurately represents complex geometries and accounts for material and geometrical nonlinearities, is the most used computational tool to simulate the mechanical response of soft tissues. However, the main drawback of FEM lies in the mechanics theory on which it is based, classical continuum Mechanics, which assumes matter is a continuum with no discontinuity. FEM must resort to complex tools such as pre-defined cohesive zones, external phase-field variables, and demanding remeshing techniques to include discontinuities. However, all approaches to equip FEM computational methods with the capability to describe material separation, such as interface elements with cohesive zone models, X-FEM, element erosion, phase-field, have some drawbacks that make them unsuitable for surgery simulation. Interface elements require a-priori knowledge of crack paths. The use of XFEM in 3D is cumbersome. Element erosion does not conserve mass. The Phase Field approach adopts a diffusive crack model instead of describing true tissue separation typical of surgical procedures. Modelling discontinuities, so difficult when using computational approaches based on classical continuum Mechanics, is instead easy for novel computational methods based on Peridynamics (PD). PD is a non-local theory of mechanics formulated with no use of spatial derivatives. Its governing equations are valid at points or surfaces of discontinuity, and it is, therefore especially suited to describe crack propagation and fragmentation problems. Moreover, PD does not require any criterium to decide the direction of crack propagation or the conditions for crack branching or coalescence; in the PD-based computational methods, cracks develop spontaneously in the way which is the most convenient from an energy point of view. Therefore, in PD computational methods, crack propagation in 3D is as easy as it is in 2D, with a remarkable advantage with respect to all other computational techniques.

Keywords: computational mechanics, peridynamics, finite element, biomechanics

Procedia PDF Downloads 68