Search results for: rejection simulation method
22167 Flowsheet Development, Simulation and Optimization of Carbon-Di-Oxide Removal System at Natural Gas Reserves by Aspen–Hysys Process Simulator
Authors: Mohammad Ruhul Amin, Nusrat Jahan
Abstract:
Natural gas is a cleaner fuel compared to the others. But it needs some treatment before it is in a state to be used. So natural gas purification is an integral part of any process where natural gas is used as raw material or fuel. There are several impurities in natural gas that have to be removed before use. CO2 is one of the major contaminants. In this project we have removed CO2 by amine process by using MEA solution. We have built up the whole amine process for removing CO2 in Aspen Hysys and simulated the process. At the end of simulation we have got very satisfactory results by using MEA solution for the removal of CO2. Simulation result shows that amine absorption process enables to reduce CO2 content from NG by 58%. HYSYS optimizer allowed us to get a perfect optimized plant. After optimization the profit of existing plant is increased by 2.34 %.Simulation and optimization by Aspen-HYSYS simulator makes available us to enormous information which will help us to further research in future.Keywords: Aspen–Hysys, CO2 removal, flowsheet development, MEA solution, natural gas optimization
Procedia PDF Downloads 50222166 Clinical Training Simulation Experience of Medical Sector Students
Authors: Tahsien Mohamed Okasha
Abstract:
Simulation is one of the emerging educational strategies that depend on the creation of scenarios to imitate what could happen in real life. At the time of COVID, we faced big obstacles in medical education, specially the clinical part and how we could apply it, the simulation was the golden key. Simulation is a very important tool of education for medical sector students, through creating a safe, changeable, quiet environment with less anxiety level for students to practice and to have repeated trials on their competencies. That impacts the level of practice, achievement, and the way of acting in real situations and experiences. A blind Random sample of students from different specialties and colleges who came and finished their training in an integrated environment was collected and tested, and the responses were graded from (1-5). The results revealed that 77% of the studied subjects agreed that dealing and interacting with different medical sector candidates in the same place was beneficial. 77% of the studied subjects agreed that simulations were challenging in thinking and decision-making skills .75% agreed that using high-fidelity manikins was helpful. 75% agree .76% agreed that working in a safe, prepared environment is helpful for realistic situations.Keywords: simulation, clinical training, education, medical sector students
Procedia PDF Downloads 3322165 Comparative Analysis of Two Modeling Approaches for Optimizing Plate Heat Exchangers
Authors: Fábio A. S. Mota, Mauro A. S. S. Ravagnani, E. P. Carvalho
Abstract:
In the present paper the design of plate heat exchangers is formulated as an optimization problem considering two mathematical modeling. The number of plates is the objective function to be minimized, considering implicitly some parameters configuration. Screening is the optimization method used to solve the problem. Thermal and hydraulic constraints are verified, not viable solutions are discarded and the method searches for the convergence to the optimum, case it exists. A case study is presented to test the applicability of the developed algorithm. Results show coherency with the literature.Keywords: plate heat exchanger, optimization, modeling, simulation
Procedia PDF Downloads 52022164 Inviscid Steady Flow Simulation Around a Wing Configuration Using MB_CNS
Authors: Muhammad Umar Kiani, Muhammad Shahbaz, Hassan Akbar
Abstract:
Simulation of a high speed inviscid steady ideal air flow around a 2D/axial-symmetry body was carried out by the use of mb_cns code. mb_cns is a program for the time-integration of the Navier-Stokes equations for two-dimensional compressible flows on a multiple-block structured mesh. The flow geometry may be either planar or axisymmetric and multiply-connected domains can be modeled by patching together several blocks. The main simulation code is accompanied by a set of pre and post-processing programs. The pre-processing programs scriptit and mb_prep start with a short script describing the geometry, initial flow state and boundary conditions and produce a discretized version of the initial flow state. The main flow simulation program (or solver as it is sometimes called) is mb_cns. It takes the files prepared by scriptit and mb_prep, integrates the discrete form of the gas flow equations in time and writes the evolved flow data to a set of output files. This output data may consist of the flow state (over the whole domain) at a number of instants in time. After integration in time, the post-processing programs mb_post and mb_cont can be used to reformat the flow state data and produce GIF or postscript plots of flow quantities such as pressure, temperature and Mach number. The current problem is an example of supersonic inviscid flow. The flow domain for the current problem (strake configuration wing) is discretized by a structured grid and a finite-volume approach is used to discretize the conservation equations. The flow field is recorded as cell-average values at cell centers and explicit time stepping is used to update conserved quantities. MUSCL-type interpolation and one of three flux calculation methods (Riemann solver, AUSMDV flux splitting and the Equilibrium Flux Method, EFM) are used to calculate inviscid fluxes across cell faces.Keywords: steady flow simulation, processing programs, simulation code, inviscid flux
Procedia PDF Downloads 43022163 Detection Method of Federated Learning Backdoor Based on Weighted K-Medoids
Authors: Xun Li, Haojie Wang
Abstract:
Federated learning is a kind of distributed training and centralized training mode, which is of great value in the protection of user privacy. In order to solve the problem that the model is vulnerable to backdoor attacks in federated learning, a backdoor attack detection method based on a weighted k-medoids algorithm is proposed. First of all, this paper collates the update parameters of the client to construct a vector group, then uses the principal components analysis (PCA) algorithm to extract the corresponding feature information from the vector group, and finally uses the improved k-medoids clustering algorithm to identify the normal and backdoor update parameters. In this paper, the backdoor is implanted in the federation learning model through the model replacement attack method in the simulation experiment, and the update parameters from the attacker are effectively detected and removed by the defense method proposed in this paper.Keywords: federated learning, backdoor attack, PCA, k-medoids, backdoor defense
Procedia PDF Downloads 11722162 Effectiveness of Simulation Resuscitation Training to Improve Self-Efficacy of Physicians and Nurses at Aga Khan University Hospital in Advanced Cardiac Life Support Courses Quasi-Experimental Study Design
Authors: Salima R. Rajwani, Tazeen Ali, Rubina Barolia, Yasmin Parpio, Nasreen Alwani, Salima B. Virani
Abstract:
Introduction: Nurses and physicians have a critical role in initiating lifesaving interventions during cardiac arrest. It is important that timely delivery of high quality Cardio Pulmonary Resuscitation (CPR) with advanced resuscitation skills and management of cardiac arrhythmias is a key dimension of code during cardiac arrest. It will decrease the chances of patient survival if the healthcare professionals are unable to initiate CPR timely. Moreover, traditional training will not prepare physicians and nurses at a competent level and their knowledge level declines over a period of time. In this regard, simulation training has been proven to be effective in promoting resuscitation skills. Simulation teaching learning strategy improves knowledge level, and skills performance during resuscitation through experiential learning without compromising patient safety in real clinical situations. The purpose of the study is to evaluate the effectiveness of simulation training in Advanced Cardiac Life Support Courses by using the selfefficacy tool. Methods: The study design is a quantitative research design and non-randomized quasi-experimental study design. The study examined the effectiveness of simulation through self-efficacy in two instructional methods; one is Medium Fidelity Simulation (MFS) and second is Traditional Training Method (TTM). The sample size was 220. Data was compiled by using the SPSS tool. The standardized simulation based training increases self-efficacy, knowledge, and skills and improves the management of patients in actual resuscitation. Results: 153 students participated in study; CG: n = 77 and EG: n = 77. The comparison was done between arms in pre and post-test. (F value was 1.69, p value is <0.195 and df was 1). There was no significant difference between arms in the pre and post-test. The interaction between arms was observed and there was no significant difference in interaction between arms in the pre and post-test. (F value was 0.298, p value is <0.586 and df is 1. However, the results showed self-efficacy scores were significantly higher within experimental group in post-test in advanced cardiac life support resuscitation courses as compared to Traditional Training Method (TTM) and had overall (p <0.0001) and F value was 143.316 (mean score was 45.01 and SD was 9.29) verses pre-test result showed (mean score was 31.15 and SD was 12.76) as compared to TTM in post-test (mean score was 29.68 and SD was 14.12) verses pre-test result showed (mean score was 42.33 and SD was 11.39). Conclusion: The standardized simulation-based training was conducted in the safe learning environment in Advanced Cardiac Life Suport Courses and physicians and nurses benefited from self-confidence, early identification of life-threatening scenarios, early initiation of CPR, and provides high-quality CPR, timely administration of medication and defibrillation, appropriate airway management, rhythm analysis and interpretation, and Return of Spontaneous Circulation (ROSC), team dynamics, debriefing, and teaching and learning strategies that will improve the patient survival in actual resuscitation.Keywords: advanced cardiac life support, cardio pulmonary resuscitation, return of spontaneous circulation, simulation
Procedia PDF Downloads 8522161 Study of Water Cluster-Amorphous Silica Collisions in the Extreme Space Environment Using the ReaxFF Reactive Force Field Molecular Dynamics Simulation Method
Authors: Ali Rahnamoun, Adri van Duin
Abstract:
The concept of high velocity particle impact on the spacecraft surface materials has been one of the important issues in the design of such materials. Among these particles, water clusters might be the most abundant and the most important particles to be studied. The importance of water clusters is that upon impact on the surface of the materials, they can cause damage to the material and also if they are sub-cooled water clusters, they can attach to the surface of the materials and cause ice accumulation on the surface which is very problematic in spacecraft and also aircraft operations. The dynamics of the collisions between amorphous silica structures and water clusters with impact velocities of 1 km/s to 10 km/s are studied using the ReaxFF reactive molecular dynamics simulation method. The initial water clusters include 150 water molecules and the water clusters are collided on the surface of amorphous fully oxidized and suboxide silica structures. These simulations show that the most abundant molecules observed on the silica surfaces, other than reflecting water molecules, are H3O+ and OH- for the water cluster impacts on suboxide and fully oxidized silica structures, respectively. The effect of impact velocity on the change of silica mass is studied. At high impact velocities the water molecules attach to the silica surface through a chemisorption process meaning that water molecule dissociates through the interaction with silica surface. However, at low impact velocities, physisorbed water molecules are also observed, which means water molecule attaches and accumulates on the silica surface. The amount of physisorbed waters molecules at low velocities is higher on the suboxide silica surfaces. The evolution of the temperatures of the water clusters during the collisions indicates that the possibility of electron excitement at impact velocities less than 10 km/s is minimal and ReaxFF reactive molecular dynamics simulation can predict the chemistry of these hypervelocity impacts. However, at impact velocities close to 10 km/s the average temperature of the impacting water clusters increase to about 2000K, with individual molecules oocasionally reaching temperatures of over 8000K and thus will be prudent to consider the concept of electron excitation at these higher impact velocities which goes beyond the current ReaxFF ability.Keywords: spacecraft materials, hypervelocity impact, reactive molecular dynamics simulation, amorphous silica
Procedia PDF Downloads 42122160 A Two-Stage Bayesian Variable Selection Method with the Extension of Lasso for Geo-Referenced Data
Authors: Georgiana Onicescu, Yuqian Shen
Abstract:
Due to the complex nature of geo-referenced data, multicollinearity of the risk factors in public health spatial studies is a commonly encountered issue, which leads to low parameter estimation accuracy because it inflates the variance in the regression analysis. To address this issue, we proposed a two-stage variable selection method by extending the least absolute shrinkage and selection operator (Lasso) to the Bayesian spatial setting, investigating the impact of risk factors to health outcomes. Specifically, in stage I, we performed the variable selection using Bayesian Lasso and several other variable selection approaches. Then, in stage II, we performed the model selection with only the selected variables from stage I and compared again the methods. To evaluate the performance of the two-stage variable selection methods, we conducted a simulation study with different distributions for the risk factors, using geo-referenced count data as the outcome and Michigan as the research region. We considered the cases when all candidate risk factors are independently normally distributed, or follow a multivariate normal distribution with different correlation levels. Two other Bayesian variable selection methods, Binary indicator, and the combination of Binary indicator and Lasso were considered and compared as alternative methods. The simulation results indicated that the proposed two-stage Bayesian Lasso variable selection method has the best performance for both independent and dependent cases considered. When compared with the one-stage approach, and the other two alternative methods, the two-stage Bayesian Lasso approach provides the highest estimation accuracy in all scenarios considered.Keywords: Lasso, Bayesian analysis, spatial analysis, variable selection
Procedia PDF Downloads 14922159 The Analysis of TRACE/FRAPTRAN in the Fuel Rods of Maanshan PWR for LBLOCA
Authors: J. R. Wang, W. Y. Li, H. T. Lin, J. H. Yang, C. Shih, S. W. Chen
Abstract:
Fuel rod analysis program transient (FRAPTRAN) code was used to study the fuel rod performance during a postulated large break loss of coolant accident (LBLOCA) in Maanshan nuclear power plant (NPP). Previous transient results from thermal hydraulic code, TRACE, with the same LBLOCA scenario, were used as input boundary conditions for FRAPTRAN. The simulation results showed that the peak cladding temperatures and the fuel center line temperatures were all below the 10CFR50.46 LOCA criteria. In addition, the maximum hoop stress was 18 MPa and the oxide thickness was 0.003 mm for the present simulation cases, which are all within the safety operation ranges. The present study confirms that this analysis method, the FRAPTRAN code combined with TRACE, is an appropriate approach to predict the fuel integrity under LBLOCA with operational ECCS.Keywords: FRAPTRAN, TRACE, LOCA, PWR
Procedia PDF Downloads 51622158 Electro-Hydrodynamic Analysis of Low-Pressure DC Glow Discharge by Lattice Boltzmann Method
Authors: Ji-Hyok Kim, Il-Gyong Paek, Yong-Jun Kim
Abstract:
We propose a numerical model based on drift-diffusion theory and lattice Boltzmann method (LBM) to analyze the electro-hydrodynamic behavior in low-pressure direct current (DC) glow discharge plasmas. We apply the drift-diffusion theory for 4-species and employ the standard lattice Boltzmann model (SLBM) for the electron, the finite difference-lattice Boltzmann model (FD-LBM) for heavy particles, and the finite difference model (FDM) for the electric potential, respectively. Our results are compared with those of other methods, and emphasize the necessity of a two-dimensional analysis for glow discharge.Keywords: glow discharge, lattice Boltzmann method, numerical analysis, plasma simulation, electro-hydrodynamic
Procedia PDF Downloads 12522157 Ocular Immunology: In Face of Immune Privilege the Eye Remains Vulnerable to Autoimmune and Inflammatory-Mediated Diseases
Authors: Husham Bayazed
Abstract:
Purpose of Presentation: The eye is one of a few sites in the body with immune privilege (IP). However, this IP is relatively easily bypassed in the face of sufficient strong local or systemic immunological responses. As immune responses are crucial elements of the repair response, the eye has developed distinct mechanisms to deliver immune responses to injury in the avascular regions of the eye. This presentation may cover and provide an overview of the mechanisms that dictate immune cell trafficking to the local ocular microenvironment in response to different autoimmune and inflammatory-mediated diseases. Recent Findings: Literature reviews declare that immune responses and inflammation play a key role in a diverse range of eye diseases. In recent years, our understanding of ocular immune responses has widely spread in ocular surface inflammation, uveitis, age-related macular degeneration (AMD), glaucoma, transplantation rejection, and other ocular diseases. It is becoming increasingly clear that multiple seemingly unrelated diseases involve immune responses with common themes in their ocular pathogenesis. Recent studies are focusing on elucidating the pathogenesis of ocular inflammatory disease to identify new targets for immunotherapy that will not only improve efficacy but also minimize adverse effects from traditional therapy. Summary: While IP was believed to protect the eye from day-to-day inflammatory insults, however, it is relatively easily breached in the face of different strong local or systemic immunological and inflammatory responses. Therefore, the ocular immune response encapsulates the full range of classical and non-classical immune responses and demonstrates many features which are reflected in other tissues, but eye tissues, by modifying these responses, may reveal unexpected and novel findings which are relevant to immune responses generally. This may have therapeutic potential for new targeting immunotherapy, restoring immune tolerance in ocular autoimmune and inflammatory diseases, and preventing rejection such as stem cells, currently being considered for treatment of worldwide blinding diseases such as AMD.Keywords: ocular diseases, immunology, immune privilege, immunotherapy
Procedia PDF Downloads 8222156 Analyzing Large Scale Recurrent Event Data with a Divide-And-Conquer Approach
Authors: Jerry Q. Cheng
Abstract:
Currently, in analyzing large-scale recurrent event data, there are many challenges such as memory limitations, unscalable computing time, etc. In this research, a divide-and-conquer method is proposed using parametric frailty models. Specifically, the data is randomly divided into many subsets, and the maximum likelihood estimator from each individual data set is obtained. Then a weighted method is proposed to combine these individual estimators as the final estimator. It is shown that this divide-and-conquer estimator is asymptotically equivalent to the estimator based on the full data. Simulation studies are conducted to demonstrate the performance of this proposed method. This approach is applied to a large real dataset of repeated heart failure hospitalizations.Keywords: big data analytics, divide-and-conquer, recurrent event data, statistical computing
Procedia PDF Downloads 17222155 Part Variation Simulations: An Industrial Case Study with an Experimental Validation
Authors: Narendra Akhadkar, Silvestre Cano, Christophe Gourru
Abstract:
Injection-molded parts are widely used in power system protection products. One of the biggest challenges in an injection molding process is shrinkage and warpage of the molded parts. All these geometrical variations may have an adverse effect on the quality of the product, functionality, cost, and time-to-market. The situation becomes more challenging in the case of intricate shapes and in mass production using multi-cavity tools. To control the effects of shrinkage and warpage, it is very important to correctly find out the input parameters that could affect the product performance. With the advances in the computer-aided engineering (CAE), different tools are available to simulate the injection molding process. For our case study, we used the MoldFlow insight tool. Our aim is to predict the spread of the functional dimensions and geometrical variations on the part due to variations in the input parameters such as material viscosity, packing pressure, mold temperature, melt temperature, and injection speed. The input parameters may vary during batch production or due to variations in the machine process settings. To perform the accurate product assembly variation simulation, the first step is to perform an individual part variation simulation to render realistic tolerance ranges. In this article, we present a method to simulate part variations coming from the input parameters variation during batch production. The method is based on computer simulations and experimental validation using the full factorial design of experiments (DoE). The robustness of the simulation model is verified through input parameter wise sensitivity analysis study performed using simulations and experiments; all the results show a very good correlation in the material flow direction. There exists a non-linear interaction between material and the input process variables. It is observed that the parameters such as packing pressure, material, and mold temperature play an important role in spread on functional dimensions and geometrical variations. This method will allow us in the future to develop accurate/realistic virtual prototypes based on trusted simulated process variation and, therefore, increase the product quality and potentially decrease the time to market.Keywords: correlation, molding process, tolerance, sensitivity analysis, variation simulation
Procedia PDF Downloads 18422154 Vine Copula Structure among Yield, Price and Weather Variables for Rating Crop Insurance Premium
Authors: Jiemiao Chen, Shuoxun Xu
Abstract:
The main goal of our research is to apply the Vine copula measuring dependency between price, temperature, and precipitation indices to calculate a fair crop insurance premium. This research is focused on Worth, Iowa, United States, over the period from 2000 to 2020, where the farmers are dependent on precipitation and average temperature during the growth period of corn. Our proposed insurance considers both the natural risk and the price risk in agricultural production. We first estimate the distributions of crops using parametric methods based on Goodness of Fit tests, and then Vine Copula is applied to model dependence between yield price, crop yield, and weather indices. Once the vine structure and its parameters are determined based on AIC/BIC criteria and forecasting price and yield are obtained from the ARIMA model, we calculate this crop insurance premium using the simulation data generated from the vine copula by the Monte Carlo Simulation method. It is shown that, compared with traditional crop insurance, our proposed insurance is more fair and thus less costly for the farmers and government.Keywords: vine copula, weather index, crop insurance premium, insurance risk management, Monte Carlo simulation
Procedia PDF Downloads 20322153 Analysis of the Simulation Merger and Economic Benefit of Local Farmers' Associations in Taiwan
Authors: Lu Yung-Hsiang, Chang Kuming, Dai Yi-Fang, Liao Ching-Yi
Abstract:
According to Taiwan’s administrative division of future land planning may lead farmer association and service areas facing recombination or merger. Thus, merger combination and the economic benefit of the farmer association are worth to be discussed. The farmer association in the merger, which may cause some then will not be consolidated, or consolidate two, or ever more to one association. However, under what condition to merge is greatest, as one of observation of this study. In addition, research without using simulation methods and only on the credit department rather whole farmer association. Therefore, this paper will use the simulation approach, and examine both the merge of farmer association and the condition under which the benefits are the greatest. The data of this study set include 266 farmer associations in Taiwan period 2012 to 2013. Empirical results showed that the number of the farmer association optimal simulation combination is 108.After the merger from the first stage can be reduced by 60% of the farmers’ association. The cost saving effects of the post-merger is not different. The cost efficiency of the farmers’ association improved it. The economies of scale and scope would decrease by the merger. The research paper hopes the finding will benefit the future merger of the farmers’ association.Keywords: simulation merger, farmer association, assurance region, data envelopment analysis
Procedia PDF Downloads 35322152 Assessment of Frying Material by Deep-Fat Frying Method
Authors: Brinda Sharma, Saakshi S. Sarpotdar
Abstract:
Deep-fat frying is popular standard method that has been studied basically to clarify the complicated mechanisms of fat decomposition at high temperatures and to assess their effects on human health. The aim of this paper is to point out the application of method engineering that has been recently improved our understanding of the fundamental principles and mechanisms concerned at different scales and different times throughout the process: pretreatment, frying, and cooling. It covers the several aspects of deep-fat drying. New results regarding the understanding of the frying method that are obtained as a results of major breakthroughs in on-line instrumentation (heat, steam flux, and native pressure sensors), within the methodology of microstructural and imaging analysis (NMR, MRI, SEM) and in software system tools for the simulation of coupled transfer and transport phenomena. Such advances have opened the approach for the creation of significant information of the behavior of varied materials and to the event of latest tools to manage frying operations via final product quality in real conditions. Lastly, this paper promotes an integrated approach to the frying method as well as numerous competencies like those of chemists, engineers, toxicologists, nutritionists, and materials scientists also as of the occupation and industrial sectors.Keywords: frying, cooling, imaging analysis (NMR, MRI, SEM), deep-fat frying
Procedia PDF Downloads 43122151 Optimal Design of Storm Water Networks Using Simulation-Optimization Technique
Authors: Dibakar Chakrabarty, Mebada Suiting
Abstract:
Rapid urbanization coupled with changes in land use pattern results in increasing peak discharge and shortening of catchment time of concentration. The consequence is floods, which often inundate roads and inhabited areas of cities and towns. Management of storm water resulting from rainfall has, therefore, become an important issue for the municipal bodies. Proper management of storm water obviously includes adequate design of storm water drainage networks. The design of storm water network is a costly exercise. Least cost design of storm water networks assumes significance, particularly when the fund available is limited. Optimal design of a storm water system is a difficult task as it involves the design of various components, like, open or closed conduits, storage units, pumps etc. In this paper, a methodology for least cost design of storm water drainage systems is proposed. The methodology proposed in this study consists of coupling a storm water simulator with an optimization method. The simulator used in this study is EPA’s storm water management model (SWMM), which is linked with Genetic Algorithm (GA) optimization method. The model proposed here is a mixed integer nonlinear optimization formulation, which takes care of minimizing the sectional areas of the open conduits of storm water networks, while satisfactorily conveying the runoff resulting from rainfall to the network outlet. Performance evaluations of the developed model show that the proposed method can be used for cost effective design of open conduit based storm water networks.Keywords: genetic algorithm (GA), optimal design, simulation-optimization, storm water network, SWMM
Procedia PDF Downloads 25322150 Molecular Dynamics Simulation for Buckling Analysis at Nanocomposite Beams
Authors: Babak Safaei, A. M. Fattahi
Abstract:
In the present study we have investigated axial buckling characteristics of nanocomposite beams reinforced by single-walled carbon nanotubes (SWCNTs). Various types of beam theories including Euler-Bernoulli beam theory, Timoshenko beam theory and Reddy beam theory were used to analyze the buckling behavior of carbon nanotube-reinforced composite beams. Generalized differential quadrature (GDQ) method was utilized to discretize the governing differential equations along with four commonly used boundary conditions. The material properties of the nanocomposite beams were obtained using molecular dynamic (MD) simulation corresponding to both short-(10,10) SWCNT and long-(10,10) SWCNT composites which were embedded by amorphous polyethylene matrix. Then the results obtained directly from MD simulations were matched with those calculated by the mixture rule to extract appropriate values of carbon nanotube efficiency parameters accounting for the scale-dependent material properties. The selected numerical results were presented to indicate the influences of nanotube volume fractions and end supports on the critical axial buckling loads of nanocomposite beams relevant to long- and short-nanotube composites.Keywords: nanocomposites, molecular dynamics simulation, axial buckling, generalized differential quadrature (GDQ)
Procedia PDF Downloads 32622149 Coefficient of Performance (COP) Optimization of an R134a Cross Vane Expander Compressor Refrigeration System
Authors: Y. D. Lim, K. S. Yap, K. T. Ooi
Abstract:
Cross Vane Expander Compressor (CVEC) is a newly invented expander-compressor combined unit, where it is introduced to replace the compressor and the expansion valve in traditional refrigeration system. The mathematical model of CVEC has been developed to examine its performance, and it was found that the energy consumption of a conventional refrigeration system was reduced by as much as 18%. It is believed that energy consumption can be further reduced by optimizing the device. In this study, the coefficient of performance (COP) of CVEC has been optimized under predetermined operational parameters and constrained main design parameters. Several main design parameters of CVEC were selected to be the variables, and the optimization was done with theoretical model in a simulation program. The theoretical model consists of geometrical model, dynamic model, heat transfer model and valve dynamics model. Complex optimization method, which is a constrained, direct search and multi-variables method was used in the study. As a result, the optimization study suggested that with an appropriate combination of design parameters, a 58% COP improvement in CVEC R134a refrigeration system is possible.Keywords: COP, cross vane expander-compressor, CVEC, design, simulation, refrigeration system, air-conditioning, R134a, multi variables
Procedia PDF Downloads 33822148 Single Pole-To-Earth Fault Detection and Location on the Tehran Railway System Using ICA and PSO Trained Neural Network
Authors: Masoud Safarishaal
Abstract:
Detecting the location of pole-to-earth faults is essential for the safe operation of the electrical system of the railroad. This paper aims to use a combination of evolutionary algorithms and neural networks to increase the accuracy of single pole-to-earth fault detection and location on the Tehran railroad power supply system. As a result, the Imperialist Competitive Algorithm (ICA) and Particle Swarm Optimization (PSO) are used to train the neural network to improve the accuracy and convergence of the learning process. Due to the system's nonlinearity, fault detection is an ideal application for the proposed method, where the 600 Hz harmonic ripple method is used in this paper for fault detection. The substations were simulated by considering various situations in feeding the circuit, the transformer, and typical Tehran metro parameters that have developed the silicon rectifier. Required data for the network learning process has been gathered from simulation results. The 600Hz component value will change with the change of the location of a single pole to the earth's fault. Therefore, 600Hz components are used as inputs of the neural network when fault location is the output of the network system. The simulation results show that the proposed methods can accurately predict the fault location.Keywords: single pole-to-pole fault, Tehran railway, ICA, PSO, artificial neural network
Procedia PDF Downloads 13022147 Socratic Style of Teaching: An Analysis of Dialectical Method
Authors: Muhammad Jawwad, Riffat Iqbal
Abstract:
The Socratic method, also known as the dialectical method and elenctic method, has significant relevance in the contemporary educational system. It can be incorporated into modern-day educational systems theoretically as well as practically. Being interactive and dialogue-based in nature, this teaching approach is followed by critical thinking and innovation. The pragmatic value of the Dialectical Method has been discussed in this article, and the limitations of the Socratic method have also been highlighted. The interactive Method of Socrates can be used in many subjects for students of different grades. The Limitations and delimitations of the Method have also been discussed for its proper implementation. This article has attempted to elaborate and analyze the teaching method of Socrates with all its pre-suppositions and Epistemological character.Keywords: Socratic method, dialectical method, knowledge, teaching, virtue
Procedia PDF Downloads 13722146 Noise and Thermal Analyses of Memristor-Based Phase Locked Loop Integrated Circuit
Authors: Naheem Olakunle Adesina
Abstract:
The memristor is considered as one of the promising candidates for mamoelectronic engineering and applications. Owing to its high compatibility with CMOS, nanoscale size, and low power consumption, memristor has been employed in the design of commonly used circuits such as phase-locked loop (PLL). In this paper, we designed a memristor-based loop filter (LF) together with other components of PLL. Following this, we evaluated the noise-rejection feature of loop filter by comparing the noise levels of input and output signals of the filter. Our SPICE simulation results showed that memristor behaves like a linear resistor at high frequencies. The result also showed that loop filter blocks the high-frequency components from phase frequency detector so as to provide a stable control voltage to the voltage controlled oscillator (VCO). In addition, we examined the effects of temperature on the performance of the designed phase locked loop circuit. A critical temperature, where there is frequency drift of VCO as a result of variations in control voltage, is identified. In conclusion, the memristor is a suitable choice for nanoelectronic systems owing to a small area, low power consumption, dense nature, high switching speed, and endurance. The proposed memristor-based loop filter, together with other components of the phase locked loop, can be designed using memristive emulator and EDA tools in current CMOS technology and simulated.Keywords: Fast Fourier Transform, hysteresis curve, loop filter, memristor, noise, phase locked loop, voltage controlled oscillator
Procedia PDF Downloads 19222145 Modelling and Simulation of Cascaded H-Bridge Multilevel Single Source Inverter Using PSIM
Authors: Gaddafi Sani Shehu, Tankut Yalcınoz, Abdullahi Bala Kunya
Abstract:
Multilevel inverters such as flying capacitor, diode-clamped, and cascaded H-bridge inverters are very popular particularly in medium and high power applications. This paper focuses on a cascaded H-bridge module using a single direct current (DC) source in order to generate an 11-level output voltage. The noble approach reduces the number of switches and gate drivers, in comparison with a conventional method. The anticipated topology produces more accurate result with an isolation transformer at high switching frequency. Different modulation techniques can be used for the multilevel inverter, but this work features modulation techniques known as selective harmonic elimination (SHE).This modulation approach reduces the number of carriers with reduction in Switching Losses, Total Harmonic Distortion (THD), and thereby increasing Power Quality (PQ). Based on the simulation result obtained, it appears SHE has the ability to eliminate selected harmonics by chopping off the fundamental output component. The performance evaluation of the proposed cascaded multilevel inverter is performed using PSIM simulation package and THD of 0.94% is obtained.Keywords: cascaded H-bridge multilevel inverter, power quality, selective harmonic elimination
Procedia PDF Downloads 42322144 Fault Tolerant Control of the Dynamical Systems Based on Internal Structure Systems
Authors: Seyed Mohammad Hashemi, Shahrokh Barati
Abstract:
The problem of fault-tolerant control (FTC) by accommodation method has been studied in this paper. The fault occurs in any system components such as actuators, sensors or internal structure of the system and leads to loss of performance and instability of the system. When a fault occurs, the purpose of the fault-tolerant control is designate strategy that can keep the control loop stable and system performance as much as possible perform it without shutting down the system. Here, the section of fault detection and isolation (FDI) system has been evaluated with regard to actuator's fault. Designing a fault detection and isolation system for a multi input-multi output (MIMO) is done by an unknown input observer, so the system is divided to several subsystems as the effect of other inputs such as disturbing given system state equations. In this observer design method, the effect of these disturbances will weaken and the only fault is detected on specific input. The results of this approach simulation can confirm the ability of the fault detection and isolation system design. After fault detection and isolation, it is necessary to redesign controller based on a suitable modification. In this regard after the use of unknown input observer theory and obtain residual signal and evaluate it, PID controller parameters redesigned for iterative. Stability of the closed loop system has proved in the presence of this method. Also, In order to soften the volatility caused by Annie variations of the PID controller parameters, modifying Sigma as a way acceptable solution used. Finally, the simulation results of three tank popular example confirm the accuracy of performance.Keywords: fault tolerant control, fault detection and isolation, actuator fault, unknown input observer
Procedia PDF Downloads 45722143 Supplemental VisCo-friction Damping for Dynamical Structural Systems
Authors: Sharad Singh, Ajay Kumar Sinha
Abstract:
Coupled dampers like viscoelastic-frictional dampers for supplemental damping are a newer technique. In this paper, innovative Visco-frictional damping models have been presented and investigated. This paper attempts to couple frictional and fluid viscous dampers into a single unit of supplemental dampers. Visco-frictional damping model is developed by series and parallel coupling of frictional and fluid viscous dampers using Maxwell and Kelvin-Voigat models. The time analysis has been performed using numerical simulation on an SDOF system with varying fundamental periods, subject to a set of 12 ground motions. The simulation was performed using the direct time integration method. MATLAB programming tool was used to carry out the numerical simulation. The response behavior has been analyzed for the varying time period and added damping. This paper compares the response reduction behavior of the two modes of coupling. This paper highlights the performance efficiency of the suggested damping models. It also presents a mathematical modeling approach to visco-frictional dampers and simultaneously suggests the suitable mode of coupling between the two sub-units.Keywords: hysteretic damping, Kelvin model, Maxwell model, parallel coupling, series coupling, viscous damping
Procedia PDF Downloads 16122142 Conceptual Model of a Residential Waste Collection System Using ARENA Software
Authors: Bruce G. Wilson
Abstract:
The collection of municipal solid waste at the curbside is a complex operation that is repeated daily under varying circumstances around the world. There have been several attempts to develop Monte Carlo simulation models of the waste collection process dating back almost 50 years. Despite this long history, the use of simulation modeling as a planning or optimization tool for waste collection is still extremely limited in practice. Historically, simulation modeling of waste collection systems has been hampered by the limitations of computer hardware and software and by the availability of representative input data. This paper outlines the development of a Monte Carlo simulation model that overcomes many of the limitations contained in previous models. The model uses a general purpose simulation software program that is easily capable of modeling an entire waste collection network. The model treats the stops on a waste collection route as a queue of work to be processed by a collection vehicle (or server). Input data can be collected from a variety of sources including municipal geographic information systems, global positioning system recorders on collection vehicles, and weigh scales at transfer stations or treatment facilities. The result is a flexible model that is sufficiently robust that it can model the collection activities in a large municipality, while providing the flexibility to adapt to changing conditions on the collection route.Keywords: modeling, queues, residential waste collection, Monte Carlo simulation
Procedia PDF Downloads 40422141 Human Factors Simulation Approach to Analyze Older Drivers’ Performance in Intersections Left-Turn Scenarios
Authors: Yassir AbdelRazig, Eren Ozguven, Ren Moses
Abstract:
While there exists a greater understanding of the differences between the driving behaviors of older and younger drivers, there is still a need to further understand how the two groups perform when attempting to perform complex intersection maneuvers. This paper looks to determine if, and to what extent, these differences exist when drivers encounter permissive left-hand turns, pedestrian traffic, two and four-lane intersections, heavy fog, and night conditions. The study will utilize a driving simulator to develop custom drivable scenarios containing one or more of the previously mentioned conditions. 32 younger and 32 older (+65 years) participants perform driving simulation scenarios and have their velocity, time to the nearest oncoming vehicle, accepted and rejected gaps, etc., recorded. The data collected from the simulator is analyzed via Raff’s method and logistic regression in order to determine and compare the critical gaps values of the two cohorts. Out of the parameters considered for this study, only the age of the driver, their experience (if they are a younger driver), the size of a gap, and the presence of pedestrians on the crosswalk proved significant. The results did not support the hypothesis that older drivers would be significantly more conservative in their critical gaps judgment and acceptance.Keywords: older drivers, simulation, left-turn, human factors
Procedia PDF Downloads 24922140 Analysis and Modeling of Graphene-Based Percolative Strain Sensor
Authors: Heming Yao
Abstract:
Graphene-based percolative strain gauges could find applications in many places such as touch panels, artificial skins or human motion detection because of its advantages over conventional strain gauges such as flexibility and transparency. These strain gauges rely on a novel sensing mechanism that depends on strain-induced morphology changes. Once a compression or tension strain is applied to Graphene-based percolative strain gauges, the overlap area between neighboring flakes becomes smaller or larger, which is reflected by the considerable change of resistance. Tiny strain change on graphene-based percolative strain sensor can act as an important leverage to tremendously increase resistance of strain sensor, which equipped graphene-based percolative strain gauges with higher gauge factor. Despite ongoing research in the underlying sensing mechanism and the limits of sensitivity, neither suitable understanding has been obtained of what intrinsic factors play the key role in adjust gauge factor, nor explanation on how the strain gauge sensitivity can be enhanced, which is undoubtedly considerably meaningful and provides guideline to design novel and easy-produced strain sensor with high gauge factor. We here simulated the strain process by modeling graphene flakes and its percolative networks. We constructed the 3D resistance network by simulating overlapping process of graphene flakes and interconnecting tremendous number of resistance elements which were obtained by fractionizing each piece of graphene. With strain increasing, the overlapping graphenes was dislocated on new stretched simulation graphene flake simulation film and a new simulation resistance network was formed with smaller flake number density. By solving the resistance network, we can get the resistance of simulation film under different strain. Furthermore, by simulation on possible variable parameters, such as out-of-plane resistance, in-plane resistance, flake size, we obtained the changing tendency of gauge factor with all these variable parameters. Compared with the experimental data, we verified the feasibility of our model and analysis. The increase of out-of-plane resistance of graphene flake and the initial resistance of sensor, based on flake network, both improved gauge factor of sensor, while the smaller graphene flake size gave greater gauge factor. This work can not only serve as a guideline to improve the sensitivity and applicability of graphene-based strain sensors in the future, but also provides method to find the limitation of gauge factor for strain sensor based on graphene flake. Besides, our method can be easily transferred to predict gauge factor of strain sensor based on other nano-structured transparent optical conductors, such as nanowire and carbon nanotube, or of their hybrid with graphene flakes.Keywords: graphene, gauge factor, percolative transport, strain sensor
Procedia PDF Downloads 42022139 Dynamic Fault Tree Analysis of Dynamic Positioning System through Monte Carlo Approach
Authors: A. S. Cheliyan, S. K. Bhattacharyya
Abstract:
Dynamic Positioning System (DPS) is employed in marine vessels of the offshore oil and gas industry. It is a computer controlled system to automatically maintain a ship’s position and heading by using its own thrusters. Reliability assessment of the same can be analyzed through conventional fault tree. However, the complex behaviour like sequence failure, redundancy management and priority of failing of events cannot be analyzed by the conventional fault trees. The Dynamic Fault Tree (DFT) addresses these shortcomings of conventional Fault Tree by defining additional gates called dynamic gates. Monte Carlo based simulation approach has been adopted for the dynamic gates. This method of realistic modeling of DPS gives meaningful insight into the system reliability and the ability to improve the same.Keywords: dynamic positioning system, dynamic fault tree, Monte Carlo simulation, reliability assessment
Procedia PDF Downloads 77722138 Temporal and Spacial Adaptation Strategies in Aerodynamic Simulation of Bluff Bodies Using Vortex Particle Methods
Authors: Dario Milani, Guido Morgenthal
Abstract:
Fluid dynamic computation of wind caused forces on bluff bodies e.g light flexible civil structures or high incidence of ground approaching airplane wings, is one of the major criteria governing their design. For such structures a significant dynamic response may result, requiring the usage of small scale devices as guide-vanes in bridge design to control these effects. The focus of this paper is on the numerical simulation of the bluff body problem involving multiscale phenomena induced by small scale devices. One of the solution methods for the CFD simulation that is relatively successful in this class of applications is the Vortex Particle Method (VPM). The method is based on a grid free Lagrangian formulation of the Navier-Stokes equations, where the velocity field is modeled by particles representing local vorticity. These vortices are being convected due to the free stream velocity as well as diffused. This representation yields the main advantages of low numerical diffusion, compact discretization as the vorticity is strongly localized, implicitly accounting for the free-space boundary conditions typical for this class of FSI problems, and a natural representation of the vortex creation process inherent in bluff body flows. When the particle resolution reaches the Kolmogorov dissipation length, the method becomes a Direct Numerical Simulation (DNS). However, it is crucial to note that any solution method aims at balancing the computational cost against the accuracy achievable. In the classical VPM method, if the fluid domain is discretized by Np particles, the computational cost is O(Np2). For the coupled FSI problem of interest, for example large structures such as long-span bridges, the aerodynamic behavior may be influenced or even dominated by small structural details such as barriers, handrails or fairings. For such geometrically complex and dimensionally large structures, resolving the complete domain with the conventional VPM particle discretization might become prohibitively expensive to compute even for moderate numbers of particles. It is possible to reduce this cost either by reducing the number of particles or by controlling its local distribution. It is also possible to increase the accuracy of the solution without increasing substantially the global computational cost by computing a correction of the particle-particle interaction in some regions of interest. In this paper different strategies are presented in order to extend the conventional VPM method to reduce the computational cost whilst resolving the required details of the flow. The methods include temporal sub stepping to increase the accuracy of the particles convection in certain regions as well as dynamically re-discretizing the particle map to locally control the global and the local amount of particles. Finally, these methods will be applied on a test case and the improvements in the efficiency as well as the accuracy of the proposed extension to the method are presented. The important benefits in terms of accuracy and computational cost of the combination of these methods will be thus presented as long as their relevant applications.Keywords: adaptation, fluid dynamic, remeshing, substepping, vortex particle method
Procedia PDF Downloads 266