Search results for: fluid flow
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5782

Search results for: fluid flow

3832 Modeling of Sand Boil near the Danube River

Authors: Edina Koch, Károly Gombás, Márton Maller

Abstract:

The Little Plain is located along the Danube river, and this area is a “hotbed” of sand boil formation. This is due to the combination of a 100-250 m thick gravel layer beneath the Little Plain with a relatively thin blanket of poor soil spreading the gravel with variable thickness. Sand boils have a tradition and history in this area. It was possible to know which sand boil started and stopped working at what water level, and some of them even have names. The authors present a 2D finite element model of groundwater flow through a selected cross-section of the Danube river, which observed activation of piping phenomena during the 2013 flood event. Soil parametrization is based on a complex site investigation program conducted along the Danube River in the Little Plain.

Keywords: site characterization, groundwater flow, numerical modeling, sand boil

Procedia PDF Downloads 83
3831 The Design of a Computer Simulator to Emulate Pathology Laboratories: A Model for Optimising Clinical Workflows

Authors: M. Patterson, R. Bond, K. Cowan, M. Mulvenna, C. Reid, F. McMahon, P. McGowan, H. Cormican

Abstract:

This paper outlines the design of a simulator to allow for the optimisation of clinical workflows through a pathology laboratory and to improve the laboratory’s efficiency in the processing, testing, and analysis of specimens. Often pathologists have difficulty in pinpointing and anticipating issues in the clinical workflow until tests are running late or in error. It can be difficult to pinpoint the cause and even more difficult to predict any issues which may arise. For example, they often have no indication of how many samples are going to be delivered to the laboratory that day or at a given hour. If we could model scenarios using past information and known variables, it would be possible for pathology laboratories to initiate resource preparations, e.g. the printing of specimen labels or to activate a sufficient number of technicians. This would expedite the clinical workload, clinical processes and improve the overall efficiency of the laboratory. The simulator design visualises the workflow of the laboratory, i.e. the clinical tests being ordered, the specimens arriving, current tests being performed, results being validated and reports being issued. The simulator depicts the movement of specimens through this process, as well as the number of specimens at each stage. This movement is visualised using an animated flow diagram that is updated in real time. A traffic light colour-coding system will be used to indicate the level of flow through each stage (green for normal flow, orange for slow flow, and red for critical flow). This would allow pathologists to clearly see where there are issues and bottlenecks in the process. Graphs would also be used to indicate the status of specimens at each stage of the process. For example, a graph could show the percentage of specimen tests that are on time, potentially late, running late and in error. Clicking on potentially late samples will display more detailed information about those samples, the tests that still need to be performed on them and their urgency level. This would allow any issues to be resolved quickly. In the case of potentially late samples, this could help to ensure that critically needed results are delivered on time. The simulator will be created as a single-page web application. Various web technologies will be used to create the flow diagram showing the workflow of the laboratory. JavaScript will be used to program the logic, animate the movement of samples through each of the stages and to generate the status graphs in real time. This live information will be extracted from an Oracle database. As well as being used in a real laboratory situation, the simulator could also be used for training purposes. ‘Bots’ would be used to control the flow of specimens through each step of the process. Like existing software agents technology, these bots would be configurable in order to simulate different situations, which may arise in a laboratory such as an emerging epidemic. The bots could then be turned on and off to allow trainees to complete the tasks required at that step of the process, for example validating test results.

Keywords: laboratory-process, optimization, pathology, computer simulation, workflow

Procedia PDF Downloads 277
3830 Generative Design Method for Cooled Additively Manufactured Gas Turbine Parts

Authors: Thomas Wimmer, Bernhard Weigand

Abstract:

The improvement of gas turbine efficiency is one of the main drivers of research and development in the gas turbine market. This has led to elevated gas turbine inlet temperatures beyond the melting point of the utilized materials. The turbine parts need to be actively cooled in order to withstand these harsh environments. However, the usage of compressor air as coolant decreases the overall gas turbine efficiency. Thus, coolant consumption needs to be minimized in order to gain the maximum advantage from higher turbine inlet temperatures. Therefore, sophisticated cooling designs for gas turbine parts aim to minimize coolant mass flow. New design space is accessible as additive manufacturing is maturing to industrial usage for the creation of hot gas flow path parts. By making use of this technology more efficient cooling schemes can be manufacture. In order to find such cooling schemes a generative design method is being developed. It generates cooling schemes randomly which adhere to a set of rules. These assure the sanity of the design. A huge amount of different cooling schemes are generated and implemented in a simulation environment where it is validated. Criteria for the fitness of the cooling schemes are coolant mass flow, maximum temperature and temperature gradients. This way the whole design space is sampled and a Pareto optimum front can be identified. This approach is applied to a flat plate, which resembles a simplified section of a hot gas flow path part. Realistic boundary conditions are applied and thermal barrier coating is accounted for in the simulation environment. The resulting cooling schemes are presented and compared to representative conventional cooling schemes. Further development of this method can give access to cooling schemes with an even better performance having higher complexity, which makes use of the available design space.

Keywords: additive manufacturing, cooling, gas turbine, heat transfer, heat transfer design, optimization

Procedia PDF Downloads 339
3829 Combustion Chamber Sizing for Energy Recovery from Furnace Process Gas: Waste to Energy

Authors: Balram Panjwani, Bernd Wittgens, Jan Erik Olsen, Stein Tore Johansen

Abstract:

The Norwegian ferroalloy industry is a world leader in sustainable production of ferrosilicon, silicon and manganese alloys with the lowest global specific energy consumption. One of the byproducts during the metal reduction process is energy rich off-gas and usually this energy is not harnessed. A novel concept for sustainable energy recovery from ferroalloy off-gas is discussed. The concept is founded on the idea of introducing a combustion chamber in the off-gas section in which energy rich off-gas mainly consisting of CO will be combusted. This will provide an additional degree of freedom for optimizing energy recovery. A well-controlled and high off-gas temperature will assure a significant increase in energy recovery and reduction of emissions to the atmosphere. Design and operation of the combustion chamber depend on many parameters, including the total power capacity of the combustion chamber, sufficient residence time for combusting the complex Poly Aromatic Hydrocarbon (PAH), NOx, as well as converting other potential pollutants. The design criteria for the combustion chamber have been identified and discussed and sizing of the combustion chamber has been carried out considering these design criteria. Computational Fluid Dynamics (CFD) has been utilized extensively for sizing the combustion chamber. The results from our CFD simulations of the flow in the combustion chamber and exploring different off-gas fuel composition are presented. In brief, the paper covers all aspect which impacts the sizing of the combustion chamber, including insulation thickness, choice of insulating material, heat transfer through extended surfaces, multi-staging and secondary air injection.

Keywords: CFD, combustion chamber, arc furnace, energy recovery

Procedia PDF Downloads 305
3828 Clustered Regularly Interspaced Short Palindromic Repeat/cas9-Based Lateral Flow and Fluorescence Diagnostics for Rapid Pathogen Detection

Authors: Mark Osborn

Abstract:

Clustered, regularly interspaced short palindromic repeat (CRISPR/Cas) proteins can be designed to bind specified DNA and RNA sequences and hold great promise for the accurate detection of nucleic acids for diagnostics. Commercially available reagents were integrated into a CRISPR/Cas9-based lateral flow assay that can detect severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) sequences with single-base specificity. This approach requires minimal equipment and represents a simplified platform for field-based deployment. A rapid, multiplex fluorescence CRISPR/Cas9 nuclease cleavage assay capable of detecting and differentiating SARS-CoV-2, influenza A and B, and respiratory syncytial virus in a single reaction was also developed. These findings provide proof of principle for CRISPR/Cas9 point-of-care diagnosis that can detect specific SARS-CoV-2 strain(s). Further, Cas9 cleavage allows for a scalable fluorescent platform for identifying respiratory viral pathogens with overlapping symptomology. Collectively, this approach is a facile platform for diagnostics with broad application to user-defined sequence interrogation and detection.

Keywords: CRISPR/Cas9, lateral flow assay, SARS-Co-V2, single-nucleotide resolution

Procedia PDF Downloads 173
3827 The Impact of Undisturbed Flow Speed on the Correlation of Aerodynamic Coefficients as a Function of the Angle of Attack for the Gyroplane Body

Authors: Zbigniew Czyz, Krzysztof Skiba, Miroslaw Wendeker

Abstract:

This paper discusses the results of aerodynamic investigation of the Tajfun gyroplane body designed by a Polish company, Aviation Artur Trendak. This gyroplane has been studied as a 1:8 scale model. Scaling objects for aerodynamic investigation is an inherent procedure in any kind of designing. If scaling, the criteria of similarity need to be satisfied. The basic criteria of similarity are geometric, kinematic and dynamic. Despite the results of aerodynamic research are often reduced to aerodynamic coefficients, one should pay attention to how values of coefficients behave if certain criteria are to be satisfied. To satisfy the dynamic criterion, for example, the Reynolds number should be focused on. This is the correlation of inertial to viscous forces. With the multiplied flow speed by the specific dimension as a numerator (with a constant kinematic viscosity coefficient), flow speed in a wind tunnel research should be increased as many times as an object is decreased. The aerodynamic coefficients specified in this research depend on the real forces that act on an object, its specific dimension, medium speed and variations in its density. Rapid prototyping with a 3D printer was applied to create the research object. The research was performed with a T-1 low-speed wind tunnel (its diameter of the measurement volume is 1.5 m) and a six-element aerodynamic internal scales, WDP1, at the Institute of Aviation in Warsaw. This T-1 wind tunnel is low-speed continuous operation with open space measurement. The research covered a number of the selected speeds of undisturbed flow, i.e. V = 20, 30 and 40 m/s, corresponding to the Reynolds numbers (as referred to 1 m) Re = 1.31∙106, 1.96∙106, 2.62∙106 for the angles of attack ranging -15° ≤ α ≤ 20°. Our research resulted in basic aerodynamic characteristics and observing the impact of undisturbed flow speed on the correlation of aerodynamic coefficients as a function of the angle of attack of the gyroplane body. If the speed of undisturbed flow in the wind tunnel changes, the aerodynamic coefficients are significantly impacted. At speed from 20 m/s to 30 m/s, drag coefficient, Cx, changes by 2.4% up to 9.9%, whereas lift coefficient, Cz, changes by -25.5% up to 15.7% if the angle of attack of 0° excluded or by -25.5% up to 236.9% if the angle of attack of 0° included. Within the same speed range, the coefficient of a pitching moment, Cmy, changes by -21.1% up to 7.3% if the angles of attack -15° and -10° excluded or by -142.8% up to 618.4% if the angle of attack -15° and -10° included. These discrepancies in the coefficients of aerodynamic forces definitely need to consider while designing the aircraft. For example, if load of certain aircraft surfaces is calculated, additional correction factors definitely need to be applied. This study allows us to estimate the discrepancies in the aerodynamic forces while scaling the aircraft. This work has been financed by the Polish Ministry of Science and Higher Education.

Keywords: aerodynamics, criteria of similarity, gyroplane, research tunnel

Procedia PDF Downloads 382
3826 Evaluation of Hard Rocks Destruction Effectiveness at Drilling

Authors: Ekaterina Leusheva, Valentin Morenov

Abstract:

Well drilling in hard rocks is coupled with high energy demands which negates the speed of the process and thus reduces overall effectiveness. Aim of this project is to develop the technique of experimental research, which would allow to select optimal washing fluid composition while adding special hardness reducing detergent reagents. Based on the analysis of existing references and conducted experiments, technique dealing with quantitative evaluation of washing fluid weakening influence on drilled rocks was developed, which considers laboratory determination of three mud properties (density, surface tension, specific electrical resistance) and three rock properties (ultimate stress, dynamic strength, micro-hardness). Developed technique can be used in the well drilling technologies and particularly while creating new compositions of drilling muds for increased destruction effectiveness of hard rocks. It can be concluded that given technique introduces coefficient of hard rocks destruction effectiveness that allows quantitative evaluation of different drilling muds on the drilling process to be taken. Correct choice of drilling mud composition with hardness reducing detergent reagents will increase drilling penetration rate and drill meterage per bit.

Keywords: detergent reagents, drilling mud, drilling process stimulation, hard rocks

Procedia PDF Downloads 535
3825 Digital Twin of Real Electrical Distribution System with Real Time Recursive Load Flow Calculation and State Estimation

Authors: Anosh Arshad Sundhu, Francesco Giordano, Giacomo Della Croce, Maurizio Arnone

Abstract:

Digital Twin (DT) is a technology that generates a virtual representation of a physical system or process, enabling real-time monitoring, analysis, and simulation. DT of an Electrical Distribution System (EDS) can perform online analysis by integrating the static and real-time data in order to show the current grid status and predictions about the future status to the Distribution System Operator (DSO), producers and consumers. DT technology for EDS also offers the opportunity to DSO to test hypothetical scenarios. This paper discusses the development of a DT of an EDS by Smart Grid Controller (SGC) application, which is developed using open-source libraries and languages. The developed application can be integrated with Supervisory Control and Data Acquisition System (SCADA) of any EDS for creating the DT. The paper shows the performance of developed tools inside the application, tested on real EDS for grid observability, Smart Recursive Load Flow (SRLF) calculation and state estimation of loads in MV feeders.

Keywords: digital twin, distributed energy resources, remote terminal units, supervisory control and data acquisition system, smart recursive load flow

Procedia PDF Downloads 91
3824 Dynamic Characterization of Shallow Aquifer Groundwater: A Lab-Scale Approach

Authors: Anthony Credoz, Nathalie Nief, Remy Hedacq, Salvador Jordana, Laurent Cazes

Abstract:

Groundwater monitoring is classically performed in a network of piezometers in industrial sites. Groundwater flow parameters, such as direction, sense and velocity, are deduced from indirect measurements between two or more piezometers. Groundwater sampling is generally done on the whole column of water inside each borehole to provide concentration values for each piezometer location. These flow and concentration values give a global ‘static’ image of potential plume of contaminants evolution in the shallow aquifer with huge uncertainties in time and space scales and mass discharge dynamic. TOTAL R&D Subsurface Environmental team is challenging this classical approach with an innovative dynamic way of characterization of shallow aquifer groundwater. The current study aims at optimizing the tools and methodologies for (i) a direct and multilevel measurement of groundwater velocities in each piezometer and, (ii) a calculation of potential flux of dissolved contaminant in the shallow aquifer. Lab-scale experiments have been designed to test commercial and R&D tools in a controlled sandbox. Multiphysics modeling were performed and took into account Darcy equation in porous media and Navier-Stockes equation in the borehole. The first step of the current study focused on groundwater flow at porous media/piezometer interface. Huge uncertainties from direct flow rate measurements in the borehole versus Darcy flow rate in the porous media were characterized during experiments and modeling. The structure and location of the tools in the borehole also impacted the results and uncertainties of velocity measurement. In parallel, direct-push tool was tested and presented more accurate results. The second step of the study focused on mass flux of dissolved contaminant in groundwater. Several active and passive commercial and R&D tools have been tested in sandbox and reactive transport modeling has been performed to validate the experiments at the lab-scale. Some tools will be selected and deployed in field assays to better assess the mass discharge of dissolved contaminants in an industrial site. The long-term subsurface environmental strategy is targeting an in-situ, real-time, remote and cost-effective monitoring of groundwater.

Keywords: dynamic characterization, groundwater flow, lab-scale, mass flux

Procedia PDF Downloads 156
3823 Hydrodynamic Characterisation of a Hydraulic Flume with Sheared Flow

Authors: Daniel Rowe, Christopher R. Vogel, Richard H. J. Willden

Abstract:

The University of Oxford’s recirculating water flume is a combined wave and current test tank with a 1 m depth, 1.1 m width, and 10 m long working section, and is capable of flow speeds up to 1 ms−1 . This study documents the hydrodynamic characteristics of the facility in preparation for experimental testing of horizontal axis tidal stream turbine models. The turbine to be tested has a rotor diameter of 0.6 m and is a modified version of one of two model-scale turbines tested in previous experimental campaigns. An Acoustic Doppler Velocimeter (ADV) was used to measure the flow at high temporal resolution at various locations throughout the flume, enabling the spatial uniformity and turbulence flow parameters to be investigated. The mean velocity profiles exhibited high levels of spatial uniformity at the design speed of the flume, 0.6 ms−1 , with variations in the three-dimensional velocity components on the order of ±1% at the 95% confidence level, along with a modest streamwise acceleration through the measurement domain, a target 5 m working section of the flume. A high degree of uniformity was also apparent for the turbulence intensity, with values ranging between 1-2% across the intended swept area of the turbine rotor. The integral scales of turbulence exhibited a far higher degree of variation throughout the water column, particularly in the streamwise and vertical scales. This behaviour is believed to be due to the high signal noise content leading to decorrelation in the sampling records. To achieve more realistic levels of vertical velocity shear in the flume, a simple procedure to practically generate target vertical shear profiles in open-channel flows is described. Here, the authors arranged a series of non-uniformly spaced parallel bars placed across the width of the flume and normal to the onset flow. By adjusting the resistance grading across the height of the working section, the downstream profiles could be modified accordingly, characterised by changes in the velocity profile power law exponent, 1/n. Considering the significant temporal variation in a tidal channel, the choice of the exponent denominator, n = 6 and n = 9, effectively provides an achievable range around the much-cited value of n = 7 observed at many tidal sites. The resulting flow profiles, which we intend to use in future turbine tests, have been characterised in detail. The results indicate non-uniform vertical shear across the survey area and reveal substantial corner flows, arising from the differential shear between the target vertical and cross-stream shear profiles throughout the measurement domain. In vertically sheared flow, the rotor-equivalent turbulence intensity ranges between 3.0-3.8% throughout the measurement domain for both bar arrangements, while the streamwise integral length scale grows from a characteristic dimension on the order of the bar width, similar to the flow downstream of a turbulence-generating grid. The experimental tests are well-defined and repeatable and serve as a reference for other researchers who wish to undertake similar investigations.

Keywords: acoustic doppler Velocimeter, experimental hydrodynamics, open-channel flow, shear profiles, tidal stream turbines

Procedia PDF Downloads 71
3822 Dynamic Stall Characterization of Low Reynolds Airfoil in Mars and Titan’s Atmosphere

Authors: Vatasta Koul, Vaibhav Sharma, Ayush Gupta, Rajesh Yadav

Abstract:

Exploratory missions to Mars and Titan have increased recently with various endeavors to find an alternate home to humankind. The use of surface rovers has its limitations due to rugged and uneven surfaces of these planetary bodies. The use of aerial robots requires the complete aerodynamic characterization of these vehicles in the atmospheric conditions of these planetary bodies. The dynamic stall phenomenon is extremely important for rotary wings performance under low Reynolds number that can be encountered in Martian and Titan’s atmosphere. The current research focuses on the aerodynamic characterization and exploration of the dynamic stall phenomenon of two different airfoils viz. E387 and Selig-Donovan7003 in Martian and Titan’s atmosphere at low Reynolds numbers of 10000 and 50000. The two-dimensional numerical simulations are conducted using commercially available finite volume solver with multi-species non-reacting mixture of gases as the working fluid. The k-epsilon (k-ε) turbulence model is used to capture the unsteady flow separation and the effect of turbulence. The dynamic characteristics are studied at a fixed different constant rotational extreme of angles of attack. This study of airfoils at different low Reynolds number and atmospheric conditions on Mars and Titan will be resulting in defining the aerodynamic characteristics of these airfoils for unmanned aerial missions for outer space exploration.

Keywords: aerodynamics, dynamic stall, E387, SD7003

Procedia PDF Downloads 121
3821 Main Control Factors of Fluid Loss in Drilling and Completion in Shunbei Oilfield by Unmanned Intervention Algorithm

Authors: Peng Zhang, Lihui Zheng, Xiangchun Wang, Xiaopan Kou

Abstract:

Quantitative research on the main control factors of lost circulation has few considerations and single data source. Using Unmanned Intervention Algorithm to find the main control factors of lost circulation adopts all measurable parameters. The degree of lost circulation is characterized by the loss rate as the objective function. Geological, engineering and fluid data are used as layers, and 27 factors such as wellhead coordinates and WOB are used as dimensions. Data classification is implemented to determine function independent variables. The mathematical equation of loss rate and 27 influencing factors is established by multiple regression method, and the undetermined coefficient method is used to solve the undetermined coefficient of the equation. Only three factors in t-test are greater than the test value 40, and the F-test value is 96.557%, indicating that the correlation of the model is good. The funnel viscosity, final shear force and drilling time were selected as the main control factors by elimination method, contribution rate method and functional method. The calculated values of the two wells used for verification differ from the actual values by -3.036m3/h and -2.374m3/h, with errors of 7.21% and 6.35%. The influence of engineering factors on the loss rate is greater than that of funnel viscosity and final shear force, and the influence of the three factors is less than that of geological factors. Quantitatively calculate the best combination of funnel viscosity, final shear force and drilling time. The minimum loss rate of lost circulation wells in Shunbei area is 10m3/h. It can be seen that man-made main control factors can only slow down the leakage, but cannot fundamentally eliminate it. This is more in line with the characteristics of karst caves and fractures in Shunbei fault solution oil and gas reservoir.

Keywords: drilling and completion, drilling fluid, lost circulation, loss rate, main controlling factors, unmanned intervention algorithm

Procedia PDF Downloads 98
3820 Experimental Characterization of Flowable Cement Pastes Made with Marble Waste

Authors: F. Messaoudi, O. Haddad, R. Bouras, S. Kaci

Abstract:

The development of self-compacting concrete (SCC) marks a huge step towards improved efficiency and working conditions on construction sites and in the precast industry. SCC flows easily into more complex shapes and through reinforcement bars, reduces the manpower required for the placement; no vibration is required to ensure correct compaction of concrete. This concrete contains a high volume of binder which is controlled by their rheological behavior. The paste consists of binders (Portland cement with or without supplementary cementitious materials), water, chemical admixtures and fillers. In this study, two series of tests were performed on self-compacting cement pastes made with marble waste additions as the mineral addition. The first series of this investigation was to determine the flow time of paste using Marsh cone, the second series was to determine the rheological parameters of the same paste namely yield stress and plastic viscosity using the rheometer Haake RheoStress 1. The results of this investigation allowed us to study the evolution of the yield stress, viscosity and the flow time Marsh cone paste as a function of the composition of the paste. A correlation between the results obtained on the flow test Marsh cone and those of the plastic viscosity on the mottled different cement pastes is proposed.

Keywords: adjuvant, rheological parameter, self-compacting cement pastes, waste marble

Procedia PDF Downloads 265
3819 Reduction of Specific Energy Consumption in Microfiltration of Bacillus velezensis Broth by Air Sparging and Turbulence Promoter

Authors: Jovana Grahovac, Ivana Pajcin, Natasa Lukic, Jelena Dodic, Aleksandar Jokic

Abstract:

To obtain purified biomass to be used in the plant pathogen biocontrol or as soil biofertilizer, it is necessary to eliminate residual broth components at the end of the fermentation process. The main drawback of membrane separation techniques is permeate flux decline due to the membrane fouling. Fouling mitigation measures increase the pressure drop along membrane channel due to the increased resistance to flow of the feed suspension, thus increasing the hydraulic power drop. At the same time, these measures lead to an increase in the permeate flux due to the reduced resistance of the filtration cake on the membrane surface. Because of these opposing effects, the energy efficiency of fouling mitigation measures is limited, and the justification of its application is provided by information on a reducing specific energy consumption compared to a case without any measures employed. In this study, the influence of static mixer (Kenics) and air-sparging (two-phase flow) on reduction of specific energy consumption (ER) was investigated. Cultivation Bacillus velezensis was carried out in the 3-L bioreactor (Biostat® Aplus) containing 2 L working volume with two parallel Rushton turbines and without internal baffles. Cultivation was carried out at 28 °C on at 150 rpm with an aeration rate of 0.75 vvm during 96 h. The experiments were carried out in a conventional cross-flow microfiltration unit. During experiments, permeate and retentate were recycled back to the broth vessel to simulate continuous process. The single channel ceramic membrane (TAMI Deutschland) used had a nominal pore size 200 nm with the length of 250 mm and an inner/external diameter of 6/10 mm. The useful membrane channel surface was 4.33×10⁻³ m². Air sparging was brought by the pressurized air connected by a three-way valve to the feed tube by a simple T-connector without diffusor. The different approaches to flux improvement are compared in terms of energy consumption. Reduction of specific energy consumption compared to microfiltration without fouling mitigation is around 49% and 63%, for use of two-phase flow and a static mixer, respectively. In the case of a combination of these two fouling mitigation methods, ER is 60%, i.e., slightly lower compared to the use of turbulence promoter alone. The reason for this result can be found in the fact that flux increase is more affected by the presence of a Kenics static mixer while sparging results in an increase of energy used during microfiltration. By comparing combined method with turbulence promoter flux enhancement method ER is negative (-7%) which can be explained by increased power consumption for air flow with moderate contribution to the flux increase. Another confirmation for this fact can be found by comparing energy consumption values for combined method with energy consumption in the case of two-phase flow. In this instance energy reduction (ER) is 22% that demonstrates that turbulence promoter is more efficient compared to two phase flow. Antimicrobial activity of Bacillus velezensis biomass against phytopathogenic isolates Xanthomonas campestris was preserved under different fouling reduction methods.

Keywords: Bacillus velezensis, microfiltration, static mixer, two-phase flow

Procedia PDF Downloads 108
3818 The Effect of Improvement Programs in the Mean Time to Repair and in the Mean Time between Failures on Overall Lead Time: A Simulation Using the System Dynamics-Factory Physics Model

Authors: Marcel Heimar Ribeiro Utiyama, Fernanda Caveiro Correia, Dario Henrique Alliprandini

Abstract:

The importance of the correct allocation of improvement programs is of growing interest in recent years. Due to their limited resources, companies must ensure that their financial resources are directed to the correct workstations in order to be the most effective and survive facing the strong competition. However, to our best knowledge, the literature about allocation of improvement programs does not analyze in depth this problem when the flow shop process has two capacity constrained resources. This is a research gap which is deeply studied in this work. The purpose of this work is to identify the best strategy to allocate improvement programs in a flow shop with two capacity constrained resources. Data were collected from a flow shop process with seven workstations in an industrial control and automation company, which process 13.690 units on average per month. The data were used to conduct a simulation with the System Dynamics-Factory Physics model. The main variables considered, due to their importance on lead time reduction, were the mean time between failures and the mean time to repair. The lead time reduction was the output measure of the simulations. Ten different strategies were created: (i) focused time to repair improvement, (ii) focused time between failures improvement, (iii) distributed time to repair improvement, (iv) distributed time between failures improvement, (v) focused time to repair and time between failures improvement, (vi) distributed time to repair and between failures improvement, (vii) hybrid time to repair improvement, (viii) hybrid time between failures improvements, (ix) time to repair improvement strategy towards the two capacity constrained resources, (x) time between failures improvement strategy towards the two capacity constrained resources. The ten strategies tested are variations of the three main strategies for improvement programs named focused, distributed and hybrid. Several comparisons among the effect of the ten strategies in lead time reduction were performed. The results indicated that for the flow shop analyzed, the focused strategies delivered the best results. When it is not possible to perform a large investment on the capacity constrained resources, companies should use hybrid approaches. An important contribution to the academy is the hybrid approach, which proposes a new way to direct the efforts of improvements. In addition, the study in a flow shop with two strong capacity constrained resources (more than 95% of utilization) is an important contribution to the literature. Another important contribution is the problem of allocation with two CCRs and the possibility of having floating capacity constrained resources. The results provided the best improvement strategies considering the different strategies of allocation of improvement programs and different positions of the capacity constrained resources. Finally, it is possible to state that both strategies, hybrid time to repair improvement and hybrid time between failures improvement, delivered best results compared to the respective distributed strategies. The main limitations of this study are mainly regarding the flow shop analyzed. Future work can further investigate different flow shop configurations like a varying number of workstations, different number of products or even different positions of the two capacity constrained resources.

Keywords: allocation of improvement programs, capacity constrained resource, hybrid strategy, lead time, mean time to repair, mean time between failures

Procedia PDF Downloads 110
3817 Intelligent Ambulance with Advance Features of Traffic Management and Telecommunication

Authors: Mamatha M. N.

Abstract:

Traffic problems, congested traffic, and flow management were recognized as major problems mostly in all the areas, which have caused a problem for the ambulance which carries the emergency patient. The proposed paper aims in the development of ambulance which reaches the nearby hospital faster even in heavy traffic scenario. This process is activated by implementing hardware in an ambulance as well as in traffic post thus allowing a smooth flow to the ambulance to reach the hospital in time. 1) The design of the vehicle to have a communication between ambulance and traffic post. 2)Electronic Health Record with Data-acquisition system 3)Telemetry of acquired biological parameters to the nearest hospital. Thus interfacing all these three different modules and integrating them on the ambulance could reach the hospital earlier than the present ambulance. The system is accurate and efficient of 99.8%.

Keywords: bio-telemetry, data acquisition, patient database, automatic traffic control

Procedia PDF Downloads 305
3816 Enhancement in Digester Efficiency and Numerical Analysis for Optimal Design Parameters of Biogas Plant Using Design of Experiment Approach

Authors: Rajneesh, Priyanka Singh

Abstract:

Biomass resources have been one of the main energy sources for mankind since the dawn of civilization. There is a vast scope to convert these energy sources into biogas which is a clean, low carbon technology for efficient management and conversion of fermentable organic wastes into a cheap and versatile fuel and bio/organic manure. Thus, in order to enhance the performance of anaerobic digester, an optimizing analysis of resultant parameters (organic dry matter (oDM) content, methane percentage, and biogas yield) has been done for a plug flow anaerobic digester having mesophilic conditions (20-40°C) with the wet fermentation process. Based on the analysis, correlations for oDM, methane percentage, and biogas yield are derived using multiple regression analysis. A statistical model is developed to correlate the operating variables using the design of experiment approach by selecting central composite design (CCD) of a response surface methodology. Results shown in the paper indicates that as the operating temperature increases the efficiency of digester gets improved provided that the pH and hydraulic retention time (HRT) remains constant. Working in an optimized range of carbon-nitrogen ratio for the plug flow digester, the output parameters show a positive change with the variation of dry matter content (DM).

Keywords: biogas, digester efficiency, design of experiment, plug flow digester

Procedia PDF Downloads 363
3815 Development of an Automatic Control System for ex vivo Heart Perfusion

Authors: Pengzhou Lu, Liming Xin, Payam Tavakoli, Zhonghua Lin, Roberto V. P. Ribeiro, Mitesh V. Badiwala

Abstract:

Ex vivo Heart Perfusion (EVHP) has been developed as an alternative strategy to expand cardiac donation by enabling resuscitation and functional assessment of hearts donated from marginal donors, which were previously not accepted. EVHP parameters, such as perfusion flow (PF) and perfusion pressure (PP) are crucial for optimal organ preservation. However, with the heart’s constant physiological changes during EVHP, such as coronary vascular resistance, manual control of these parameters is rendered imprecise and cumbersome for the operator. Additionally, low control precision and the long adjusting time may lead to irreversible damage to the myocardial tissue. To solve this problem, an automatic heart perfusion system was developed by applying a Human-Machine Interface (HMI) and a Programmable-Logic-Controller (PLC)-based circuit to control PF and PP. The PLC-based control system collects the data of PF and PP through flow probes and pressure transducers. It has two control modes: the RPM-flow mode and the pressure mode. The RPM-flow control mode is an open-loop system. It influences PF through providing and maintaining the desired speed inputted through the HMI to the centrifugal pump with a maximum error of 20 rpm. The pressure control mode is a closed-loop system where the operator selects a target Mean Arterial Pressure (MAP) to control PP. The inputs of the pressure control mode are the target MAP, received through the HMI, and the real MAP, received from the pressure transducer. A PID algorithm is applied to maintain the real MAP at the target value with a maximum error of 1mmHg. The precision and control speed of the RPM-flow control mode were examined by comparing the PLC-based system to an experienced operator (EO) across seven RPM adjustment ranges (500, 1000, 2000 and random RPM changes; 8 trials per range) tested in a random order. System’s PID algorithm performance in pressure control was assessed during 10 EVHP experiments using porcine hearts. Precision was examined through monitoring the steady-state pressure error throughout perfusion period, and stabilizing speed was tested by performing two MAP adjustment changes (4 trials per change) of 15 and 20mmHg. A total of 56 trials were performed to validate the RPM-flow control mode. Overall, the PLC-based system demonstrated the significantly faster speed than the EO in all trials (PLC 1.21±0.03, EO 3.69±0.23 seconds; p < 0.001) and greater precision to reach the desired RPM (PLC 10±0.7, EO 33±2.7 mean RPM error; p < 0.001). Regarding pressure control, the PLC-based system has the median precision of ±1mmHg error and the median stabilizing times in changing 15 and 20mmHg of MAP are 15 and 19.5 seconds respectively. The novel PLC-based control system was 3 times faster with 60% less error than the EO for RPM-flow control. In pressure control mode, it demonstrates a high precision and fast stabilizing speed. In summary, this novel system successfully controlled perfusion flow and pressure with high precision, stability and a fast response time through a user-friendly interface. This design may provide a viable technique for future development of novel heart preservation and assessment strategies during EVHP.

Keywords: automatic control system, biomedical engineering, ex-vivo heart perfusion, human-machine interface, programmable logic controller

Procedia PDF Downloads 163
3814 Impact of the Dog-Technic for D1-D4 and Longitudinal Stroke Technique for Diaphragm on Peak Expiratory Flow (PEF) in Asthmatic Patients

Authors: Victoria Eugenia Garnacho-Garnacho, Elena Sonsoles Rodriguez-Lopez, Raquel Delgado-Delgado, Alvaro Otero-Campos, Jesus Guodemar-Perez, Angelo Michelle Vagali, Juan Pablo Hervas-Perez

Abstract:

Asthma is a heterogeneous disease which has always had a drug treatment. Osteopathic treatment that we propose is aimed, seen through a dorsal manipulation (Dog Technic D1-D4) and a technique for diaphragm (Longitudinal Stroke) forced expiratory flow in spirometry changes there are in particular that there is an increase in the volumes of the Peak Flow and Post intervention and effort and that the application of these two techniques together is more powerful if we applied only a Longitudinal (Stroke). Also rating if this type of treatment will have repercussions on breathlessness, a very common symptom in asthma. And finally to investigate if provided vertebra pain decreased after a manipulation. Methods—Participants were recruited between students and professors of the University, aged 18-65, patients (n = 18) were assigned randomly to one of the two groups, group 1 (longitudinal Stroke and manipulation dorsal Dog Technic) and group 2 (diaphragmatic technique, Longitudinal Stroke). The statistical analysis is characterized by the comparison of the main indicator of obstruction of via area PEF (peak expiratory flow) in various situations through the peak flow meter Datospir Peak-10. The measurements were carried out in four phases: at rest, after the stress test, after the treatment, after treatment and the stress test. After each stress test was evaluated, through the Borg scale, the level of Dyspnea on each patient, regardless of the group. In Group 1 in addition to these parameters was calculated using an algometer spinous pain before and after the manipulation. All data were taken at the minute. Results—12 Group 1 (Dog Technic and Longitudinal Stroke) patients responded positively to treatment, there was an increase of 5.1% and 6.1% of the post-treatment PEF and post-treatment, and effort. The results of the scale of Borg by which we measure the level of Dyspnea were positive, a 54.95%, patients noted an improvement in breathing. In addition was confirmed through the means of both groups group 1 in which two techniques were applied was 34.05% more effective than group 2 in which applied only a. After handling pain fell by 38% of the cases. Conclusions—The impact of the technique of Dog-Technic for D1-D4 and the Longitudinal Stroke technique for diaphragm in the volumes of peak expiratory flow (PEF) in asthmatic patients were positive, there was a change of the PEF Post intervention and post-treatment, and effort and showed the most effective group in which only a technique was applied. Furthermore this type of treatment decreased facilitated vertebrae pain and was efficient in the improvement of Dyspnea and the general well-being of the patient.

Keywords: ANS, asthma, manipulation, manual therapy, osteopathic

Procedia PDF Downloads 281
3813 Trace Element Compositions of Placer Gold Samples: Implication for Gold Exploration in Northern Cameroon

Authors: Yanick Blaise Ketchaya, Taofa Zhou

Abstract:

The type of primary source of gold deposit can be explored by using the study of trace element analysis of placer gold which is a valuable exploration tool. Au-bearing deposits are investigated through the placer gold, which is an important indicator mineral. The hydrothermal fluid interacting with diverse geological settings exerts an important function on the chemical composition of gold. Consequently, alluvial gold particles from the placer deposits within the Gamba district in northern Cameroon were examined by an electron probe microanalyzer (EPMA) to show discriminant chemical signatures. The gold grains from a different locality show the same trace element composition, which appears to be in a solid solution in Au. These trace element compositions, contained in gold grains, indicate a homogeneous source. The placer gold particles have significant chemical characteristics (low Ag content), consistent with a mesothermal source. The gold particle signatures in the Gamba district, with high Te and Bi contents, reflect the chemical characteristics of the felsic host rock superimposed on the chemical signature of the hydrothermal fluid.

Keywords: hypogene source, Northern Cameroon, placer gold, trace element

Procedia PDF Downloads 95
3812 Rheological Behavior of Fresh Activated Sludge

Authors: Salam K. Al-Dawery

Abstract:

Despite of few research works on municipal sludge, still there is a lack of actual data. Thus, this work was focused on the conditioning and rheology of fresh activated sludge. The effect of cationic polyelectrolyte has been investigated at different concentrations and pH values in a comparative fashion. Yield stress is presented in all results indicating the minimum stress that necessary to reach flow conditions. Connections between particle-particle is the reason for this yield stress, also, the addition of polyelectrolyte causes strong bonds between particles and water resulting in the aggregation of particles which required higher shear stress in order to flow. The results from the experiments indicate that the cationic polyelectrolytes have significant effluence on the sludge characteristic and water quality such as turbidity, SVI, zone settling rate and shear stress.

Keywords: rheology, polyelectrolyte, settling volume index, turbidity

Procedia PDF Downloads 347
3811 Experimental Investigation of Hull Form for Electric Driven Ferry

Authors: Vasilij Djackov, Tomas Zapnickas, Evgenii Iamshchikov, Lukas Norkevicius, Rima Mickeviciene, Larisa Vasiljeva

Abstract:

In this paper, the resistance and pitching values of the test of an electric ferry are presented. The research was carried out in the open flow channel of Klaipėda University with a multi-axis dynamometer. The received model resistance values were recalculated to the real vessel and the preliminary chosen propulsion unit power was compared. After analyzing the results of the pitching of the model, it was concluded that the shape of the hull needs to be further improved, taking into account the possible uneven weight distribution at the ends of the ferry. Further investigation of the hull of the electric ferry is recommended, including experiments with various water depths and activation of propulsion units.

Keywords: electrical ferry, model tests, open flow channel, pitching, resistance

Procedia PDF Downloads 82
3810 A Mathematical Model of Power System State Estimation for Power Flow Solution

Authors: F. Benhamida, A. Graa, L. Benameur, I. Ziane

Abstract:

The state estimation of the electrical power system operation state is very important for supervising task. With the nonlinearity of the AC power flow model, the state estimation problem (SEP) is a nonlinear mathematical problem with many local optima. This paper treat the mathematical model for the SEP and the monitoring of the nonlinear systems of great dimensions with an application on power electrical system, the modelling, the analysis and state estimation synthesis in order to supervise the power system behavior. in fact, it is very difficult, to see impossible, (for reasons of accessibility, techniques and/or of cost) to measure the excessive number of the variables of state in a large-sized system. It is thus important to develop software sensors being able to produce a reliable estimate of the variables necessary for the diagnosis and also for the control.

Keywords: power system, state estimation, robustness, observability

Procedia PDF Downloads 506
3809 M-Number of Aortic Cannulas Applied During Hypothermic Cardiopulmonary Bypass

Authors: Won-Gon Kim

Abstract:

A standardized system to describe the pressure-flow characteristics of a given cannula has recently been proposed and has been termed ‘the M-number’. Using three different sizes of aortic cannulas in 50 pediatric cardiac patients on hypothermic cardiopulmonary bypass, we analyzed the correlation between experimentally and clinically derived M-numbers, and found this was positive. Clinical M-numbers were typically 0.35 to 0.55 greater than experimental M-numbers, and correlated inversely with a patient's temperature change; this was most probably due to increased blood viscosity, arising from hypothermia. This inverse relationship was more marked in higher M-number cannulas. The clinical data obtained in this study suggest that experimentally derived M-numbers correlate strongly with clinical performance of the cannula, and that the influence of temperature is significant.

Keywords: cardiopulmonary bypass, M-number, aortic cannula, pressure-flow characteristics

Procedia PDF Downloads 232
3808 Numerical and Experimental Investigation of Air Distribution System of Larder Type Refrigerator

Authors: Funda Erdem Şahnali, Ş. Özgür Atayılmaz, Tolga N. Aynur

Abstract:

Almost all of the domestic refrigerators operate on the principle of the vapor compression refrigeration cycle and removal of heat from the refrigerator cabinets is done via one of the two methods: natural convection or forced convection. In this study, airflow and temperature distributions inside a 375L no-frost type larder cabinet, in which cooling is provided by forced convection, are evaluated both experimentally and numerically. Airflow rate, compressor capacity and temperature distribution in the cooling chamber are known to be some of the most important factors that affect the cooling performance and energy consumption of a refrigerator. The objective of this study is to evaluate the original temperature distribution in the larder cabinet, and investigate for better temperature distribution solutions throughout the refrigerator domain via system optimizations that could provide uniform temperature distribution. The flow visualization and airflow velocity measurements inside the original refrigerator are performed via Stereoscopic Particle Image Velocimetry (SPIV). In addition, airflow and temperature distributions are investigated numerically with Ansys Fluent. In order to study the heat transfer inside the aforementioned refrigerator, forced convection theories covering the following cases are applied: closed rectangular cavity representing heat transfer inside the refrigerating compartment. The cavity volume has been represented with finite volume elements and is solved computationally with appropriate momentum and energy equations (Navier-Stokes equations). The 3D model is analyzed as transient, with k-ε turbulence model and SIMPLE pressure-velocity coupling for turbulent flow situation. The results obtained with the 3D numerical simulations are in quite good agreement with the experimental airflow measurements using the SPIV technique. After Computational Fluid Dynamics (CFD) analysis of the baseline case, the effects of three parameters: compressor capacity, fan rotational speed and type of shelf (glass or wire) are studied on the energy consumption; pull down time, temperature distributions in the cabinet. For each case, energy consumption based on experimental results is calculated. After the analysis, the main effective parameters for temperature distribution inside a cabin and energy consumption based on CFD simulation are determined and simulation results are supplied for Design of Experiments (DOE) as input data for optimization. The best configuration with minimum energy consumption that provides minimum temperature difference between the shelves inside the cabinet is determined.

Keywords: air distribution, CFD, DOE, energy consumption, experimental, larder cabinet, refrigeration, uniform temperature

Procedia PDF Downloads 90
3807 Removal of Rhodamine B from Aqueous Solution Using Natural Clay by Fixed Bed Column Method

Authors: A. Ghribi, M. Bagane

Abstract:

The discharge of dye in industrial effluents is of great concern because their presence and accumulation have a toxic or carcinogenic effect on living species. The removal of such compounds at such low levels is a difficult problem. The adsorption process is an effective and attractive proposition for the treatment of dye contaminated wastewater. Activated carbon adsorption in fixed beds is a very common technology in the treatment of water and especially in processes of decolouration. However, it is expensive and the powdered one is difficult to be separated from aquatic system when it becomes exhausted or the effluent reaches the maximum allowable discharge level. The regeneration of exhausted activated carbon by chemical and thermal procedure is also expensive and results in loss of the sorbent. The focus of this research was to evaluate the adsorption potential of the raw clay in removing rhodamine B from aqueous solutions using a laboratory fixed-bed column. The continuous sorption process was conducted in this study in order to simulate industrial conditions. The effect of process parameters, such as inlet flow rate, adsorbent bed height, and initial adsorbate concentration on the shape of breakthrough curves was investigated. A glass column with an internal diameter of 1.5 cm and height of 30 cm was used as a fixed-bed column. The pH of feed solution was set at 8.5. Experiments were carried out at different bed heights (5 - 20 cm), influent flow rates (1.6- 8 mL/min) and influent rhodamine B concentrations (20 - 80 mg/L). The obtained results showed that the adsorption capacity increases with the bed depth and the initial concentration and it decreases at higher flow rate. The column regeneration was possible for four adsorption–desorption cycles. The clay column study states the value of the excellent adsorption capacity for the removal of rhodamine B from aqueous solution. Uptake of rhodamine B through a fixed-bed column was dependent on the bed depth, influent rhodamine B concentration, and flow rate.

Keywords: adsorption, breakthrough curve, clay, fixed bed column, rhodamine b, regeneration

Procedia PDF Downloads 263
3806 Environmental Engineering Case Study of Waste Water Treatement

Authors: Harold Jideofor

Abstract:

Wastewater treatment consists of applying known technology to improve or upgrade the quality of a wastewater. Usually wastewater treatment will involve collecting the wastewater in a central, segregated location (the Wastewater Treatment Plant) and subjecting the wastewater to various treatment processes. Most often, since large volumes of wastewater are involved, treatment processes are carried out on continuously flowing wastewaters (continuous flow or "open" systems) rather than as "batch" or a series of periodic treatment processes in which treatment is carried out on parcels or "batches" of wastewaters. While most wastewater treatment processes are continuous flow, certain operations, such as vacuum filtration, involving storage of sludge, the addition of chemicals, filtration and removal or disposal of the treated sludge, are routinely handled as periodic batch operations.

Keywords: wastewater treatment, environmental engineering, waste water

Procedia PDF Downloads 564
3805 Methanol Steam Reforming with Heat Recovery for Hydrogen-Rich Gas Production

Authors: Horng-Wen Wu, Yi Chao, Rong-Fang Horng

Abstract:

This study is to develop a methanol steam reformer with a heat recovery zone, which recovers heat from exhaust gas of a diesel engine, and to investigate waste heat recovery ratio at the required reaction temperature. The operation conditions of the reformer are reaction temperature (200 °C, 250 °C, and 300 °C), steam to carbonate (S/C) ratio (0.9, 1.1, and 1.3), and N2 volume flow rate (40 cm3/min, 70 cm3/min, and 100 cm3/min). Finally, the hydrogen concentration, the CO, CO2, and N2 concentrations are measured and recorded to calculate methanol conversion efficiency, hydrogen flow rate, and assisting combustion gas and impeding combustion gas ratio. The heat source of this reformer comes from electric heater and waste heat of exhaust gas from diesel engines. The objective is to recover waste heat from the engine and to make more uniform temperature distribution within the reformer. It is beneficial for the reformer to enhance the methanol conversion efficiency and hydrogen-rich gas production. Experimental results show that the highest hydrogen flow rate exists at N2 of the volume rate 40 cm3/min and reforming reaction temperature of 300 °C and the value is 19.6 l/min. With the electric heater and heat recovery from exhaust gas, the maximum heat recovery ratio is 13.18 % occurring at water-methanol (S/C) ratio of 1.3 and the reforming reaction temperature of 300 °C.

Keywords: heat recovery, hydrogen-rich production, methanol steam reformer, methanol conversion efficiency

Procedia PDF Downloads 453
3804 Experimental and Computational Fluid Dynamics Analysis of Horizontal Axis Wind Turbine

Authors: Saim Iftikhar Awan, Farhan Ali

Abstract:

Wind power has now become one of the most important resources of renewable energy. The machine which extracts kinetic energy from wind is wind turbine. This work is all about the electrical power analysis of horizontal axis wind turbine to check the efficiency of different configurations of wind turbines to get maximum output and comparison of experimental and Computational Fluid Dynamics (CFD) results. Different experiments have been performed to obtain that configuration with the help of which we can get the maximum electrical power output by changing the different parameters like the number of blades, blade shape, wind speed, etc. in first step experimentation is done, and then the similar configuration is designed in 3D CAD software. After a series of experiments, it has been found that the turbine with four blades at an angle of 75° gives maximum power output and increase in wind speed increases the power output. The models designed on CAD software are imported on ANSYS-FLUENT to predict mechanical power. This mechanical power is then converted into electrical power, and the results were approximately the same in both cases. In the end, a comparison has been done to compare the results of experiments and ANSYS-FLUENT.

Keywords: computational analysis, power efficiency, wind energy, wind turbine

Procedia PDF Downloads 145
3803 Predicting Loss of Containment in Surface Pipeline using Computational Fluid Dynamics and Supervised Machine Learning Model to Improve Process Safety in Oil and Gas Operations

Authors: Muhammmad Riandhy Anindika Yudhy, Harry Patria, Ramadhani Santoso

Abstract:

Loss of containment is the primary hazard that process safety management is concerned within the oil and gas industry. Escalation to more serious consequences all begins with the loss of containment, starting with oil and gas release from leakage or spillage from primary containment resulting in pool fire, jet fire and even explosion when reacted with various ignition sources in the operations. Therefore, the heart of process safety management is avoiding loss of containment and mitigating its impact through the implementation of safeguards. The most effective safeguard for the case is an early detection system to alert Operations to take action prior to a potential case of loss of containment. The detection system value increases when applied to a long surface pipeline that is naturally difficult to monitor at all times and is exposed to multiple causes of loss of containment, from natural corrosion to illegal tapping. Based on prior researches and studies, detecting loss of containment accurately in the surface pipeline is difficult. The trade-off between cost-effectiveness and high accuracy has been the main issue when selecting the traditional detection method. The current best-performing method, Real-Time Transient Model (RTTM), requires analysis of closely positioned pressure, flow and temperature (PVT) points in the pipeline to be accurate. Having multiple adjacent PVT sensors along the pipeline is expensive, hence generally not a viable alternative from an economic standpoint.A conceptual approach to combine mathematical modeling using computational fluid dynamics and a supervised machine learning model has shown promising results to predict leakage in the pipeline. Mathematical modeling is used to generate simulation data where this data is used to train the leak detection and localization models. Mathematical models and simulation software have also been shown to provide comparable results with experimental data with very high levels of accuracy. While the supervised machine learning model requires a large training dataset for the development of accurate models, mathematical modeling has been shown to be able to generate the required datasets to justify the application of data analytics for the development of model-based leak detection systems for petroleum pipelines. This paper presents a review of key leak detection strategies for oil and gas pipelines, with a specific focus on crude oil applications, and presents the opportunities for the use of data analytics tools and mathematical modeling for the development of robust real-time leak detection and localization system for surface pipelines. A case study is also presented.

Keywords: pipeline, leakage, detection, AI

Procedia PDF Downloads 173