Search results for: universe modelling
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1966

Search results for: universe modelling

1756 Numerical Modelling of Immiscible Fluids Flow in Oil Reservoir Rocks during Enhanced Oil Recovery Processes

Authors: Zahreddine Hafsi, Manoranjan Mishra , Sami Elaoud

Abstract:

Ensuring the maximum recovery rate of oil from reservoir rocks is a challenging task that requires preliminary numerical analysis of different techniques used to enhance the recovery process. After conventional oil recovery processes and in order to retrieve oil left behind after the primary recovery phase, water flooding in one of several techniques used for enhanced oil recovery (EOR). In this research work, EOR via water flooding is numerically modeled, and hydrodynamic instabilities resulted from immiscible oil-water flow in reservoir rocks are investigated. An oil reservoir is a porous medium consisted of many fractures of tiny dimensions. For modeling purposes, the oil reservoir is considered as a collection of capillary tubes which provides useful insights into how fluids behave in the reservoir pore spaces. Equations governing oil-water flow in oil reservoir rocks are developed and numerically solved following a finite element scheme. Numerical results are obtained using Comsol Multiphysics software. The two phase Darcy module of COMSOL Multiphysics allows modelling the imbibition process by the injection of water (as wetting phase) into an oil reservoir. Van Genuchten, Brooks Corey and Levrett models were considered as retention models and obtained flow configurations are compared, and the governing parameters are discussed. For the considered retention models it was found that onset of instabilities viz. fingering phenomenon is highly dependent on the capillary pressure as well as the boundary conditions, i.e., the inlet pressure and the injection velocity.

Keywords: capillary pressure, EOR process, immiscible flow, numerical modelling

Procedia PDF Downloads 129
1755 Modelling Optimal Control of Diabetes in the Workplace

Authors: Eunice Christabel Chukwu

Abstract:

Introduction: Diabetes is a chronic medical condition which is characterized by high levels of glucose in the blood and urine; it is usually diagnosed by means of a glucose tolerance test (GTT). Diabetes can cause a range of health problems if left unmanaged, as it can lead to serious complications. It is essential to manage the condition effectively, particularly in the workplace where the impact on work productivity can be significant. This paper discusses the modelling of optimal control of diabetes in the workplace using a control theory approach. Background: Diabetes mellitus is a condition caused by too much glucose in the blood. Insulin, a hormone produced by the pancreas, controls the blood sugar level by regulating the production and storage of glucose. In diabetes, there may be a decrease in the body’s ability to respond to insulin or a decrease in insulin produced by the pancreas which will lead to abnormalities in the metabolism of carbohydrates, proteins, and fats. In addition to the health implications, the condition can also have a significant impact on work productivity, as employees with uncontrolled diabetes are at risk of absenteeism, reduced performance, and increased healthcare costs. While several interventions are available to manage diabetes, the most effective approach is to control blood glucose levels through a combination of lifestyle modifications and medication. Methodology: The control theory approach involves modelling the dynamics of the system and designing a controller that can regulate the system to achieve optimal performance. In the case of diabetes, the system dynamics can be modelled using a mathematical model that describes the relationship between insulin, glucose, and other variables. The controller can then be designed to regulate the glucose levels to maintain them within a healthy range. Results: The modelling of optimal control of diabetes in the workplace using a control theory approach has shown promising results. The model has been able to predict the optimal dose of insulin required to maintain glucose levels within a healthy range, taking into account the individual’s lifestyle, medication regimen, and other relevant factors. The approach has also been used to design interventions that can improve diabetes management in the workplace, such as regular glucose monitoring and education programs. Conclusion: The modelling of optimal control of diabetes in the workplace using a control theory approach has significant potential to improve diabetes management and work productivity. By using a mathematical model and a controller to regulate glucose levels, the approach can help individuals with diabetes to achieve optimal health outcomes while minimizing the impact of the condition on their work performance. Further research is needed to validate the model and develop interventions that can be implemented in the workplace.

Keywords: mathematical model, blood, insulin, pancreas, model, glucose

Procedia PDF Downloads 59
1754 Identifying Barriers of Implementing Building Information Modelling in Construction

Authors: Kasra HosseinMostofi, Mohamadamin Oyar Hossein, Reza Mehdizadeh Anvigh

Abstract:

BIM is an innovative concept for the majority of firms operating in industry. BIM offers a new paradigm to design, construct, operate, and maintain a facility. However, even with the most conscientious use, stakeholders can run into trouble during its implementation on a project or within an organization. At times, project stakeholders are unaware of the challenges that they can face with the implementation at the project level or an organizational level. Therefore, the study aimed to identify and compile barriers associated with the BIM implementation at the project and organizational level, as per the literature. Despite the fact that innumerable advantageous involved in exploiting BIM, there are some barriers to implement it properly. These barriers have been proved as impediments for academicians and members of construction team project to take the maximum advantage of its utilization. Although some research has been conducted to identify these barriers regarding BIM implementation in construction industry, more research is needed to be carried out among academicians to identify these barriers in institutions, and most importantly, to make suggestions for eliminating these obstacles.

Keywords: building information modelling, construction, design and construction, designers

Procedia PDF Downloads 184
1753 Scaling Strategy of a New Experimental Rig for Wheel-Rail Contact

Authors: Meysam Naeimi, Zili Li, Rolf Dollevoet

Abstract:

A new small–scale test rig developed for rolling contact fatigue (RCF) investigations in wheel–rail material. This paper presents the scaling strategy of the rig based on dimensional analysis and mechanical modelling. The new experimental rig is indeed a spinning frame structure with multiple wheel components over a fixed rail-track ring, capable of simulating continuous wheel-rail contact in a laboratory scale. This paper describes the dimensional design of the rig, to derive its overall scaling strategy and to determine the key elements’ specifications. Finite element (FE) modelling is used to simulate the mechanical behavior of the rig with two sample scale factors of 1/5 and 1/7. The results of FE models are compared with the actual railway system to observe the effectiveness of the chosen scales. The mechanical properties of the components and variables of the system are finally determined through the design process.

Keywords: new test rig, rolling contact fatigue, rail, small scale

Procedia PDF Downloads 481
1752 Numerical Modelling of Dry Stone Masonry Structures Based on Finite-Discrete Element Method

Authors: Ž. Nikolić, H. Smoljanović, N. Živaljić

Abstract:

This paper presents numerical model based on finite-discrete element method for analysis of the structural response of dry stone masonry structures under static and dynamic loads. More precisely, each discrete stone block is discretized by finite elements. Material non-linearity including fracture and fragmentation of discrete elements as well as cyclic behavior during dynamic load are considered through contact elements which are implemented within a finite element mesh. The application of the model was conducted on several examples of these structures. The performed analysis shows high accuracy of the numerical results in comparison with the experimental ones and demonstrates the potential of the finite-discrete element method for modelling of the response of dry stone masonry structures.

Keywords: dry stone masonry structures, dynamic load, finite-discrete element method, static load

Procedia PDF Downloads 412
1751 Modelling of Rate-Dependent Hysteresis of Polypyrrole Dual Sensing-Actuators for Precise Position Control

Authors: Johanna Schumacher, Toribio F. Otero, Victor H. Pascual

Abstract:

Bending dual sensing-actuators based on electroactive polymers are faradaic motors meaning the consumed charge determines the actuator’s tip position. During actuation, consumed charges during oxidation and reduction result in different tip positions showing dynamic hysteresis effects with errors up to 25%. For a precise position control of these actuators, the characterization of the hysteresis effect due to irreversible reactions is crucial. Here, the investigation and modelling of dynamic hysteresis effects of polypyrrole-dodezylbenzenesulfonate (PPyDBS) actuators under ambient working conditions are presented. The hysteresis effect is studied for charge consumption at different frequencies and a rate-dependent hysteresis model is derived. The hysteresis model is implemented as closed loop system and is verified experimentally.

Keywords: dual sensing-actuator, electroactive polymers, hysteresis, position control

Procedia PDF Downloads 383
1750 2D RF ICP Torch Modelling with Fluid Plasma

Authors: Mokhtar Labiod, Nabil Ikhlef, Keltoum Bouherine, Olivier Leroy

Abstract:

A numerical model for the radio-frequency (RF) Argon discharge chamber is developed to simulate the low pressure low temperature inductively coupled plasma. This model will be of fundamental importance in the design of the plasma magnetic control system. Electric and magnetic fields inside the discharge chamber are evaluated by solving a magnetic vector potential equation. To start with, the equations of the ideal magnetohydrodynamics theory will be presented describing the basic behaviour of magnetically confined plasma and equations are discretized with finite element method in cylindrical coordinates. The discharge chamber is assumed to be axially symmetric and the plasma is treated as a compressible gas. Plasma generation due to ionization is added to the continuity equation. Magnetic vector potential equation is solved for the electromagnetic fields. A strong dependence of the plasma properties on the discharge conditions and the gas temperature is obtained.

Keywords: direct-coupled model, magnetohydrodynamic, modelling, plasma torch simulation

Procedia PDF Downloads 431
1749 Application of Terminal Sliding Mode Control to the Stabilization of the Indoor Temperature in Buildings

Authors: Pawel Skruch, Marek Dlugosz

Abstract:

The paper starts with a general model of the temperature dynamics in buildings. The modelling approach relies on thermodynamics, in particular heat transfer, principles. The model considers heat loses by conduction and ventilation and internal heat gains. The parameters of the model can be determined uniquely from the geometry of the building and from thermal properties of construction materials. The model is presented using state space notation and this form is used in the control design procedure. A sliding surface is defined by the system output and the desired trajectory. The control law is designed to force the trajectory of the system from any initial condition to the sliding surface in finite time. The trajectory of the system after reaching the sliding surface remains on it. A simulation example is included to verify the approach and to demonstrate the achievable performance improvement by the proposed solution in the temperature control in buildings.

Keywords: modelling, building, temperature dynamics, sliding-mode control, sliding surface

Procedia PDF Downloads 546
1748 Calculating All Dark Energy and Dark Matter Effects Through Dynamic Gravity Theory

Authors: Sean Kinney

Abstract:

In 1666, Newton created the Law of Universal Gravitation. And in 1915, Einstein improved it to incorporate factors such as time dilation and gravitational lensing. But currently, there is a problem with this “universal” law. The math doesn’t work outside the confines of our solar system. And something is missing; any evidence of what gravity actually is and how it manifest. This paper explores the notion that gravity must obey the law of conservation of energy as all other forces in this universe have been shown to do. Explaining exactly what gravity is and how it manifests itself. And looking at many different implications that would be created are explained. And finally, using the math of Dynamic Gravity to calculate Dark Energy and Dark Matter effects to explain all observations without the need of exotic measures.

Keywords: gravity, dynamic gravity, dark matter, dark energy

Procedia PDF Downloads 104
1747 A Cosmic Time Dilation Model for the Week of Creation

Authors: Kwok W. Cheung

Abstract:

A scientific interpretation of creation reconciling the beliefs of six literal days of creation and a 13.7-billion-year-old universe currently perceived by most modern cosmologists is proposed. We hypothesize that the reference timeframe of God’s creation is associated with some cosmic time different from the earth's time. We show that the scale factor of earth time to cosmic time can be determined by the solution of the Friedmann equations. Based on this scale factor and some basic assumptions, we derive a Cosmic Time Dilation model that harmonizes the literal meaning of creation days and scientific discoveries with remarkable accuracy.

Keywords: cosmological expansion, time dilation, creation, genesis, relativity, Big Bang, biblical hermeneutics

Procedia PDF Downloads 88
1746 Choral Singers' Preference for Expressive Priming Techniques

Authors: Shawn Michael Condon

Abstract:

Current research on teaching expressivity mainly involves instrumentalists. This study focuses on choral singers’ preference of priming techniques based on four methods for teaching expressivity. 112 choral singers answered the survey about their preferred methods for priming expressivity (vocal modelling, using metaphor, tapping into felt emotions, and drawing on past experiences) in three conditions (active, passive, and instructor). Analysis revealed higher preference for drawing on past experience among more experienced singers. The most preferred technique in the passive and instructor roles was vocal modelling, with metaphors and tapping into felt emotions favoured in an active role. Priming techniques are often used in combination with other methods to enhance singing technique or expressivity and are dependent upon the situation, repertoire, and the preferences of the instructor and performer.

Keywords: emotion, expressivity, performance, singing, teaching

Procedia PDF Downloads 154
1745 Modelling and Simulation Efforts in Scale-Up and Characterization of Semi-Solid Dosage Forms

Authors: Saurav S. Rath, Birendra K. David

Abstract:

Generic pharmaceutical industry has to operate in strict timelines of product development and scale-up from lab to plant. Hence, detailed product & process understanding and implementation of appropriate mechanistic modelling and Quality-by-design (QbD) approaches are imperative in the product life cycle. This work provides example cases of such efforts in topical dosage products. Topical products are typically in the form of emulsions, gels, thick suspensions or even simple solutions. The efficacy of such products is determined by characteristics like rheology and morphology. Defining, and scaling up the right manufacturing process with a given set of ingredients, to achieve the right product characteristics presents as a challenge to the process engineer. For example, the non-Newtonian rheology varies not only with CPPs and CMAs but also is an implicit function of globule size (CQA). Hence, this calls for various mechanistic models, to help predict the product behaviour. This paper focusses on such models obtained from computational fluid dynamics (CFD) coupled with population balance modelling (PBM) and constitutive models (like shear, energy density). In a special case of the use of high shear homogenisers (HSHs) for the manufacture of thick emulsions/gels, this work presents some findings on (i) scale-up algorithm for HSH using shear strain, a novel scale-up parameter for estimating mixing parameters, (ii) non-linear relationship between viscosity and shear imparted into the system, (iii) effect of hold time on rheology of product. Specific examples of how this approach enabled scale-up across 1L, 10L, 200L, 500L and 1000L scales will be discussed.

Keywords: computational fluid dynamics, morphology, quality-by-design, rheology

Procedia PDF Downloads 268
1744 Modelling of Moisture Loss and Oil Uptake during Deep-Fat Frying of Plantain

Authors: James A. Adeyanju, John O. Olajide, Akinbode A. Adedeji

Abstract:

A predictive mathematical model based on the fundamental principles of mass transfer was developed to simulate the moisture content and oil content during Deep-Fat Frying (DFF) process of dodo. The resulting governing equation, that is, partial differential equation that describes rate of moisture loss and oil uptake was solved numerically using explicit Finite Difference Technique (FDT). Computer codes were written in MATLAB environment for the implementation of FDT at different frying conditions and moisture loss as well as oil uptake simulation during DFF of dodo. Plantain samples were sliced into 5 mm thickness and fried at different frying oil temperatures (150, 160 and 170 ⁰C) for periods varying from 2 to 4 min. The comparison between the predicted results and experimental data for the validation of the model showed reasonable agreement. The correlation coefficients between the predicted and experimental values of moisture and oil transfer models ranging from 0.912 to 0.947 and 0.895 to 0.957, respectively. The predicted results could be further used for the design, control and optimization of deep-fat frying process.

Keywords: frying, moisture loss, modelling, oil uptake

Procedia PDF Downloads 446
1743 Improving Our Understanding of the in vivo Modelling of Psychotic Disorders

Authors: Zsanett Bahor, Cristina Nunes-Fonseca, Gillian L. Currie, Emily S. Sena, Lindsay D.G. Thomson, Malcolm R. Macleod

Abstract:

Psychosis is ranked as the third most disabling medical condition in the world by the World Health Organization. Despite a substantial amount of research in recent years, available treatments are not universally effective and have a wide range of adverse side effects. Since many clinical drug candidates are identified through in vivo modelling, a deeper understanding of these models, and their strengths and limitations, might help us understand reasons for difficulties in psychosis drug development. To provide an unbiased summary of the preclinical psychosis literature we performed a systematic electronic search of PubMed for publications modelling a psychotic disorder in vivo, identifying 14,721 relevant studies. Double screening of 11,000 publications from this dataset so far established 2403 animal studies of psychosis, with the most common model being schizophrenia (95%). 61% of these models are induced using pharmacological agents. For all the models only 56% of publications test a therapeutic treatment. We propose a systematic review of these studies to assess the prevalence of reporting of measures to reduce risk of bias, and a meta-analysis to assess the internal and external validity of these animal models. Our findings are likely to be relevant to future preclinical studies of psychosis as this generation of strong empirical evidence has the potential to identify weaknesses, areas for improvement and make suggestions on refinement of experimental design. Such a detailed understanding of the data which inform what we think we know will help improve the current attrition rate between bench and bedside in psychosis research.

Keywords: animal models, psychosis, systematic review, schizophrenia

Procedia PDF Downloads 289
1742 A Design of an Augmented Reality Based Virtual Heritage Application

Authors: Stephen Barnes, Ian Mills, Frances Cleary

Abstract:

Augmented and virtual reality-based applications offer many benefits for the heritage and tourism sector. This technology provides a platform to showcase the regions of interest to people without the need for them to be physically present, which has had a positive impact on enticing tourists to visit those locations. However, the technology also provides the opportunity to present historical artefacts in a form that accurately represents their original, intended appearance. Three sites of interest were identified in the Lingaun Valley in South East Ireland, wherein virtual representations of site-specific artefacts of interest were created via a multidisciplinary team encompassing archaeology, art history, 3D modelling, design, and software development. The collated information has been presented to users via an augmented reality mobile-based application that provides information in an engaging manner that encourages an interest in history as well as visits to the sites in the Lingaun Valley.

Keywords: augmented reality, virtual heritage, 3D modelling, archaeology, virtual representation

Procedia PDF Downloads 80
1741 Software Component Identification from Its Object-Oriented Code: Graph Metrics Based Approach

Authors: Manel Brichni, Abdelhak-Djamel Seriai

Abstract:

Systems are increasingly complex. To reduce their complexity, an abstract view of the system can simplify its development. To overcome this problem, we propose a method to decompose systems into subsystems while reducing their coupling. These subsystems represent components. Consisting of an existing object-oriented systems, the main idea of our approach is based on modelling as graphs all entities of an oriented object source code. Such modelling is easy to handle, so we can apply restructuring algorithms based on graph metrics. The particularity of our approach consists in integrating in addition to standard metrics, such as coupling and cohesion, some graph metrics giving more precision during the components identi cation. To treat this problem, we relied on the ROMANTIC approach that proposed a component-based software architecture recovery from an object oriented system.

Keywords: software reengineering, software component and interfaces, metrics, graphs

Procedia PDF Downloads 500
1740 Calculating All Dark Energy and Dark Matter Effects through Dynamic Gravity Theory

Authors: Sean Michael Kinney

Abstract:

In 1666, Newton created the Law of Universal Gravitation. And in 1915, Einstein improved it to incorporate factors such as time dilation and gravitational lensing. But currently, there is a problem with this “universal” law. The math doesn’t work outside the confines of our solar system. And something is missing; any evidence of what gravity actually is and how it manifests. This paper explores the notion that gravity must obey the law of conservation of energy as all other forces in this universe have been shown to do. Explaining exactly what gravity is and how it manifests itself. And looking at many different implications that would be created are explained. And finally, use the math of Dynamic gravity to calculate Dark Energy and Dark Matter effects to explain all observations without the need for exotic measures.

Keywords: dynamic gravity, gravity, dark matter, dark energy

Procedia PDF Downloads 76
1739 Optimal Placement and Sizing of Distributed Generation in Microgrid for Power Loss Reduction and Voltage Profile Improvement

Authors: Ferinar Moaidi, Mahdi Moaidi

Abstract:

Environmental issues and the ever-increasing in demand of electrical energy make it necessary to have distributed generation (DG) resources in the power system. In this research, in order to realize the goals of reducing losses and improving the voltage profile in a microgrid, the allocation and sizing of DGs have been used. The proposed Genetic Algorithm (GA) is described from the array of artificial intelligence methods for solving the problem. The algorithm is implemented on the IEEE 33 buses network. This study is presented in two scenarios, primarily to illustrate the effect of location and determination of DGs has been done to reduce losses and improve the voltage profile. On the other hand, decisions made with the one-level assumptions of load are not universally accepted for all levels of load. Therefore, in this study, load modelling is performed and the results are presented for multi-levels load state.

Keywords: distributed generation, genetic algorithm, microgrid, load modelling, loss reduction, voltage improvement

Procedia PDF Downloads 142
1738 Peril´s Environment of Energetic Infrastructure Complex System, Modelling by the Crisis Situation Algorithms

Authors: Jiří F. Urbánek, Alena Oulehlová, Hana Malachová, Jiří J. Urbánek Jr.

Abstract:

Crisis situations investigation and modelling are introduced and made within the complex system of energetic critical infrastructure, operating on peril´s environments. Every crisis situations and perils has an origin in the emergency/ crisis event occurrence and they need critical/ crisis interfaces assessment. Here, the emergency events can be expected - then crisis scenarios can be pre-prepared by pertinent organizational crisis management authorities towards their coping; or it may be unexpected - without pre-prepared scenario of event. But the both need operational coping by means of crisis management as well. The operation, forms, characteristics, behaviour and utilization of crisis management have various qualities, depending on real critical infrastructure organization perils, and prevention training processes. An aim is always - better security and continuity of the organization, which successful obtainment needs to find and investigate critical/ crisis zones and functions in critical infrastructure organization models, operating in pertinent perils environment. Our DYVELOP (Dynamic Vector Logistics of Processes) method is disposables for it. Here, it is necessary to derive and create identification algorithm of critical/ crisis interfaces. The locations of critical/ crisis interfaces are the flags of crisis situation in organization of critical infrastructure models. Then, the model of crisis situation will be displayed at real organization of Czech energetic crisis infrastructure subject in real peril environment. These efficient measures are necessary for the infrastructure protection. They will be derived for peril mitigation, crisis situation coping and for environmentally friendly organization survival, continuity and its sustainable development advanced possibilities.

Keywords: algorithms, energetic infrastructure complex system, modelling, peril´s environment

Procedia PDF Downloads 402
1737 Storms Dynamics in the Black Sea in the Context of the Climate Changes

Authors: Eugen Rusu

Abstract:

The objective of the work proposed is to perform an analysis of the wave conditions in the Black Sea basin. This is especially focused on the spatial and temporal occurrences and on the dynamics of the most extreme storms in the context of the climate changes. A numerical modelling system, based on the spectral phase averaged wave model SWAN, has been implemented and validated against both in situ measurements and remotely sensed data, all along the sea. Moreover, a successive correction method for the assimilation of the satellite data has been associated with the wave modelling system. This is based on the optimal interpolation of the satellite data. Previous studies show that the process of data assimilation improves considerably the reliability of the results provided by the modelling system. This especially concerns the most sensitive cases from the point of view of the accuracy of the wave predictions, as the extreme storm situations are. Following this numerical approach, it has to be highlighted that the results provided by the wave modelling system above described are in general in line with those provided by some similar wave prediction systems implemented in enclosed or semi-enclosed sea basins. Simulations of this wave modelling system with data assimilation have been performed for the 30-year period 1987-2016. Considering this database, the next step was to analyze the intensity and the dynamics of the higher storms encountered in this period. According to the data resulted from the model simulations, the western side of the sea is considerably more energetic than the rest of the basin. In this western region, regular strong storms provide usually significant wave heights greater than 8m. This may lead to maximum wave heights even greater than 15m. Such regular strong storms may occur several times in one year, usually in the wintertime, or in late autumn, and it can be noticed that their frequency becomes higher in the last decade. As regards the case of the most extreme storms, significant wave heights greater than 10m and maximum wave heights close to 20m (and even greater) may occur. Such extreme storms, which in the past were noticed only once in four or five years, are more recent to be faced almost every year in the Black Sea, and this seems to be a consequence of the climate changes. The analysis performed included also the dynamics of the monthly and annual significant wave height maxima as well as the identification of the most probable spatial and temporal occurrences of the extreme storm events. Finally, it can be concluded that the present work provides valuable information related to the characteristics of the storm conditions and on their dynamics in the Black Sea. This environment is currently subjected to high navigation traffic and intense offshore and nearshore activities and the strong storms that systematically occur may produce accidents with very serious consequences.

Keywords: Black Sea, extreme storms, SWAN simulations, waves

Procedia PDF Downloads 246
1736 LTE Modelling of a DC Arc Ignition on Cold Electrodes

Authors: O. Ojeda Mena, Y. Cressault, P. Teulet, J. P. Gonnet, D. F. N. Santos, MD. Cunha, M. S. Benilov

Abstract:

The assumption of plasma in local thermal equilibrium (LTE) is commonly used to perform electric arc simulations for industrial applications. This assumption allows to model the arc using a set of magneto-hydromagnetic equations that can be solved with a computational fluid dynamic code. However, the LTE description is only valid in the arc column, whereas in the regions close to the electrodes the plasma deviates from the LTE state. The importance of these near-electrode regions is non-trivial since they define the energy and current transfer between the arc and the electrodes. Therefore, any accurate modelling of the arc must include a good description of the arc-electrode phenomena. Due to the modelling complexity and computational cost of solving the near-electrode layers, a simplified description of the arc-electrode interaction was developed in a previous work to study a steady high-pressure arc discharge, where the near-electrode regions are introduced at the interface between arc and electrode as boundary conditions. The present work proposes a similar approach to simulate the arc ignition in a free-burning arc configuration following an LTE description of the plasma. To obtain the transient evolution of the arc characteristics, appropriate boundary conditions for both the near-cathode and the near-anode regions are used based on recent publications. The arc-cathode interaction is modeled using a non-linear surface heating approach considering the secondary electron emission. On the other hand, the interaction between the arc and the anode is taken into account by means of the heating voltage approach. From the numerical modelling, three main stages can be identified during the arc ignition. Initially, a glow discharge is observed, where the cold non-thermionic cathode is uniformly heated at its surface and the near-cathode voltage drop is in the order of a few hundred volts. Next, a spot with high temperature is formed at the cathode tip followed by a sudden decrease of the near-cathode voltage drop, marking the glow-to-arc discharge transition. During this stage, the LTE plasma also presents an important increase of the temperature in the region adjacent to the hot spot. Finally, the near-cathode voltage drop stabilizes at a few volts and both the electrode and plasma temperatures reach the steady solution. The results after some seconds are similar to those presented for thermionic cathodes.

Keywords: arc-electrode interaction, thermal plasmas, electric arc simulation, cold electrodes

Procedia PDF Downloads 121
1735 Building Information Modelling Based Value for Money Assessment in Public-Private Partnership

Authors: Guoqian Ren, Haijiang Li, Jisong Zhang

Abstract:

Over the past 40 years, urban development has undergone large-scale, high-speed expansion, beyond what was previously considered normal and in a manner not proportionally related to population growth or physical considerations. With more scientific and refined decision-making in the urban construction process, new urbanization approaches, aligned with public-private partnerships (PPPs) which evolved in the early 1990s, have become acceptable and, in some situations, even better solutions to outstanding urban municipal construction projects, especially in developing countries. However, as the main driving force to deal with urban public services, PPPs are still problematic regarding value for money (VFM) process in most large-scale construction projects. This paper therefore reviews recent PPP articles in popular project management journals and relevant toolkits, published in the last 10 years, to identify the indicators that influence VFM within PPPs across regions. With increasing concerns about profitability and environmental and social impacts, the current PPP structure requires a more integrated platform to manage multi-performance project life cycles. Building information modelling (BIM), a popular approach to the procurement process in AEC sectors, provides the potential to ensure VFM while also working in tandem with the semantic approach to holistically measure life cycle costs (LCC) and achieve better sustainability. This paper suggests that BIM applied to the entire PPP life cycle could support holistic decision-making regarding VFM processes and thus meet service targets.

Keywords: public-private partnership, value for money, building information modelling, semantic approach

Procedia PDF Downloads 208
1734 Modelling of Atomic Force Microscopic Nano Robot's Friction Force on Rough Surfaces

Authors: M. Kharazmi, M. Zakeri, M. Packirisamy, J. Faraji

Abstract:

Micro/Nanorobotics or manipulation of nanoparticles by Atomic Force Microscopic (AFM) is one of the most important solutions for controlling the movement of atoms, particles and micro/nano metrics components and assembling of them to design micro/nano-meter tools. Accurate modelling of manipulation requires identification of forces and mechanical knowledge in the Nanoscale which are different from macro world. Due to the importance of the adhesion forces and the interaction of surfaces at the nanoscale several friction models were presented. In this research, friction and normal forces that are applied on the AFM by using of the dynamic bending-torsion model of AFM are obtained based on Hurtado-Kim friction model (HK), Johnson-Kendall-Robert contact model (JKR) and Greenwood-Williamson roughness model (GW). Finally, the effect of standard deviation of asperities height on the normal load, friction force and friction coefficient are studied.

Keywords: atomic force microscopy, contact model, friction coefficient, Greenwood-Williamson model

Procedia PDF Downloads 198
1733 Study on the Legend of Dayi in China

Authors: Zhiguo Ju

Abstract:

Both ancient written documents and archaeological studies showed that the ancient Chinese was people who worshiped the Sun God. There is a legend of Dayi in China, however, told a story of Dayi, an ancient hero who sacrificed his life to shoot nine suns out of the ten in the sky, restored the order of the universe, and saved the people’s life. By investigating its oral inheritance and folk customs, we found that the story was originated from Rizhao, the east coralline in east China. This could provide valuable cultural resources and could be used in local tourism. How the Sun-worshipping-people developed such a contradictory legend, how was it be carried by generations, and how to use it as a cultural source to promote the local tourism was discussed.

Keywords: Legend of Dayi, sun worship, ancient Chinese, China

Procedia PDF Downloads 689
1732 Development of an Automatic Calibration Framework for Hydrologic Modelling Using Approximate Bayesian Computation

Authors: A. Chowdhury, P. Egodawatta, J. M. McGree, A. Goonetilleke

Abstract:

Hydrologic models are increasingly used as tools to predict stormwater quantity and quality from urban catchments. However, due to a range of practical issues, most models produce gross errors in simulating complex hydraulic and hydrologic systems. Difficulty in finding a robust approach for model calibration is one of the main issues. Though automatic calibration techniques are available, they are rarely used in common commercial hydraulic and hydrologic modelling software e.g. MIKE URBAN. This is partly due to the need for a large number of parameters and large datasets in the calibration process. To overcome this practical issue, a framework for automatic calibration of a hydrologic model was developed in R platform and presented in this paper. The model was developed based on the time-area conceptualization. Four calibration parameters, including initial loss, reduction factor, time of concentration and time-lag were considered as the primary set of parameters. Using these parameters, automatic calibration was performed using Approximate Bayesian Computation (ABC). ABC is a simulation-based technique for performing Bayesian inference when the likelihood is intractable or computationally expensive to compute. To test the performance and usefulness, the technique was used to simulate three small catchments in Gold Coast. For comparison, simulation outcomes from the same three catchments using commercial modelling software, MIKE URBAN were used. The graphical comparison shows strong agreement of MIKE URBAN result within the upper and lower 95% credible intervals of posterior predictions as obtained via ABC. Statistical validation for posterior predictions of runoff result using coefficient of determination (CD), root mean square error (RMSE) and maximum error (ME) was found reasonable for three study catchments. The main benefit of using ABC over MIKE URBAN is that ABC provides a posterior distribution for runoff flow prediction, and therefore associated uncertainty in predictions can be obtained. In contrast, MIKE URBAN just provides a point estimate. Based on the results of the analysis, it appears as though ABC the developed framework performs well for automatic calibration.

Keywords: automatic calibration framework, approximate bayesian computation, hydrologic and hydraulic modelling, MIKE URBAN software, R platform

Procedia PDF Downloads 303
1731 Crude Oil Electrostatic Mathematical Modelling on an Existing Industrial Plant

Authors: Fatemeh Yazdanmehr, Iulian Nistor

Abstract:

The scope of the current study is the prediction of water separation in a two-stage industrial crude oil desalting plant. This research study was focused on developing a desalting operation in an existing production unit of one Iranian heavy oil field with 75 MBPD capacity. Because of some operational issues, such as oil dehydration at high temperatures, the optimization of the desalter operational parameters was essential. The mathematical desalting is modeled based on the population balance method. The existing operational data is used for tuning and validation of the accuracy of the modeling. The inlet oil temperature to desalter used was decreased from 110°C to 80°C, and the desalted electrical field was increased from 0.75 kv to 2.5 kv. The proposed condition for the desalter also meets the water oil specification. Based on these conditions of desalter, the oil recovery is increased by 574 BBL/D, and the gas flaring decrease by 2.8 MMSCF/D. Depending on the oil price, the additional production of oil can increase the annual income by about $15 MM and reduces greenhouse gas production caused by gas flaring.

Keywords: desalter, demulsification, modelling, water-oil separation, crude oil emulsion

Procedia PDF Downloads 74
1730 Customer Data Analysis Model Using Business Intelligence Tools in Telecommunication Companies

Authors: Monica Lia

Abstract:

This article presents a customer data analysis model using business intelligence tools for data modelling, transforming, data visualization and dynamic reports building. Economic organizational customer’s analysis is made based on the information from the transactional systems of the organization. The paper presents how to develop the data model starting for the data that companies have inside their own operational systems. The owned data can be transformed into useful information about customers using business intelligence tool. For a mature market, knowing the information inside the data and making forecast for strategic decision become more important. Business Intelligence tools are used in business organization as support for decision-making.

Keywords: customer analysis, business intelligence, data warehouse, data mining, decisions, self-service reports, interactive visual analysis, and dynamic dashboards, use cases diagram, process modelling, logical data model, data mart, ETL, star schema, OLAP, data universes

Procedia PDF Downloads 427
1729 Text Analysis to Support Structuring and Modelling a Public Policy Problem-Outline of an Algorithm to Extract Inferences from Textual Data

Authors: Claudia Ehrentraut, Osama Ibrahim, Hercules Dalianis

Abstract:

Policy making situations are real-world problems that exhibit complexity in that they are composed of many interrelated problems and issues. To be effective, policies must holistically address the complexity of the situation rather than propose solutions to single problems. Formulating and understanding the situation and its complex dynamics, therefore, is a key to finding holistic solutions. Analysis of text based information on the policy problem, using Natural Language Processing (NLP) and Text analysis techniques, can support modelling of public policy problem situations in a more objective way based on domain experts knowledge and scientific evidence. The objective behind this study is to support modelling of public policy problem situations, using text analysis of verbal descriptions of the problem. We propose a formal methodology for analysis of qualitative data from multiple information sources on a policy problem to construct a causal diagram of the problem. The analysis process aims at identifying key variables, linking them by cause-effect relationships and mapping that structure into a graphical representation that is adequate for designing action alternatives, i.e., policy options. This study describes the outline of an algorithm used to automate the initial step of a larger methodological approach, which is so far done manually. In this initial step, inferences about key variables and their interrelationships are extracted from textual data to support a better problem structuring. A small prototype for this step is also presented.

Keywords: public policy, problem structuring, qualitative analysis, natural language processing, algorithm, inference extraction

Procedia PDF Downloads 588
1728 Modelling of Pipe Jacked Twin Tunnels in a Very Soft Clay

Authors: Hojjat Mohammadi, Randall Divito, Gary J. E. Kramer

Abstract:

Tunnelling and pipe jacking in very soft soils (fat clays), even with an Earth Pressure Balance tunnel boring machine (EPBM), can cause large ground displacements. In this study, the short-term and long-term ground and tunnel response is predicted for twin, pipe-jacked EPBM 3 meter diameter tunnels with a narrow pillar width. Initial modelling indicated complete closure of the annulus gap at the tail shield onto the centrifugally cast, glass-fiber-reinforced, polymer mortar jacking pipe (FRP). Numerical modelling was employed to simulate the excavation and support installation sequence, examine the ground response during excavation, confirm the adequacy of the pillar width and check the structural adequacy of the installed pipe. In the numerical models, Mohr-Coulomb constitutive model with the effect of unloading was adopted for the fat clays, while for the bedrock layer, the generalized Hoek-Brown was employed. The numerical models considered explicit excavation sequences and different levels of ground convergence prior to support installation. The well-studied excavation sequences made the analysis possible for this study on a very soft clay, otherwise, obtaining the convergency in the numerical analysis would be impossible. The predicted results indicate that the ground displacements around the tunnel and its effect on the pipe would be acceptable despite predictions of large zones of plastic behaviour around the tunnels and within the entire pillar between them due to excavation-induced ground movements.

Keywords: finite element modeling (FEM), pipe-jacked tunneling, very soft clay, EPBM

Procedia PDF Downloads 80
1727 Design and Optimization of a Customized External Fixation Device for Lower Limb Injuries

Authors: Mohammed S. Alqahtani, Paulo J. Bartolo

Abstract:

External fixation is a common technique for the treatment and stabilization of bone fractures. Different designs have been proposed by companies and research groups, but all of them present limitations such as high weight, not comfortable to use, and not customized to individual patients. This paper proposes a lightweight customized external fixator, overcoming some of these limitations. External fixators are designed using a set of techniques such as medical imaging, CAD modelling, finite element analysis, and full factorial design of experiments. Key design parameters are discussed, and the optimal set of parameters is used to design the final external fixator. Numerical simulations are used to validate design concepts. Results present an optimal external fixation design with weight reduction of 13% without compromising its stiffness and structural integrity. External fixators are also designed to be additively manufactured, allowing to develop a strategy for personalization.

Keywords: computer-aided design modelling, external fixation, finite element analysis, full factorial, personalization

Procedia PDF Downloads 157