Search results for: generalized linear model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18914

Search results for: generalized linear model

17354 Geomechanical Numerical Modeling of Well Wall in Drilling with Finite Difference Method

Authors: Marzieh Zarei

Abstract:

Well instability is one of the most fundamental challenges faced by the oil and gas industry. Well wall stability analysis is a gap to be filled in the oil industry. The collection of static data such as well logging leads to the construction of a geomechanical numerical model, which will help in assessing the probable risks in future drilling. In this paper, geomechanical model was designed, and mechanical properties of the rock was determined at all points of the model. It was found the safe mud window was determined and the minimum and maximum mud pressures were determined in the ranges of 70-60 MPa and 110-100 MPa, respectively.

Keywords: geomechanics, numerical model, well stability, in-situ stress, underbalanced drilling

Procedia PDF Downloads 116
17353 Simulation of Behaviour Dynamics and Optimization of the Energy System

Authors: Iva Dvornik, Sandro Božić, Žana Božić Brkić

Abstract:

System-dynamic simulating modelling is one of the most appropriate and successful scientific methods of the complex, non-linear, natural, technical and organizational systems. In the recent practice its methodology proved to be efficient in solving the problems of control, behavior, sensitivity and flexibility of the system dynamics behavior having a high degree of complexity, all these by computing simulation i.e. “under laboratory conditions” what means without any danger for observed realities. This essay deals with the research of the gas turbine dynamic process as well as the operating pump units and transformation of gas energy into hydraulic energy has been simulated. In addition, system mathematical model has been also researched (gas turbine- centrifugal pumps – pipeline pressure system – storage vessel).

Keywords: system dynamics, modelling, centrifugal pump, turbine, gases, continuous and discrete simulation, heuristic optimisation

Procedia PDF Downloads 99
17352 Statistical Analysis of Extreme Flow (Regions of Chlef)

Authors: Bouthiba Amina

Abstract:

The estimation of the statistics bound to the precipitation represents a vast domain, which puts numerous challenges to meteorologists and hydrologists. Sometimes, it is necessary, to approach in value the extreme events for sites where there is little, or no datum, as well as their periods of return. The search for a model of the frequency of the heights of daily rains dresses a big importance in operational hydrology: It establishes a basis for predicting the frequency and intensity of floods by estimating the amount of precipitation in past years. The most known and the most common approach is the statistical approach, It consists in looking for a law of probability that fits best the values observed by the random variable " daily maximal rain " after a comparison of various laws of probability and methods of estimation by means of tests of adequacy. Therefore, a frequent analysis of the annual series of daily maximal rains was realized on the data of 54 pluviometric stations of the pond of high and average. This choice was concerned with five laws usually applied to the study and the analysis of frequent maximal daily rains. The chosen period is from 1970 to 2013. It was of use to the forecast of quantiles. The used laws are the law generalized by extremes to three components, those of the extreme values to two components (Gumbel and log-normal) in two parameters, the law Pearson typifies III and Log-Pearson III in three parameters. In Algeria, Gumbel's law has been used for a long time to estimate the quantiles of maximum flows. However, and we will check and choose the most reliable law.

Keywords: return period, extreme flow, statistics laws, Gumbel, estimation

Procedia PDF Downloads 72
17351 Study of the Protection of Induction Motors

Authors: Bencheikh Abdellah

Abstract:

In this paper, we present a mathematical model dedicated to the simulation breaks bars in a three-phase cage induction motor. This model is based on a mesh circuit representing the rotor cage. The tested simulation allowed us to demonstrate the effectiveness of this model to describe the behavior of the machine in a healthy state, failure.

Keywords: AC motors, squirrel cage, diagnostics, MATLAB, SIMULINK

Procedia PDF Downloads 431
17350 Dynamic Model of Heterogeneous Markets with Imperfect Information for the Optimization of Company's Long-Time Strategy

Authors: Oleg Oborin

Abstract:

This paper is dedicated to the development of the model, which can be used to evaluate the effectiveness of long-term corporate strategies and identify the best strategies. The theoretical model of the relatively homogenous product market (such as iron and steel industry, mobile services or road transport) has been developed. In the model, the market consists of a large number of companies with different internal characteristics and objectives. The companies can perform mergers and acquisitions in order to increase their market share. The model allows the simulation of long-time dynamics of the market (for a period longer than 20 years). Therefore, a large number of simulations on random input data was conducted in the framework of the model. After that, the results of the model were compared with the dynamics of real markets, such as the US steel industry from the beginning of the XX century to the present day, and the market of mobile services in Germany for the period between 1990 and 2015.

Keywords: Economic Modelling, Long-Time Strategy, Mergers and Acquisitions, Simulation

Procedia PDF Downloads 358
17349 Modeling and Energy Analysis of Limestone Decomposition with Microwave Heating

Authors: Sofia N. Gonçalves, Duarte M. S. Albuquerque, José C. F. Pereira

Abstract:

The energy transition is spurred by structural changes in energy demand, supply, and prices. Microwave technology was first proposed as a faster alternative for cooking food. It was found that food heated instantly when interacting with high-frequency electromagnetic waves. The dielectric properties account for a material’s ability to absorb electromagnetic energy and dissipate this energy in the form of heat. Many energy-intense industries could benefit from electromagnetic heating since many of the raw materials are dielectric at high temperatures. Limestone sedimentary rock is a dielectric material intensively used in the cement industry to produce unslaked lime. A numerical 3D model was implemented in COMSOL Multiphysics to study the limestone continuous processing under microwave heating. The model solves the two-way coupling between the Energy equation and Maxwell’s equations as well as the coupling between heat transfer and chemical interfaces. Complementary, a controller was implemented to optimize the overall heating efficiency and control the numerical model stability. This was done by continuously matching the cavity impedance and predicting the required energy for the system, avoiding energy inefficiencies. This controller was developed in MATLAB and successfully fulfilled all these goals. The limestone load influence on thermal decomposition and overall process efficiency was the main object of this study. The procedure considered the Verification and Validation of the chemical kinetics model separately from the coupled model. The chemical model was found to correctly describe the chosen kinetic equation, and the coupled model successfully solved the equations describing the numerical model. The interaction between flow of material and electric field Poynting vector revealed to influence limestone decomposition, as a result from the low dielectric properties of limestone. The numerical model considered this effect and took advantage from this interaction. The model was demonstrated to be highly unstable when solving non-linear temperature distributions. Limestone has a dielectric loss response that increases with temperature and has low thermal conductivity. For this reason, limestone is prone to produce thermal runaway under electromagnetic heating, as well as numerical model instabilities. Five different scenarios were tested by considering a material fill ratio of 30%, 50%, 65%, 80%, and 100%. Simulating the tube rotation for mixing enhancement was proven to be beneficial and crucial for all loads considered. When uniform temperature distribution is accomplished, the electromagnetic field and material interaction is facilitated. The results pointed out the inefficient development of the electric field within the bed for 30% fill ratio. The thermal efficiency showed the propensity to stabilize around 90%for loads higher than 50%. The process accomplished a maximum microwave efficiency of 75% for the 80% fill ratio, sustaining that the tube has an optimal fill of material. Electric field peak detachment was observed for the case with 100% fill ratio, justifying the lower efficiencies compared to 80%. Microwave technology has been demonstrated to be an important ally for the decarbonization of the cement industry.

Keywords: CFD numerical simulations, efficiency optimization, electromagnetic heating, impedance matching, limestone continuous processing

Procedia PDF Downloads 167
17348 Channel Characteristics and Morphometry of a Part of Umtrew River, Meghalaya

Authors: Pratyashi Phukan, Ranjan Saikia

Abstract:

Morphometry incorporates quantitative study of the area ,altitude,volume, slope profiles of a land and drainage basin characteristics of the area concerned.Fluvial geomorphology includes the consideration of linear,areal and relief aspects of a fluvially originated drainage basin. The linear aspect deals with the hierarchical orders of streams, numbers, and lenghts of stream segments and various relationship among them.The areal aspect includes the analysis of basin perimeters,basin shape, basin area, and related morphometric laws. The relief aspect incorporates besides hypsometric, climographic and altimetric analysis,the study of absolute and relative reliefs, relief ratios, average slope, etc. In this paper we have analysed the relationship among stream velocity, channel shape,sediment load,channel width,channel depth, etc.

Keywords: morphometry, hydraulic geometry, Umtrew river, Meghalaya

Procedia PDF Downloads 450
17347 Life Prediction Method of Lithium-Ion Battery Based on Grey Support Vector Machines

Authors: Xiaogang Li, Jieqiong Miao

Abstract:

As for the problem of the grey forecasting model prediction accuracy is low, an improved grey prediction model is put forward. Firstly, use trigonometric function transform the original data sequence in order to improve the smoothness of data , this model called SGM( smoothness of grey prediction model), then combine the improved grey model with support vector machine , and put forward the grey support vector machine model (SGM - SVM).Before the establishment of the model, we use trigonometric functions and accumulation generation operation preprocessing data in order to enhance the smoothness of the data and weaken the randomness of the data, then use support vector machine (SVM) to establish a prediction model for pre-processed data and select model parameters using genetic algorithms to obtain the optimum value of the global search. Finally, restore data through the "regressive generate" operation to get forecasting data. In order to prove that the SGM-SVM model is superior to other models, we select the battery life data from calce. The presented model is used to predict life of battery and the predicted result was compared with that of grey model and support vector machines.For a more intuitive comparison of the three models, this paper presents root mean square error of this three different models .The results show that the effect of grey support vector machine (SGM-SVM) to predict life is optimal, and the root mean square error is only 3.18%. Keywords: grey forecasting model, trigonometric function, support vector machine, genetic algorithms, root mean square error

Keywords: Grey prediction model, trigonometric functions, support vector machines, genetic algorithms, root mean square error

Procedia PDF Downloads 454
17346 Two Dimensional Steady State Modeling of Temperature Profile and Heat Transfer of Electrohydrodynamically Enhanced Micro Heat Pipe

Authors: H. Shokouhmand, M. Tajerian

Abstract:

A numerical investigation of laminar forced convection flows through a square cross section micro heat pipe by applying electrohydrodynamic (EHD) field has been studied. In the present study, pentane is selected as working fluid. Temperature and velocity profiles and heat transfer enhancement in the micro heat pipe by using EHD field at the two-dimensional and single phase fluid flow in steady state regime have been numerically calculated. At this model, only Coulomb force is considered. The study has been carried out for the Reynolds number 10 to 100 and EHD force field up to 8 KV. Coupled, non-linear equations governed on the model (continuity, momentum, and energy equations) have been solved simultaneously by CFD numerical methods. Steady state behavior of affecting parameters, e.g. friction factor, average temperature, Nusselt number and heat transfer enhancement criteria, have been evaluated. It has been observed that by increasing Reynolds number, the effect of EHD force became more significant and for smaller Reynolds numbers the rate of heat transfer enhancement criteria is increased. By obtaining and plotting the mentioned parameters, it has been shown that the EHD field enhances the heat transfer process. The numerical results show that by increasing EHD force field the absolute value of Nusselt number and friction factor increases and average temperature of fluid flow decreases. But the increasing rate of Nusselt number is greater than increasing value of friction factor, which makes applying EHD force field for heat transfer enhancement in micro heat pipes acceptable and applicable. The numerical results of model are in good agreement with the experimental results available in the literature.

Keywords: micro heat pipe, electrohydrodynamic force, Nusselt number, average temperature, friction factor

Procedia PDF Downloads 260
17345 Optimal Beam for Accelerator Driven Systems

Authors: M. Paraipan, V. M. Javadova, S. I. Tyutyunnikov

Abstract:

The concept of energy amplifier or accelerator driven system (ADS) involves the use of a particle accelerator coupled with a nuclear reactor. The accelerated particle beam generates a supplementary source of neutrons, which allows the subcritical functioning of the reactor, and consequently a safe exploitation. The harder neutron spectrum realized ensures a better incineration of the actinides. The almost generalized opinion is that the optimal beam for ADS is represented by protons with energy around 1 GeV (gigaelectronvolt). In the present work, a systematic analysis of the energy gain for proton beams with energy from 0.5 to 3 GeV and ion beams from deuteron to neon with energies between 0.25 and 2 AGeV is performed. The target is an assembly of metallic U-Pu-Zr fuel rods in a bath of lead-bismuth eutectic coolant. The rods length is 150 cm. A beryllium converter with length 110 cm is used in order to maximize the energy released in the target. The case of a linear accelerator is considered, with a beam intensity of 1.25‧10¹⁶ p/s, and a total accelerator efficiency of 0.18 for proton beam. These values are planned to be achieved in the European Spallation Source project. The energy gain G is calculated as the ratio between the energy released in the target to the energy spent to accelerate the beam. The energy released is obtained through simulation with the code Geant4. The energy spent is calculating by scaling from the data about the accelerator efficiency for the reference particle (proton). The analysis concerns the G values, the net power produce, the accelerator length, and the period between refueling. The optimal energy for proton is 1.5 GeV. At this energy, G reaches a plateau around a value of 8 and a net power production of 120 MW (megawatt). Starting with alpha, ion beams have a higher G than 1.5 GeV protons. A beam of 0.25 AGeV(gigaelectronvolt per nucleon) ⁷Li realizes the same net power production as 1.5 GeV protons, has a G of 15, and needs an accelerator length 2.6 times lower than for protons, representing the best solution for ADS. Beams of ¹⁶O or ²⁰Ne with energy 0.75 AGeV, accelerated in an accelerator with the same length as 1.5 GeV protons produce approximately 900 MW net power, with a gain of 23-25. The study of the evolution of the isotopes composition during irradiation shows that the increase in power production diminishes the period between refueling. For a net power produced of 120 MW, the target can be irradiated approximately 5000 days without refueling, but only 600 days when the net power reaches 1 GW (gigawatt).

Keywords: accelerator driven system, ion beam, electrical power, energy gain

Procedia PDF Downloads 131
17344 Modelling Export Dynamics in the CSEE Countries Using GVAR Model

Authors: S. Jakšić, B. Žmuk

Abstract:

The paper investigates the key factors of export dynamics for a set of Central and Southeast European (CSEE) countries in the context of current economic and financial crisis. In order to model the export dynamics a Global Vector Auto Regressive (GVAR) model is defined. As opposed to models which model each country separately, the GVAR combines all country models in a global model which enables obtaining important information on spill-over effects in the context of globalization and rising international linkages. The results of the study indicate that for most of the CSEE countries, exports are mainly driven by domestic shocks, both in the short run and in the long run. This study is the first application of the GVAR model to studying the export dynamics in the CSEE countries and therefore the results of the study present an important empirical contribution.

Keywords: export, GFEVD, global VAR, international trade, weak exogeneity

Procedia PDF Downloads 295
17343 Prediction of B-Cell Epitope for 24 Mite Allergens: An in Silico Approach towards Epitope-Based Immune Therapeutics

Authors: Narjes Ebrahimi, Soheila Alyasin, Navid Nezafat, Hossein Esmailzadeh, Younes Ghasemi, Seyed Hesamodin Nabavizadeh

Abstract:

Immunotherapy with allergy vaccines is of great importance in allergen-specific immunotherapy. In recent years, B-cell epitope-based vaccines have attracted considerable attention and the prediction of epitopes is crucial to design these types of allergy vaccines. B-cell epitopes might be linear or conformational. The prerequisite for the identification of conformational epitopes is the information about allergens' tertiary structures. Bioinformatics approaches have paved the way towards the design of epitope-based allergy vaccines through the prediction of tertiary structures and epitopes. Mite allergens are one of the major allergy contributors. Several mite allergens can elicit allergic reactions; however, their structures and epitopes are not well established. So, B-cell epitopes of various groups of mite allergens (24 allergens in 6 allergen groups) were predicted in the present work. Tertiary structures of 17 allergens with unknown structure were predicted and refined with RaptorX and GalaxyRefine servers, respectively. The predicted structures were further evaluated by Rampage, ProSA-web, ERRAT and Verify 3D servers. Linear and conformational B-cell epitopes were identified with Ellipro, Bcepred, and DiscoTope 2 servers. To improve the accuracy level, consensus epitopes were selected. Fifty-four conformational and 133 linear consensus epitopes were predicted. Furthermore, overlapping epitopes in each allergen group were defined, following the sequence alignment of the allergens in each group. The predicted epitopes were also compared with the experimentally identified epitopes. The presented results provide valuable information for further studies about allergy vaccine design.

Keywords: B-cell epitope, Immunotherapy, In silico prediction, Mite allergens, Tertiary structure

Procedia PDF Downloads 153
17342 Simplified 3R2C Building Thermal Network Model: A Case Study

Authors: S. M. Mahbobur Rahman

Abstract:

Whole building energy simulation models are widely used for predicting future energy consumption, performance diagnosis and optimum control.  Black box building energy modeling approach has been heavily studied in the past decade. The thermal response of a building can also be modeled using a network of interconnected resistors (R) and capacitors (C) at each node called R-C network. In this study, a model building, Case 600, as described in the “Standard Method of Test for the Evaluation of Building Energy Analysis Computer Program”, ASHRAE standard 140, is studied along with a 3R2C thermal network model and the ASHRAE clear sky solar radiation model. Although building an energy model involves two important parts of building component i.e., the envelope and internal mass, the effect of building internal mass is not considered in this study. All the characteristic parameters of the building envelope are evaluated as on Case 600. Finally, monthly building energy consumption from the thermal network model is compared with a simple-box energy model within reasonable accuracy. From the results, 0.6-9.4% variation of monthly energy consumption is observed because of the south-facing windows.

Keywords: ASHRAE case study, clear sky solar radiation model, energy modeling, thermal network model

Procedia PDF Downloads 135
17341 Asymptotic Expansion of the Korteweg-de Vries-Burgers Equation

Authors: Jian-Jun Shu

Abstract:

It is common knowledge that many physical problems (such as non-linear shallow-water waves and wave motion in plasmas) can be described by the Korteweg-de Vries (KdV) equation, which possesses certain special solutions, known as solitary waves or solitons. As a marriage of the KdV equation and the classical Burgers (KdVB) equation, the Korteweg-de Vries-Burgers (KdVB) equation is a mathematical model of waves on shallow water surfaces in the presence of viscous dissipation. Asymptotic analysis is a method of describing limiting behavior and is a key tool for exploring the differential equations which arise in the mathematical modeling of real-world phenomena. By using variable transformations, the asymptotic expansion of the KdVB equation is presented in this paper. The asymptotic expansion may provide a good gauge on the validation of the corresponding numerical scheme.

Keywords: asymptotic expansion, differential equation, Korteweg-de Vries-Burgers (KdVB) equation, soliton

Procedia PDF Downloads 243
17340 Maturity Model for Agro-Industrial Logistics

Authors: Erika Tatiana Ruiz, Wilson Adarme Jaimes

Abstract:

This abstract presents the methodology for improving the logistics processes of agricultural production units belonging to the coffee, cocoa, and fruit sectors, starting from the fundamental concepts and detailing each of the phases to carry out the diagnosis, which will be the basis for the formulation of its action plan and implementation of the maturity model. As a result of this work, the maturity model is formulated to improve logistics processes. This model seeks to: generate a progressive model that is useful for all productive units belonging to these sectors at the national level, regardless of their initial conditions, focus on the improvement of logistics processes as a strategy that contributes to improving the competitiveness of the agricultural sector in Colombia and spread the implementation of good logistics practices in postharvest in all departments of the country through autonomous tools. This model has been built through a series of steps that allow the evaluation and improvement of the logistics dimensions or indicators. The potential improvements for each dimension provide the foundation on which to advance to the next level. Within the maturity model, a methodology is indicated for the design and execution of strategies to improve its logistics processes, taking into account the current state of each production unit.

Keywords: agroindustrial, characterization, logistics, maturity model, processes

Procedia PDF Downloads 134
17339 Super Harmonic Nonlinear Lateral Vibration of an Axially Moving Beam with Rotating Prismatic Joint

Authors: M. Najafi, S. Bab, F. Rahimi Dehgolan

Abstract:

The motion of an axially moving beam with rotating prismatic joint with a tip mass on the end is analyzed to investigate the nonlinear vibration and dynamic stability of the beam. The beam is moving with a harmonic axially and rotating velocity about a constant mean velocity. A time-dependent partial differential equation and boundary conditions with the aid of the Hamilton principle are derived to describe the beam lateral deflection. After the partial differential equation is discretized by the Galerkin method, the method of multiple scales is applied to obtain analytical solutions. Frequency response curves are plotted for the super harmonic resonances of the first and the second modes. The effects of non-linear term and mean velocity are investigated on the steady state response of the axially moving beam. The results are validated with numerical simulations.

Keywords: super harmonic resonances, non-linear vibration, axially moving beam, Galerkin method

Procedia PDF Downloads 385
17338 Reducing Hazardous Materials Releases from Railroad Freights through Dynamic Trip Plan Policy

Authors: Omar A. Abuobidalla, Mingyuan Chen, Satyaveer S. Chauhan

Abstract:

Railroad transportation of hazardous materials freights is important to the North America economics that supports the national’s supply chain. This paper introduces various extensions of the dynamic hazardous materials trip plan problems. The problem captures most of the operational features of a real-world railroad transportations systems that dynamically initiates a set of blocks and assigns each shipment to a single block path or multiple block paths. The dynamic hazardous materials trip plan policies have distinguishing features that are integrating the blocking plan, and the block activation decisions. We also present a non-linear mixed integer programming formulation for each variant and present managerial insights based on a hypothetical railroad network. The computation results reveal that the dynamic car scheduling policies are not only able to take advantage of the capacity of the network but also capable of diminishing the population, and environment risks by rerouting the active blocks along the least risky train services without sacrificing the cost advantage of the railroad. The empirical results of this research illustrate that the issue of integrating the blocking plan, and the train makeup of the hazardous materials freights must receive closer attentions.

Keywords: dynamic car scheduling, planning and scheduling hazardous materials freights, airborne hazardous materials, gaussian plume model, integrated blocking and routing plans, box model

Procedia PDF Downloads 202
17337 Starlink Satellite Collision Probability Simulation Based on Simplified Geometry Model

Authors: Toby Li, Julian Zhu

Abstract:

In this paper, a model based on a simplified geometry is introduced to give a very conservative collision probability prediction for the Starlink satellite in its most densely clustered region. Under the model in this paper, the probability of collision for Starlink satellite where it clustered most densely is found to be 8.484 ∗ 10^−4. It is found that the predicted collision probability increased nonlinearly with the increased safety distance set. This simple model provides evidence that the continuous development of maneuver avoidance systems is necessary for the future of the orbital safety of satellites under the harsher Lower Earth Orbit environment.

Keywords: Starlink, collision probability, debris, geometry model

Procedia PDF Downloads 74
17336 MLProxy: SLA-Aware Reverse Proxy for Machine Learning Inference Serving on Serverless Computing Platforms

Authors: Nima Mahmoudi, Hamzeh Khazaei

Abstract:

Serving machine learning inference workloads on the cloud is still a challenging task at the production level. The optimal configuration of the inference workload to meet SLA requirements while optimizing the infrastructure costs is highly complicated due to the complex interaction between batch configuration, resource configurations, and variable arrival process. Serverless computing has emerged in recent years to automate most infrastructure management tasks. Workload batching has revealed the potential to improve the response time and cost-effectiveness of machine learning serving workloads. However, it has not yet been supported out of the box by serverless computing platforms. Our experiments have shown that for various machine learning workloads, batching can hugely improve the system’s efficiency by reducing the processing overhead per request. In this work, we present MLProxy, an adaptive reverse proxy to support efficient machine learning serving workloads on serverless computing systems. MLProxy supports adaptive batching to ensure SLA compliance while optimizing serverless costs. We performed rigorous experiments on Knative to demonstrate the effectiveness of MLProxy. We showed that MLProxy could reduce the cost of serverless deployment by up to 92% while reducing SLA violations by up to 99% that can be generalized across state-of-the-art model serving frameworks.

Keywords: serverless computing, machine learning, inference serving, Knative, google cloud run, optimization

Procedia PDF Downloads 171
17335 Study of the Behavior of Bolted Joints with and Without Reinforcement

Authors: Karim Akkouche

Abstract:

Many methods have been developed for characterizing the behavior of bolted joints. However, in the presence of a certain model of stiffeners, no orientation was given in relation to their modeling. To this end, multitude of coarse errors can arise in the reproduction of the propagation of efforts and in representation of the modes of deformations. Considering these particularities, a numerical investigation was carried out in our laboratory. In this paper we will present a comparative study between three types of assemblies. A non-linear 3D modeling was chosen, given that it takes into consideration geometric and material non-linearity, using the Finite Element calculation code ABAQUS. Initially, we evaluated the influence of the presence of each stiffener on the "global" behavior of the assemblies, this by analyzing their Moment-Rotation curves, also by referring to the classification system proposed by NF EN 1993- 1.8 which is based on the resisting moment Mj-Rd and the initial stiffness Sj.int. In a second step, we evaluated the "local" behavior of their components by referring to the stress-strain curves.

Keywords: assembly, post-beam, end plate, nonlinearity

Procedia PDF Downloads 68
17334 Modeling and Validation of Microspheres Generation in the Modified T-Junction Device

Authors: Lei Lei, Hongbo Zhang, Donald J. Bergstrom, Bing Zhang, K. Y. Song, W. J. Zhang

Abstract:

This paper presents a model for a modified T-junction device for microspheres generation. The numerical model is developed using a commercial software package: COMSOL Multiphysics. In order to test the accuracy of the numerical model, multiple variables, such as the flow rate of cross-flow, fluid properties, structure, and geometry of the microdevice are applied. The results from the model are compared with the experimental results in the diameter of the microsphere generated. The comparison shows a good agreement. Therefore the model is useful in further optimization of the device and feedback control of microsphere generation if any.

Keywords: CFD modeling, validation, microsphere generation, modified T-junction

Procedia PDF Downloads 694
17333 Modeling User Context Using CEAR Diagram

Authors: Ravindra Dastikop, G. S. Thyagaraju, U. P. Kulkarni

Abstract:

Even though the number of context aware applications is increasing day by day along with the users, till today there is no generic programming paradigm for context aware applications. This situation could be remedied by design and developing the appropriate context modeling and programming paradigm for context aware applications. In this paper, we are proposing the static context model and metrics for validating the expressiveness and understandability of the model. The proposed context modeling is a way of describing a situation of user using context entities , attributes and relationships .The model which is an extended and hybrid version of ER model, ontology model and Graphical model is specifically meant for expressing and understanding the user situation in context aware environment. The model is useful for understanding context aware problems, preparing documentation and designing programs and databases. The model makes use of context entity attributes relationship (CEAR) diagram for representation of association between the context entities and attributes. We have identified a new set of graphical notations for improving the expressiveness and understandability of context from the end user perspective .

Keywords: user context, context entity, context entity attributes, situation, sensors, devices, relationships, actors, expressiveness, understandability

Procedia PDF Downloads 336
17332 Comparative Analysis of Simulation-Based and Mixed-Integer Linear Programming Approaches for Optimizing Building Modernization Pathways Towards Decarbonization

Authors: Nico Fuchs, Fabian Wüllhorst, Laura Maier, Dirk Müller

Abstract:

The decarbonization of building stocks necessitates the modernization of existing buildings. Key measures for this include reducing energy demands through insulation of the building envelope, replacing heat generators, and installing solar systems. Given limited financial resources, it is impractical to modernize all buildings in a portfolio simultaneously; instead, prioritization of buildings and modernization measures for a given planning horizon is essential. Optimization models for modernization pathways can assist portfolio managers in this prioritization. However, modeling and solving these large-scale optimization problems, often represented as mixed-integer problems (MIP), necessitates simplifying the operation of building energy systems particularly with respect to system dynamics and transient behavior. This raises the question of which level of simplification remains sufficient to accurately account for realistic costs and emissions of building energy systems, ensuring a fair comparison of different modernization measures. This study addresses this issue by comparing a two-stage simulation-based optimization approach with a single-stage mathematical optimization in a mixed-integer linear programming (MILP) formulation. The simulation-based approach serves as a benchmark for realistic energy system operation but requires a restriction of the solution space to discrete choices of modernization measures, such as the sizing of heating systems. After calculating the operation of different energy systems in terms of the resulting final energy demands in simulation models on a first stage, the results serve as input for a second stage MILP optimization, where the design of each building in the portfolio is optimized. In contrast to the simulation-based approach, the MILP-based approach can capture a broader variety of modernization measures due to the efficiency of MILP solvers but necessitates simplifying the building energy system operation. Both approaches are employed to determine the cost-optimal design and dimensioning of several buildings in a portfolio to meet climate targets within limited yearly budgets, resulting in a modernization pathway for the entire portfolio. The comparison reveals that the MILP formulation successfully captures design decisions of building energy systems, such as the selection of heating systems and the modernization of building envelopes. However, the results regarding the optimal dimensioning of heating technologies differ from the results of the two-stage simulation-based approach, as the MILP model tends to overestimate operational efficiency, highlighting the limitations of the MILP approach.

Keywords: building energy system optimization, model accuracy in optimization, modernization pathways, building stock decarbonization

Procedia PDF Downloads 16
17331 Image-Based (RBG) Technique for Estimating Phosphorus Levels of Different Crops

Authors: M. M. Ali, Ahmed Al- Ani, Derek Eamus, Daniel K. Y. Tan

Abstract:

In this glasshouse study, we developed the new image-based non-destructive technique for detecting leaf P status of different crops such as cotton, tomato and lettuce. Plants were allowed to grow on nutrient media containing different P concentrations, i.e. 0%, 50% and 100% of recommended P concentration (P0 = no P, L; P1 = 2.5 mL 10 L-1 of P and P2 = 5 mL 10 L-1 of P as NaH2PO4). After 10 weeks of growth, plants were harvested and data on leaf P contents were collected using the standard destructive laboratory method and at the same time leaf images were collected by a handheld crop image sensor. We calculated leaf area, leaf perimeter and RGB (red, green and blue) values of these images. This data was further used in the linear discriminant analysis (LDA) to estimate leaf P contents, which successfully classified these plants on the basis of leaf P contents. The data indicated that P deficiency in crop plants can be predicted using the image and morphological data. Our proposed non-destructive imaging method is precise in estimating P requirements of different crop species.

Keywords: image-based techniques, leaf area, leaf P contents, linear discriminant analysis

Procedia PDF Downloads 372
17330 Effects of Crushed Waste Aggregate from the Manufacture of Clay Bricks on Rendering Cement Mortar Performance

Authors: Benmalek M. Larbi, R. Harbi, S. Boukor

Abstract:

This paper reports an experimental work that aimed to investigate the effects of clay brick waste, as part of fine aggregate, on rendering mortar performance. The brick, in crushed form, was from a local brick manufacturer that was rejected due to being of-standard. It was used to replace 33.33 %, 50 %, 66.66 % and 100 % by weight of the quarry sand in mortar. Effects of the brick replacement on the mortar key properties intended for wall plastering were investigated; these are workability, compressive strength, flexural strength, linear shrinkage, water absorption by total immersion and by capillary suction. The results showed that as the brick replacement level increased, the mortar workability reduced. The linear shrinkage increases over time and decreases with the introduction of brick waste. The compressive and flexural strengths decrease with the increase of brick waste because of their great water absorption.

Keywords: clay brick waste, mortar, properties, quarry sand

Procedia PDF Downloads 251
17329 Nondestructive Testing for Reinforced Concrete Buildings with Active Infrared Thermography

Authors: Huy Q. Tran, Jungwon Huh, Kiseok Kwak, Choonghyun Kang

Abstract:

Infrared thermography (IRT) technique has been proven to be a good method for nondestructive evaluation of concrete material. In the building, a broad range of applications has been used such as subsurface defect inspection, energy loss, and moisture detection. The purpose of this research is to consider the qualitative and quantitative performance of reinforced concrete deteriorations using active infrared thermography technique. An experiment of three different heating regimes was conducted on a concrete slab in the laboratory. The thermal characteristics of the IRT method, i.e., absolute contrast and observation time, are investigated. A linear relationship between the observation time and the real depth was established with a well linear regression R-squared of 0.931. The results showed that the absolute contrast above defective area increases with the rise of the size of delamination and the heating time. In addition, the depth of delamination can be predicted by using the proposal relationship of this study.

Keywords: concrete building, infrared thermography, nondestructive evaluation, subsurface delamination

Procedia PDF Downloads 278
17328 The Role of Team Efficacy and Coaching on the Relationships between Distributive and Procedural Justice and Job Engagement

Authors: Yoonhee Cho, Gye-Hoon Hong

Abstract:

This study focuses on the roles of distributive and procedural justice on job engagement. Additionally, the study focuses on whether situational factors such as team efficacy and team leaders’ coaching moderate the relationship between distributive and procedural justice and job engagement. Ordinary linear regression was used to analyze data from seven South Korean Companies (total N=346). Results confirmed the hypothesized model indicating that both distributive and procedural justices were positively related to job engagement of employees. Team efficacy and team leaders’ coaching moderated the relationship between distributive justice and job engagement whereas it brought non-significant result found for procedural justice. The facts that two types of justice and the interactive effects of two situational variables were different implied that different managerial strategies should be used when job engagement was to be enhanced.

Keywords: coaching, distributive justice, job engagement, procedural justice, team efficacy

Procedia PDF Downloads 548
17327 A Strategic Partner Evaluation Model for the Project Based Enterprises

Authors: Woosik Jang, Seung H. Han

Abstract:

The optimal partner selection is one of the most important factors to pursue the project’s success. However, in practice, there is a gaps in perception of success depending on the role of the enterprises for the projects. This frequently makes a relations between the partner evaluation results and the project’s final performances, insufficiently. To meet this challenges, this study proposes a strategic partner evaluation model considering the perception gaps between enterprises. A total 3 times of survey was performed; factor selection, perception gap analysis, and case application. After then total 8 factors are extracted from independent sample t-test and Borich model to set-up the evaluation model. Finally, through the case applications, only 16 enterprises are re-evaluated to “Good” grade among the 22 “Good” grade from existing model. On the contrary, 12 enterprises are re-evaluated to “Good” grade among the 19 “Bad” grade from existing model. Consequently, the perception gaps based evaluation model is expected to improve the decision making quality and also enhance the probability of project’s success.

Keywords: partner evaluation model, project based enterprise, decision making, perception gap, project performance

Procedia PDF Downloads 152
17326 Development of a Model for Predicting Radiological Risks in Interventional Cardiology

Authors: Stefaan Carpentier, Aya Al Masri, Fabrice Leroy, Thibault Julien, Safoin Aktaou, Malorie Martin, Fouad Maaloul

Abstract:

Introduction: During an 'Interventional Radiology (IR)' procedure, the patient's skin-dose may become very high for a burn, necrosis, and ulceration to appear. In order to prevent these deterministic effects, a prediction of the peak skin-dose for the patient is important in order to improve the post-operative care to be given to the patient. The objective of this study is to estimate, before the intervention, the patient dose for ‘Chronic Total Occlusion (CTO)’ procedures by selecting relevant clinical indicators. Materials and methods: 103 procedures were performed in the ‘Interventional Cardiology (IC)’ department using a Siemens Artis Zee image intensifier that provides the Air Kerma of each IC exam. Peak Skin Dose (PSD) was measured for each procedure using radiochromic films. Patient parameters such as sex, age, weight, and height were recorded. The complexity index J-CTO score, specific to each intervention, was determined by the cardiologist. A correlation method applied to these indicators allowed to specify their influence on the dose. A predictive model of the dose was created using multiple linear regressions. Results: Out of 103 patients involved in the study, 5 were excluded for clinical reasons and 2 for placement of radiochromic films outside the exposure field. 96 2D-dose maps were finally used. The influencing factors having the highest correlation with the PSD are the patient's diameter and the J-CTO score. The predictive model is based on these parameters. The comparison between estimated and measured skin doses shows an average difference of 0.85 ± 0.55 Gy for doses of less than 6 Gy. The mean difference between air-Kerma and PSD is 1.66 Gy ± 1.16 Gy. Conclusion: Using our developed method, a first estimate of the dose to the skin of the patient is available before the start of the procedure, which helps the cardiologist in carrying out its intervention. This estimation is more accurate than that provided by the Air-Kerma.

Keywords: chronic total occlusion procedures, clinical experimentation, interventional radiology, patient's peak skin dose

Procedia PDF Downloads 128
17325 Performance and Limitations of Likelihood Based Information Criteria and Leave-One-Out Cross-Validation Approximation Methods

Authors: M. A. C. S. Sampath Fernando, James M. Curran, Renate Meyer

Abstract:

Model assessment, in the Bayesian context, involves evaluation of the goodness-of-fit and the comparison of several alternative candidate models for predictive accuracy and improvements. In posterior predictive checks, the data simulated under the fitted model is compared with the actual data. Predictive model accuracy is estimated using information criteria such as the Akaike information criterion (AIC), the Bayesian information criterion (BIC), the Deviance information criterion (DIC), and the Watanabe-Akaike information criterion (WAIC). The goal of an information criterion is to obtain an unbiased measure of out-of-sample prediction error. Since posterior checks use the data twice; once for model estimation and once for testing, a bias correction which penalises the model complexity is incorporated in these criteria. Cross-validation (CV) is another method used for examining out-of-sample prediction accuracy. Leave-one-out cross-validation (LOO-CV) is the most computationally expensive variant among the other CV methods, as it fits as many models as the number of observations. Importance sampling (IS), truncated importance sampling (TIS) and Pareto-smoothed importance sampling (PSIS) are generally used as approximations to the exact LOO-CV and utilise the existing MCMC results avoiding expensive computational issues. The reciprocals of the predictive densities calculated over posterior draws for each observation are treated as the raw importance weights. These are in turn used to calculate the approximate LOO-CV of the observation as a weighted average of posterior densities. In IS-LOO, the raw weights are directly used. In contrast, the larger weights are replaced by their modified truncated weights in calculating TIS-LOO and PSIS-LOO. Although, information criteria and LOO-CV are unable to reflect the goodness-of-fit in absolute sense, the differences can be used to measure the relative performance of the models of interest. However, the use of these measures is only valid under specific circumstances. This study has developed 11 models using normal, log-normal, gamma, and student’s t distributions to improve the PCR stutter prediction with forensic data. These models are comprised of four with profile-wide variances, four with locus specific variances, and three which are two-component mixture models. The mean stutter ratio in each model is modeled as a locus specific simple linear regression against a feature of the alleles under study known as the longest uninterrupted sequence (LUS). The use of AIC, BIC, DIC, and WAIC in model comparison has some practical limitations. Even though, IS-LOO, TIS-LOO, and PSIS-LOO are considered to be approximations of the exact LOO-CV, the study observed some drastic deviations in the results. However, there are some interesting relationships among the logarithms of pointwise predictive densities (lppd) calculated under WAIC and the LOO approximation methods. The estimated overall lppd is a relative measure that reflects the overall goodness-of-fit of the model. Parallel log-likelihood profiles for the models conditional on equal posterior variances in lppds were observed. This study illustrates the limitations of the information criteria in practical model comparison problems. In addition, the relationships among LOO-CV approximation methods and WAIC with their limitations are discussed. Finally, useful recommendations that may help in practical model comparisons with these methods are provided.

Keywords: cross-validation, importance sampling, information criteria, predictive accuracy

Procedia PDF Downloads 384