Search results for: time delayed SIR epidemic model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30844

Search results for: time delayed SIR epidemic model

29344 Estimating Gait Parameter from Digital RGB Camera Using Real Time AlphaPose Learning Architecture

Authors: Murad Almadani, Khalil Abu-Hantash, Xinyu Wang, Herbert Jelinek, Kinda Khalaf

Abstract:

Gait analysis is used by healthcare professionals as a tool to gain a better understanding of the movement impairment and track progress. In most circumstances, monitoring patients in their real-life environments with low-cost equipment such as cameras and wearable sensors is more important. Inertial sensors, on the other hand, cannot provide enough information on angular dynamics. This research offers a method for tracking 2D joint coordinates using cutting-edge vision algorithms and a single RGB camera. We provide an end-to-end comprehensive deep learning pipeline for marker-less gait parameter estimation, which, to our knowledge, has never been done before. To make our pipeline function in real-time for real-world applications, we leverage the AlphaPose human posture prediction model and a deep learning transformer. We tested our approach on the well-known GPJATK dataset, which produces promising results.

Keywords: gait analysis, human pose estimation, deep learning, real time gait estimation, AlphaPose, transformer

Procedia PDF Downloads 114
29343 Stock Movement Prediction Using Price Factor and Deep Learning

Authors: Hy Dang, Bo Mei

Abstract:

The development of machine learning methods and techniques has opened doors for investigation in many areas such as medicines, economics, finance, etc. One active research area involving machine learning is stock market prediction. This research paper tries to consider multiple techniques and methods for stock movement prediction using historical price or price factors. The paper explores the effectiveness of some deep learning frameworks for forecasting stock. Moreover, an architecture (TimeStock) is proposed which takes the representation of time into account apart from the price information itself. Our model achieves a promising result that shows a potential approach for the stock movement prediction problem.

Keywords: classification, machine learning, time representation, stock prediction

Procedia PDF Downloads 143
29342 A New Car-Following Model with Consideration of the Brake Light

Authors: Zhiyuan Tang, Ju Zhang, Wenyuan Wu

Abstract:

In this research, a car-following model with consideration of the status of the brake light is proposed. The numerical results show that the stability of the traffic flow is improved. The ability of the brake light to reduce car accident is also showed.

Keywords: brake light, car-following model, traffic flow, regional planning, transportation

Procedia PDF Downloads 575
29341 Collision Avoidance Based on Model Predictive Control for Nonlinear Octocopter Model

Authors: Doğan Yıldız, Aydan Müşerref Erkmen

Abstract:

The controller of the octocopter is mostly based on the PID controller. For complex maneuvers, PID controllers have limited performance capability like in collision avoidance. When an octocopter needs avoidance from an obstacle, it must instantly show an agile maneuver. Also, this kind of maneuver is affected severely by the nonlinear characteristic of octocopter. When these kinds of limitations are considered, the situation is highly challenging for the PID controller. In the proposed study, these challenges are tried to minimize by using the model predictive controller (MPC) for collision avoidance with a nonlinear octocopter model. The aim is to show that MPC-based collision avoidance has the capability to deal with fast varying conditions in case of obstacle detection and diminish the nonlinear effects of octocopter with varying disturbances.

Keywords: model predictive control, nonlinear octocopter model, collision avoidance, obstacle detection

Procedia PDF Downloads 187
29340 An Alternative Richards’ Growth Model Based on Hyperbolic Sine Function

Authors: Samuel Oluwafemi Oyamakin, Angela Unna Chukwu

Abstract:

Richrads growth equation being a generalized logistic growth equation was improved upon by introducing an allometric parameter using the hyperbolic sine function. The integral solution to this was called hyperbolic Richards growth model having transformed the solution from deterministic to a stochastic growth model. Its ability in model prediction was compared with the classical Richards growth model an approach which mimicked the natural variability of heights/diameter increment with respect to age and therefore provides a more realistic height/diameter predictions using the coefficient of determination (R2), Mean Absolute Error (MAE) and Mean Square Error (MSE) results. The Kolmogorov-Smirnov test and Shapiro-Wilk test was also used to test the behavior of the error term for possible violations. The mean function of top height/Dbh over age using the two models under study predicted closely the observed values of top height/Dbh in the hyperbolic Richards nonlinear growth models better than the classical Richards growth model.

Keywords: height, diameter at breast height, DBH, hyperbolic sine function, Pinus caribaea, Richards' growth model

Procedia PDF Downloads 387
29339 Semi-Empirical Modeling of Heat Inactivation of Enterococci and Clostridia During the Hygienisation in Anaerobic Digestion Process

Authors: Jihane Saad, Thomas Lendormi, Caroline Le Marechal, Anne-marie Pourcher, Céline Druilhe, Jean-louis Lanoiselle

Abstract:

Agricultural anaerobic digestion consists in the conversion of animal slurry and manure into biogas and digestate. They need, however, to be treated at 70 ºC during 60 min before anaerobic digestion according to the European regulation (EC n°1069/2009 & EU n°142/2011). The impact of such heat treatment on the outcome of bacteria has been poorly studied up to now. Moreover, a recent study¹ has shown that enterococci and clostridia are still detected despite the application of such thermal treatment, questioning the relevance of this approach for the hygienisation of digestate. The aim of this study is to establish the heat inactivation kinetics of two species of enterococci (Enterococcus faecalis and Enterococcus faecium) and two species of clostridia (Clostridioides difficile and Clostridium novyi as a non-toxic model for Clostridium botulinum of group III). A pure culture of each strain was prepared in a specific sterile medium at concentration of 10⁴ – 10⁷ MPN / mL (Most Probable number), depending on the bacterial species. Bacterial suspensions were then filled in sterilized capillary tubes and placed in a water or oil bath at desired temperature for a specific period of time. Each bacterial suspension was enumerated using a MPN approach, and tests were repeated three times for each temperature/time couple. The inactivation kinetics of the four indicator bacteria is described using the Weibull model and the classical Bigelow model of first-order kinetics. The Weibull model takes biological variation, with respect to thermal inactivation, into account and is basically a statistical model of distribution of inactivation times as the classical first-order approach is a special case of the Weibull model. The heat treatment at 70 ºC / 60 min contributes to a reduction greater than 5 log10 for E. faecium and E. faecalis. However, it results only in a reduction of about 0.7 log10 for C. difficile and an increase of 0.5 log10 for C. novyi. Application of treatments at higher temperatures is required to reach a reduction greater or equal to 3 log10 for C. novyi (such as 30 min / 100 ºC, 13 min / 105 ºC, 3 min / 110 ºC, and 1 min / 115 ºC), raising the question of the relevance of the application of heat treatment at 70 ºC / 60 min for these spore-forming bacteria. To conclude, the heat treatment (70 ºC / 60 min) defined by the European regulation is sufficient to inactivate non-sporulating bacteria. Higher temperatures (> 100 ºC) are required as far as spore-forming bacteria concerns to reach a 3 log10 reduction (sporicidal activity).

Keywords: heat treatment, enterococci, clostridia, inactivation kinetics

Procedia PDF Downloads 105
29338 DNA Methylation Changes Caused by Lawsone

Authors: Zuzana Poborilova, Anna B. Ohlsson, Torkel Berglund, Anna Vildova, Petr Babula

Abstract:

Lawsone is a pigment that occurs naturally in plants. It has been used as a skin and hair dye for a long time. Moreover, its different biological activities have been reported. The present study focused on the effect of lawsone on a plant cell model represented by tobacco BY-2 cell suspension culture, which is used as a model comparable with the HeLa cells. It has been shown that lawsone inhibits the cell growth in the concentration-dependent manner. In addition, changes in DNA methylation level have been determined. We observed decreasing level of DNA methylation in the presence of increasing concentrations of lawsone. These results were accompanied with overproduction of reactive oxygen species (ROS). Since epigenetic modifications can be caused by different stress factors, there could be a connection between the changes in the level of DNA methylation and ROS production caused by lawsone.

Keywords: DNA methylation, lawsone, naphthoquinone, reactive oxygen species

Procedia PDF Downloads 422
29337 Fair Value Accounting and Evolution of the Ohlson Model

Authors: Mohamed Zaher Bouaziz

Abstract:

Our study examines the Ohlson Model, which links a company's market value to its equity and net earnings, in the context of the evolution of the Canadian accounting model, characterized by more extensive use of fair value and a broader measure of performance after IFRS adoption. Our hypothesis is that if equity is reported at its fair value, this valuation is closely linked to market capitalization, so the weight of earnings weakens or even disappears in the Ohlson Model. Drawing on Canada's adoption of the International Financial Reporting Standards (IFRS), our results support our hypothesis that equity appears to include most of the relevant information for investors, while earnings have become less important. However, the predictive power of earnings does not disappear.

Keywords: fair value accounting, Ohlson model, IFRS adoption, value-relevance of equity and earnings

Procedia PDF Downloads 183
29336 Modelling of Solidification in a Latent Thermal Energy Storage with a Finned Tube Bundle Heat Exchanger Unit

Authors: Remo Waser, Simon Maranda, Anastasia Stamatiou, Ludger J. Fischer, Joerg Worlitschek

Abstract:

In latent heat storage, a phase change material (PCM) is used to store thermal energy. The heat transfer rate during solidification is limited and considered as a key challenge in the development of latent heat storages. Thus, finned heat exchangers (HEX) are often utilized to increase the heat transfer rate of the storage system. In this study, a new modeling approach to calculating the heat transfer rate in latent thermal energy storages with complex HEX geometries is presented. This model allows for an optimization of the HEX design in terms of costs and thermal performance of the system. Modeling solidification processes requires the calculation of time-dependent heat conduction with moving boundaries. Commonly used computational fluid dynamic (CFD) methods enable the analysis of the heat transfer in complex HEX geometries. If applied to the entire storage, the drawback of this approach is the high computational effort due to small time steps and fine computational grids required for accurate solutions. An alternative to describe the process of solidification is the so-called temperature-based approach. In order to minimize the computational effort, a quasi-stationary assumption can be applied. This approach provides highly accurate predictions for tube heat exchangers. However, it shows unsatisfactory results for more complex geometries such as finned tube heat exchangers. The presented simulation model uses a temporal and spatial discretization of heat exchanger tube. The spatial discretization is based on the smallest possible symmetric segment of the HEX. The heat flow in each segment is calculated using finite volume method. Since the heat transfer fluid temperature can be derived using energy conservation equations, the boundary conditions at the inner tube wall is dynamically updated for each time step and segment. The model allows a prediction of the thermal performance of latent thermal energy storage systems using complex HEX geometries with considerably low computational effort.

Keywords: modelling of solidification, finned tube heat exchanger, latent thermal energy storage

Procedia PDF Downloads 263
29335 A Constitutive Model of Ligaments and Tendons Accounting for Fiber-Matrix Interaction

Authors: Ratchada Sopakayang, Gerhard A. Holzapfel

Abstract:

In this study, a new constitutive model is developed to describe the hyperelastic behavior of collagenous tissues with a parallel arrangement of collagen fibers such as ligaments and tendons. The model is formulated using a continuum approach incorporating the structural changes of the main tissue components: collagen fibers, proteoglycan-rich matrix and fiber-matrix interaction. The mechanical contribution of the interaction between the fibers and the matrix is simply expressed by a coupling term. The structural change of the collagen fibers is incorporated in the constitutive model to describe the activation of the fibers under tissue straining. Finally, the constitutive model can easily describe the stress-stretch nonlinearity which occurs when a ligament/tendon is axially stretched. This study shows that the interaction between the fibers and the matrix contributes to the mechanical tissue response. Therefore, the model may lead to a better understanding of the physiological mechanisms of ligaments and tendons under axial loading.

Keywords: constitutive model, fiber-matrix, hyperelasticity, interaction, ligament, tendon

Procedia PDF Downloads 294
29334 Transfer Function Model-Based Predictive Control for Nuclear Core Power Control in PUSPATI TRIGA Reactor

Authors: Mohd Sabri Minhat, Nurul Adilla Mohd Subha

Abstract:

The 1MWth PUSPATI TRIGA Reactor (RTP) in Malaysia Nuclear Agency has been operating more than 35 years. The existing core power control is using conventional controller known as Feedback Control Algorithm (FCA). It is technically challenging to keep the core power output always stable and operating within acceptable error bands for the safety demand of the RTP. Currently, the system could be considered unsatisfactory with power tracking performance, yet there is still significant room for improvement. Hence, a new design core power control is very important to improve the current performance in tracking and regulating reactor power by controlling the movement of control rods that suit the demand of highly sensitive of nuclear reactor power control. In this paper, the proposed Model Predictive Control (MPC) law was applied to control the core power. The model for core power control was based on mathematical models of the reactor core, MPC, and control rods selection algorithm. The mathematical models of the reactor core were based on point kinetics model, thermal hydraulic models, and reactivity models. The proposed MPC was presented in a transfer function model of the reactor core according to perturbations theory. The transfer function model-based predictive control (TFMPC) was developed to design the core power control with predictions based on a T-filter towards the real-time implementation of MPC on hardware. This paper introduces the sensitivity functions for TFMPC feedback loop to reduce the impact on the input actuation signal and demonstrates the behaviour of TFMPC in term of disturbance and noise rejections. The comparisons of both tracking and regulating performance between the conventional controller and TFMPC were made using MATLAB and analysed. In conclusion, the proposed TFMPC has satisfactory performance in tracking and regulating core power for controlling nuclear reactor with high reliability and safety.

Keywords: core power control, model predictive control, PUSPATI TRIGA reactor, TFMPC

Procedia PDF Downloads 237
29333 Using the Structural Equation Model to Explain the Effect of Supervisory Practices on Regulatory Density

Authors: Jill Round

Abstract:

In the economic system, the financial sector plays a crucial role as an intermediary between market participants, other financial institutions, and customers. Financial institutions such as banks have to make decisions to satisfy the demands of all the participants by keeping abreast of regulatory change. In recent years, progress has been made regarding frameworks, development of rules, standards, and processes to manage risks in the banking sector. The increasing focus of regulators and policymakers placed on risk management, corporate governance, and the organization’s culture is of special interest as it requires a well-resourced risk controlling function, compliance function, and internal audit function. In the past years, the relevance of these functions that make up the so-called Three Lines of Defense has moved from the backroom to the boardroom. The approach of the model can vary based on the various organizational characteristics. Due to the intense regulatory requirements, organizations operating in the financial sector have more mature models. In less regulated industries there is more cloudiness about what tasks are allocated where. All parties strive to achieve their objectives through the effective management of risks and serve the identical stakeholders. Today, the Three Lines of Defense model is used throughout the world. The research looks at trends and emerging issues in the professions of the Three Lines of Defense within the banking sector. The answers are believed to helping to explain the increasing regulatory requirements for the banking sector. While the number of supervisory practices increases the risk management requirements intensify and demand more regulatory compliance at the same time. The Structural Equation Modeling (SEM) is applied by making use of conducted surveys in the research field. It aims to describe (i) the theoretical model regarding the applicable linearity relationships, (ii) the causal relationship between multiple predictors (exogenous) and multiple dependent variables (endogenous), (iii) taking into consideration the unobservable variables and (iv) the measurement errors. The surveys conducted on the research field suggest that the observable variables are caused by various latent variables. The SEM consists of the 1) measurement model and the 2) structural model. There is a detectable correlation regarding the cause-effect relationship among the performed supervisory practices and the increasing scope of regulation. Supervisory practices reinforce the regulatory density. In the past, controls were placed after supervisory practices were conducted or incidents occurred. In further research, it is of interest to examine, whether risk management is proactive, reactive to incidents and supervisory practices or can be both at the same time.

Keywords: risk management, structural equation model, supervisory practice, three lines of defense

Procedia PDF Downloads 218
29332 Burnback Analysis of Star Grain Using Level-Set Technique

Authors: Ali Yasin, Ali Kamran, Muhammad Safdar

Abstract:

In order to reduce the hefty cost involved in terms of time and project cost, the development and application of advanced numerical tools to address the burn-back analysis problem in solid rocket motor design and development is the need of time. Several advanced numerical schemes have been developed in recent times, but their usage in the design of propellant grain of solid rocket motors is very rare. In this paper, an advanced numerical technique named the Level-Set method has been utilized for the burn-back analysis of star grain to study the effect of geometrical parameters on ballistic performance indicators such as solid loading, neutrality, and sliver percentage. In the level set technique, simple finite difference methods may fail quickly and require more sophisticated non-oscillatory schemes for feasible long-time simulation. For internal ballistic calculations, a simplified equilibrium pressure method is utilized. Preliminary results of the operative conditions, for all the combustion time, of star grain burn-back using level set techniques are compared with published results using CAD technique to test the developed numerical model.

Keywords: solid rocket motor, internal ballistic, level-set technique, star grain

Procedia PDF Downloads 120
29331 Approach to Study the Workability of Concrete with the Fractal Model

Authors: Achouri Fatima, Chouicha Kaddour

Abstract:

The main parameters affecting the workability are the water content, particle size, and the total surface of the grains, as long as the mixing water begins by wetting the surface of the grains and then fills the voids between the grains to form entrapped water, the quantity of water remaining is called free water. The aim is to undertake a fractal approach through the relationship between the concrete formulation parameters and workability, to develop this approach a series of concrete taken from the literature was investigated by varying formulation parameters such as G / S, the quantity of cement C and the quantity of mixing water E. We also call on other model as the model for the thickness of the water layer and model of the thickness of the paste layer to judge their relevance, hence the following results : the relevance of the model of the thickness of the water layer is considered relevant when there is a variation in the water quantity, the model of the thickness of the layer of the paste is only applicable if we consider that the paste is made with the grain value Dmax = 2.85: value from which we see a stable model.

Keywords: concrete, fractal method, paste thickness, water thickness, workability

Procedia PDF Downloads 375
29330 Virtual Assessment of Measurement Error in the Fractional Flow Reserve

Authors: Keltoum Chahour, Mickael Binois

Abstract:

Due to a lack of standardization during the invasive fractional flow reserve (FFR) procedure, the index is subject to many sources of uncertainties. In this paper, we investigate -through simulation- the effect of the (FFR) device position and configuration on the obtained value of the (FFR) fraction. For this purpose, we use computational fluid dynamics (CFD) in a 3D domain corresponding to a diseased arterial portion. The (FFR) pressure captor is introduced inside it with a given length and coefficient of bending to capture the (FFR) value. To get over the computational limitations, basically, the time of the simulation is about 2h 15min for one (FFR) value; we generate a Gaussian Process (GP) model for (FFR) prediction. The (GP) model indicates good accuracy and demonstrates the effective error in the measurement created by the random configuration of the pressure captor.

Keywords: fractional flow reserve, Gaussian processes, computational fluid dynamics, drift

Procedia PDF Downloads 128
29329 Experimental and Numerical Investigation of “Machining Induced Residual Stresses” during Orthogonal Machining of Alloy Steel AISI 4340

Authors: Theena Thayalan, K. N. Ramesh Babu

Abstract:

Machining induced residual stress (RS) is one of the most important surface integrity parameters that characterize the near surface layer of a mechanical component, which plays a crucial role in controlling the performance, especially its fatigue life. Since experimental determination of RS is expensive and time consuming, it would be of great benefit if they could be predicted. In such case, it would be possible to select the cutting parameters required to produce a favorable RS profile. In the present study, an effort has been made to develop a 'two dimensional finite element model (FEM)' to simulate orthogonal cutting process and to predict surface and sub-surface RS using the commercial FEA software DEFORM-2D. The developed finite element model has been validated through experimental investigation of RS. In the experimentation, the orthogonal cutting tests were carried out on AISI 4340 by varying the cutting speed (VC) and uncut chip thickness (f) at three levels and the surface & sub-surface RS has been measured using XRD and Electro polishing techniques. The comparison showed that the RS obtained using developed numerical model is in reasonable agreement with that of experimental data.

Keywords: FEM, machining, residual stress, XRF

Procedia PDF Downloads 345
29328 R&D Diffusion and Productivity in a Globalized World: Country Capabilities in an MRIO Framework

Authors: S. Jimenez, R.Duarte, J.Sanchez-Choliz, I. Villanua

Abstract:

There is a certain consensus in economic literature about the factors that have influenced in historical differences in growth rates observed between developed and developing countries. However, it is less clear what elements have marked different paths of growth in developed economies in recent decades. R&D has always been seen as one of the major sources of technological progress, and productivity growth, which is directly influenced by technological developments. Following recent literature, we can say that ‘innovation pushes the technological frontier forward’ as well as encourage future innovation through the creation of externalities. In other words, productivity benefits from innovation are not fully appropriated by innovators, but it also spread through the rest of the economies encouraging absorptive capacities, what have become especially important in a context of increasing fragmentation of production This paper aims to contribute to this literature in two ways, first, exploring alternative indexes of R&D flows embodied in inter-country, inter-sectorial flows of good and services (as approximation to technology spillovers) capturing structural and technological characteristic of countries and, second, analyzing the impact of direct and embodied R&D on the evolution of labor productivity at the country/sector level in recent decades. The traditional way of calculation through a multiregional input-output framework assumes that all countries have the same capabilities to absorb technology, but it is not, each one has different structural features and, this implies, different capabilities as part of literature, claim. In order to capture these differences, we propose to use a weight based on specialization structure indexes; one related with the specialization of countries in high-tech sectors and the other one based on a dispersion index. We propose these two measures because, as far as we understood, country capabilities can be captured through different ways; countries specialization in knowledge-intensive sectors, such as Chemicals or Electrical Equipment, or an intermediate technology effort across different sectors. Results suggest the increasing importance of country capabilities while increasing the trade openness. Besides, if we focus in the country rankings, we can observe that with high-tech weighted R&D embodied countries as China, Taiwan and Germany arose the top five despite not having the highest intensities of R&D expenditure, showing the importance of country capabilities. Additionally, through a fixed effects panel data model we show that, in fact, R&D embodied is important to explain labor productivity increases, in fact, even more that direct R&D investments. This is reflecting that globalization is more important than has been said until now. However, it is true that almost all analysis done in relation with that consider the effect of t-1 direct R&D intensity over economic growth. Nevertheless, from our point of view R&D evolve as a delayed flow and it is necessary some time to be able to see its effects on the economy, as some authors have already claimed. Our estimations tend to corroborate this hypothesis obtaining a gap between 4-5 years.

Keywords: economic growth, embodied, input-output, technology

Procedia PDF Downloads 120
29327 Efficient Chiller Plant Control Using Modern Reinforcement Learning

Authors: Jingwei Du

Abstract:

The need of optimizing air conditioning systems for existing buildings calls for control methods designed with energy-efficiency as a primary goal. The majority of current control methods boil down to two categories: empirical and model-based. To be effective, the former heavily relies on engineering expertise and the latter requires extensive historical data. Reinforcement Learning (RL), on the other hand, is a model-free approach that explores the environment to obtain an optimal control strategy often referred to as “policy”. This research adopts Proximal Policy Optimization (PPO) to improve chiller plant control, and enable the RL agent to collaborate with experienced engineers. It exploits the fact that while the industry lacks historical data, abundant operational data is available and allows the agent to learn and evolve safely under human supervision. Thanks to the development of language models, renewed interest in RL has led to modern, online, policy-based RL algorithms such as the PPO. This research took inspiration from “alignment”, a process that utilizes human feedback to finetune the pretrained model in case of unsafe content. The methodology can be summarized into three steps. First, an initial policy model is generated based on minimal prior knowledge. Next, the prepared PPO agent is deployed so feedback from both critic model and human experts can be collected for future finetuning. Finally, the agent learns and adapts itself to the specific chiller plant, updates the policy model and is ready for the next iteration. Besides the proposed approach, this study also used traditional RL methods to optimize the same simulated chiller plants for comparison, and it turns out that the proposed method is safe and effective at the same time and needs less to no historical data to start up.

Keywords: chiller plant, control methods, energy efficiency, proximal policy optimization, reinforcement learning

Procedia PDF Downloads 18
29326 Kalman Filter for Bilinear Systems with Application

Authors: Abdullah E. Al-Mazrooei

Abstract:

In this paper, we present a new kind of the bilinear systems in the form of state space model. The evolution of this system depends on the product of state vector by its self. The well known Lotak Volterra and Lorenz models are special cases of this new model. We also present here a generalization of Kalman filter which is suitable to work with the new bilinear model. An application to real measurements is introduced to illustrate the efficiency of the proposed algorithm.

Keywords: bilinear systems, state space model, Kalman filter, application, models

Procedia PDF Downloads 437
29325 Investigation of Different Control Stratgies for UPFC Decoupled Model and the Impact of Location on Control Parameters

Authors: S. A. Al-Qallaf, S. A. Al-Mawsawi, A. Haider

Abstract:

In order to evaluate the performance of a unified power flow controller (UPFC), mathematical models for steady state and dynamic analysis are to be developed. The steady state model is mainly concerned with the incorporation of the UPFC in load flow studies. Several load flow models for UPFC have been introduced in literature, and one of the most reliable models is the decoupled UPFC model. In spite of UPFC decoupled load flow model simplicity, it is more robust compared to other UPFC load flow models and it contains unique capabilities. Some shortcoming such as additional set of nonlinear equations are to be solved separately after the load flow solution is obtained. The aim of this study is to investigate the different control strategies that can be realized in the decoupled load flow model (individual control and combined control), and the impact of the location of the UPFC in the network on its control parameters.

Keywords: UPFC, decoupled model, load flow, control parameters

Procedia PDF Downloads 548
29324 Modeling Environmental, Social, and Governance Financial Assets with Lévy Subordinated Processes and Option Pricing

Authors: Abootaleb Shirvani, Svetlozar Rachev

Abstract:

ESG stands for Environmental, Social, and Governance and is a non-financial factor that investors use to specify material risks and growth opportunities in their analysis process. ESG ratings provide a quantitative measure of socially responsible investment, and it is essential to incorporate ESG ratings when modeling the dynamics of asset returns. In this article, we propose a triple subordinated Lévy process for incorporating numeric ESG ratings into dynamic asset pricing theory to model the time series properties of the stock returns. The motivation for introducing three layers of subordinator is twofold. The first two layers of subordinator capture the skew and fat-tailed properties of the stock return distribution that cannot be explained well by the existing Lévy subordinated model. The third layer of the subordinator introduces ESG valuation and incorporates numeric ESG ratings into dynamic asset pricing theory and option pricing. We employ the triple subordinator Lévy model for developing the ESG-valued stock return model, derive the implied ESG score surfaces for Microsoft, Apple, and Amazon stock returns, and compare the shape of the ESG implied surface scores for these stocks.

Keywords: ESG scores, dynamic asset pricing theory, multiple subordinated modeling, Lévy processes, option pricing

Procedia PDF Downloads 77
29323 Swallowing Outcomes in Supraglottic Cancer Patients after Trans-Oral Robotic Surgery (TORS) Provided with Early Dysphagia Management Using Standardized Functional and Objective Measures

Authors: Hitesh Gupta, Surender Dabas

Abstract:

TORS is increasingly gaining widespread use and has been explored as minimally invasive surgery for the treatment of supraglottic cancer (SGC). Being a central critical role of Supraglottis in deglutition, swallowing outcomes post TORS remain a most important factor. Available published studies show inconsistent swallowing outcomes and are deficient in standardized outcome measures, description of swallowing recovery and rehabilitation. So, the objective of this study is to find out swallowing outcomes in SGC patients after TORS provided with early dysphagia management using standardized measures. Prospectively 16 patients were recruited in the study who underwent TORS for primary tumor of Supraglottis, involving one or more sub-sites or invading to sites other than Supraglottis at the BLK Super Specialty Hospital, New Delhi from March 2019 to June 2020. All patients were evaluated for dysphagia with subsequent swallowing rehabilitation on post operative day 3 in the hospital or at the time of discharge, whichever was earlier. Functional oral intake scale (FOIS) and penetration-aspiration score (PAS) were used as outcome measures to quantify swallowing recovery at one month and six month post operatively. Post TORS, patients achieved functional swallow in less than one month, where resection was limited to Supraglottis, while the recovery was delayed in patients with extended resection to tongue base or hypopharynx. Overall, out of Total 16 cases including all supraglottis sub-catagories, 13 (81%) could remove their NG tube (FOIS ≥5 and PAS=1 ) within 6 months. In which 8 cases(62%) achieved functional swallow in less than one month. Swallowing outcomes post TORS supraglottic laryngectomy are favorable if provided with early dysphagia management (or swallowing rehabilitation).

Keywords: dysphagia, supraglottic cancer, swallowing, TORS

Procedia PDF Downloads 100
29322 DNA Methylation 6mA and Histone Methylation Involved in Multi-/Trans-Generational Reproductive Effects in Caenorhabditis elegans Induced by Atrazine

Authors: Jiechen Yin, Xiang Hong, Ran Liu

Abstract:

Atrazine (ATR), a widely used triazine herbicide, is an environmental endocrine disruptor that can cause health problems. However, whether there are multi/trans-generational reproductive impacts of ATR have not been studied to our best knowledge. Therefore, in this study, Caenorhabditis elegans was used as a preferable model organism to identify the multi/trans-generational reproductive toxicity of ATR. L1 larvae were exposed to different concentrations (0.0004–40 mg/L) of ATR for 48 h. Successive generations (F1 to F5) were fed without ATR and consecutive exposure. The results showed that ATR exposure during P0 decreased fecundity, including a reduction in fertilized eggs, oocytes, and ovulation rate, delayed gonadal development, and decreased the relative area of the gonad arm and germ cell number. Furthermore, continuous ATR exposure (P0–F5) causes a significant increase in reproductive toxicity in subsequent generations, although no significant toxicity occurred in the P0 generation after exposure to environmental-related concentrations, suggesting that ATR exposure might have cumulative effects. Likewise, parental exposure to ATR caused transgenerational toxicity impairments. Interestingly, reproductive toxicity not development toxicity was transmitted to several generations (F1–F4), and the F2 generation showed the most notable changes. QRT-PCR results showed that genes related to DNA methylation 6mA (damt-1, nmad-1) and histone H3 methylation (mes-4, met-2, set-25, set-2, and utx-1) can also be passed on to offspring. The function of H3K4 and H3K9 methylation were explored by using loss-of-function mutants for set-2, set-25, and met-2. Transmissible reproductive toxicity was absent in met-2(n4256), set-2(ok952), and set-25(n5021) mutants, which suggests that the histone methyltransferases H3K4 and H3K9 activity are indispensable for the transgenerational effect of ATR. Finally, the downstream genes of DNA methylation and histone H3 methylation were determined. ATR upregulated the expression of ZC317.7, hsp-6, and hsp-60. Mitochondrial stress in parental generation dependent transcription 6mA modifiers may establish these epigenetic marks in progeny.

Keywords: ATR, Caenorhabditis elegans, multi-/trans-generation, reproductive toxicity

Procedia PDF Downloads 66
29321 Exploring Deep Neural Network Compression: An Overview

Authors: Ghorab Sara, Meziani Lila, Rubin Harvey Stuart

Abstract:

The rapid growth of deep learning has led to intricate and resource-intensive deep neural networks widely used in computer vision tasks. However, their complexity results in high computational demands and memory usage, hindering real-time application. To address this, research focuses on model compression techniques. The paper provides an overview of recent advancements in compressing neural networks and categorizes the various methods into four main approaches: network pruning, quantization, network decomposition, and knowledge distillation. This paper aims to provide a comprehensive outline of both the advantages and limitations of each method.

Keywords: model compression, deep neural network, pruning, knowledge distillation, quantization, low-rank decomposition

Procedia PDF Downloads 38
29320 Modelling Spatial Dynamics of Terrorism

Authors: André Python

Abstract:

To this day, terrorism persists as a worldwide threat, exemplified by the recent deadly attacks in January 2015 in Paris and the ongoing massacres perpetrated by ISIS in Iraq and Syria. In response to this threat, states deploy various counterterrorism measures, the cost of which could be reduced through effective preventive measures. In order to increase the efficiency of preventive measures, policy-makers may benefit from accurate predictive models that are able to capture the complex spatial dynamics of terrorism occurring at a local scale. Despite empirical research carried out at country-level that has confirmed theories explaining the diffusion processes of terrorism across space and time, scholars have failed to assess diffusion’s theories on a local scale. Moreover, since scholars have not made the most of recent statistical modelling approaches, they have been unable to build up predictive models accurate in both space and time. In an effort to address these shortcomings, this research suggests a novel approach to systematically assess the theories of terrorism’s diffusion on a local scale and provide a predictive model of the local spatial dynamics of terrorism worldwide. With a focus on the lethal terrorist events that occurred after 9/11, this paper addresses the following question: why and how does lethal terrorism diffuse in space and time? Based on geolocalised data on worldwide terrorist attacks and covariates gathered from 2002 to 2013, a binomial spatio-temporal point process is used to model the probability of terrorist attacks on a sphere (the world), the surface of which is discretised in the form of Delaunay triangles and refined in areas of specific interest. Within a Bayesian framework, the model is fitted through an integrated nested Laplace approximation - a recent fitting approach that computes fast and accurate estimates of posterior marginals. Hence, for each location in the world, the model provides a probability of encountering a lethal terrorist attack and measures of volatility, which inform on the model’s predictability. Diffusion processes are visualised through interactive maps that highlight space-time variations in the probability and volatility of encountering a lethal attack from 2002 to 2013. Based on the previous twelve years of observation, the location and lethality of terrorist events in 2014 are statistically accurately predicted. Throughout the global scope of this research, local diffusion processes such as escalation and relocation are systematically examined: the former process describes an expansion from high concentration areas of lethal terrorist events (hotspots) to neighbouring areas, while the latter is characterised by changes in the location of hotspots. By controlling for the effect of geographical, economical and demographic variables, the results of the model suggest that the diffusion processes of lethal terrorism are jointly driven by contagious and non-contagious factors that operate on a local scale – as predicted by theories of diffusion. Moreover, by providing a quantitative measure of predictability, the model prevents policy-makers from making decisions based on highly uncertain predictions. Ultimately, this research may provide important complementary tools to enhance the efficiency of policies that aim to prevent and combat terrorism.

Keywords: diffusion process, terrorism, spatial dynamics, spatio-temporal modeling

Procedia PDF Downloads 345
29319 Rheological Model for Describing Spunlace Nonwoven Behavior

Authors: Sana Ridene, Soumaya Sayeb, Houda Helali, Mohammed Ben Hassen

Abstract:

Nonwoven structures have a range of applications which include Medical, filtration, geotextile and recently this unconventional fabric is finding a niche in fashion apparel. In this paper, a modified form of Vangheluwe rheological model is used to describe the mechanical behavior of nonwovens fabrics in uniaxial tension. This model is an association in parallel of three Maxwell elements characterized by damping coefficients η1, η2 and η3 and E1, E2, E3 elastic modulus and a nonlinear spring C. The model is verified experimentally with two types of nonwovens (50% viscose /50% Polyester) and (40% viscose/60% Polyester) and a range of three square weights values. Comparative analysis of the theoretical model and the experimental results of tensile test proofs a high correlation between them. The proposed model can fairly well replicate the behavior of nonwoven fabrics during relaxation and sample traction. This allowed us to predict the mechanical behavior in tension and relaxation of fabrics starting only from their technical parameters (composition and weight).

Keywords: mechanical behavior, tensile strength, relaxation, rheological model

Procedia PDF Downloads 403
29318 Thermal Modelling and Experimental Comparison for a Moving Pantograph Strip

Authors: Nicolas Delcey, Philippe Baucour, Didier Chamagne, Geneviève Wimmer, Auditeau Gérard, Bausseron Thomas, Bouger Odile, Blanvillain Gérard

Abstract:

This paper proposes a thermal study of the catenary/pantograph interface for a train in motion. A 2.5D complex model of the pantograph strip has been defined and created by a coupling between a 1D and a 2D model. Experimental and simulation results are presented and with a comparison allow validating the 2.5D model. Some physical phenomena are described and presented with the help of the model such as the stagger motion thermal effect, particular heats and the effect of the material characteristics. Finally it is possible to predict the critical thermal configuration during a train trip.

Keywords: electro-thermal studies, mathematical optimizations, multi-physical approach, numerical model, pantograph strip wear

Procedia PDF Downloads 321
29317 The Quantity and Quality of Teacher Talking Time in EFL Classroom

Authors: Hanan Abufares Elkhimry

Abstract:

Looking for more effective teaching and learning approaches, teaching instructors have been telling trainee teachers to decrease their talking time, but the problem is how best to do this. Doing classroom research, specifically in the area of teacher talking time (TTT), is worthwhile, as it could improve the quality of teaching languages, as the learners are the ones who should be practicing and using the language. This work hopes to ascertain if teachers consider this need in a way that provides the students with the opportunities to increase their production of language. This is a question that is worthwhile answering. As many researchers have found, TTT should be decreased to 30% of classroom talking time and STT should be increased up to 70%. Other researchers agree with this, but add that it should be with awareness of the quality of teacher talking time. Therefore, this study intends to investigate the balance between quantity and quality of teacher talking time in the EFL classroom. For this piece of research and in order to capture the amount of talking in a four classrooms. The amount of talking time was measured. A Checklist was used to assess the quality of the talking time In conclusion, In order to improve the quality of TTT, the results showed that teachers may use more or less than 30% of the classroom talking time and still produce a successful classroom learning experience. As well as, the important factors that can affect TTT is the English level of the students. This was clear in the classroom observations, where the highest TTT recorded was with the lowest English level group.

Keywords: teacher talking time TTT, learning experience, classroom research, effective teaching

Procedia PDF Downloads 409
29316 Generating Swarm Satellite Data Using Long Short-Term Memory and Generative Adversarial Networks for the Detection of Seismic Precursors

Authors: Yaxin Bi

Abstract:

Accurate prediction and understanding of the evolution mechanisms of earthquakes remain challenging in the fields of geology, geophysics, and seismology. This study leverages Long Short-Term Memory (LSTM) networks and Generative Adversarial Networks (GANs), a generative model tailored to time-series data, for generating synthetic time series data based on Swarm satellite data, which will be used for detecting seismic anomalies. LSTMs demonstrated commendable predictive performance in generating synthetic data across multiple countries. In contrast, the GAN models struggled to generate synthetic data, often producing non-informative values, although they were able to capture the data distribution of the time series. These findings highlight both the promise and challenges associated with applying deep learning techniques to generate synthetic data, underscoring the potential of deep learning in generating synthetic electromagnetic satellite data.

Keywords: LSTM, GAN, earthquake, synthetic data, generative AI, seismic precursors

Procedia PDF Downloads 28
29315 Consensus Problem of High-Order Multi-Agent Systems under Predictor-Based Algorithm

Authors: Cheng-Lin Liu, Fei Liu

Abstract:

For the multi-agent systems with agent's dynamics described by high-order integrator, and usual consensus algorithm composed of the state coordination control parts is proposed. Under communication delay, consensus algorithm in asynchronously-coupled form just can make the agents achieve a stationary consensus, and sufficient consensus condition is obtained based on frequency-domain analysis. To recover the original consensus state of the high-order agents without communication delay, besides, a predictor-based consensus algorithm is constructed via multiplying the delayed neighboring agents' states by a delay-related compensation part, and sufficient consensus condition is also obtained. Simulation illustrates the correctness of the results.

Keywords: high-order dynamic agents, communication delay, consensus, predictor-based algorithm

Procedia PDF Downloads 567