Search results for: times series analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30662

Search results for: times series analysis

30392 Sensitivity Analysis for 14 Bus Systems in a Distribution Network with Distributed Generators

Authors: Lakshya Bhat, Anubhav Shrivastava, Shiva Rudraswamy

Abstract:

There has been a formidable interest in the area of Distributed Generation in recent times. A wide number of loads are addressed by Distributed Generators and have better efficiency too. The major disadvantage in Distributed Generation is voltage control- is highlighted in this paper. The paper addresses voltage control at buses in IEEE 14 Bus system by regulating reactive power. An analysis is carried out by selecting the most optimum location in placing the Distributed Generators through load flow analysis and seeing where the voltage profile rises. MATLAB programming is used for simulation of voltage profile in the respective buses after introduction of DG’s. A tolerance limit of +/-5% of the base value has to be maintained. To maintain the tolerance limit, 3 methods are used. Sensitivity analysis of 3 methods for voltage control is carried out to determine the priority among the methods.

Keywords: distributed generators, distributed system, reactive power, voltage control, sensitivity analysis

Procedia PDF Downloads 672
30391 Scheduling Jobs with Stochastic Processing Times or Due Dates on a Server to Minimize the Number of Tardy Jobs

Authors: H. M. Soroush

Abstract:

The problem of scheduling products and services for on-time deliveries is of paramount importance in today’s competitive environments. It arises in many manufacturing and service organizations where it is desirable to complete jobs (products or services) with different weights (penalties) on or before their due dates. In such environments, schedules should frequently decide whether to schedule a job based on its processing time, due-date, and the penalty for tardy delivery to improve the system performance. For example, it is common to measure the weighted number of late jobs or the percentage of on-time shipments to evaluate the performance of a semiconductor production facility or an automobile assembly line. In this paper, we address the problem of scheduling a set of jobs on a server where processing times or due-dates of jobs are random variables and fixed weights (penalties) are imposed on the jobs’ late deliveries. The goal is to find the schedule that minimizes the expected weighted number of tardy jobs. The problem is NP-hard to solve; however, we explore three scenarios of the problem wherein: (i) both processing times and due-dates are stochastic; (ii) processing times are stochastic and due-dates are deterministic; and (iii) processing times are deterministic and due-dates are stochastic. We prove that special cases of these scenarios are solvable optimally in polynomial time, and introduce efficient heuristic methods for the general cases. Our computational results show that the heuristics perform well in yielding either optimal or near optimal sequences. The results also demonstrate that the stochasticity of processing times or due-dates can affect scheduling decisions. Moreover, the proposed problem is general in the sense that its special cases reduce to some new and some classical stochastic single machine models.

Keywords: number of late jobs, scheduling, single server, stochastic

Procedia PDF Downloads 457
30390 Downside Risk Analysis of the Nigerian Stock Market: A Value at Risk Approach

Authors: Godwin Chigozie Okpara

Abstract:

This paper using standard GARCH, EGARCH, and TARCH models on day of the week return series (of 246 days) from the Nigerian Stock market estimated the model variants’ VaR. An asymmetric return distribution and fat-tail phenomenon in financial time series were considered by estimating the models with normal, student t and generalized error distributions. The analysis based on Akaike Information Criterion suggests that the EGARCH model with student t innovation distribution can furnish more accurate estimate of VaR. In the light of this, we apply the likelihood ratio tests of proportional failure rates to VaR derived from EGARCH model in order to determine the short and long positions VaR performances. The result shows that as alpha ranges from 0.05 to 0.005 for short positions, the failure rate significantly exceeds the prescribed quintiles while it however shows no significant difference between the failure rate and the prescribed quantiles for long positions. This suggests that investors and portfolio managers in the Nigeria stock market have long trading position or can buy assets with concern on when the asset prices will fall. Precisely, the VaR estimates for the long position range from -4.7% for 95 percent confidence level to -10.3% for 99.5 percent confidence level.

Keywords: downside risk, value-at-risk, failure rate, kupiec LR tests, GARCH models

Procedia PDF Downloads 412
30389 Time-Series Load Data Analysis for User Power Profiling

Authors: Mahdi Daghmhehci Firoozjaei, Minchang Kim, Dima Alhadidi

Abstract:

In this paper, we present a power profiling model for smart grid consumers based on real time load data acquired smart meters. It profiles consumers’ power consumption behaviour using the dynamic time warping (DTW) clustering algorithm. Due to the invariability of signal warping of this algorithm, time-disordered load data can be profiled and consumption features be extracted. Two load types are defined and the related load patterns are extracted for classifying consumption behaviour by DTW. The classification methodology is discussed in detail. To evaluate the performance of the method, we analyze the time-series load data measured by a smart meter in a real case. The results verify the effectiveness of the proposed profiling method with 90.91% true positive rate for load type clustering in the best case.

Keywords: power profiling, user privacy, dynamic time warping, smart grid

Procedia PDF Downloads 102
30388 Finite Element Modeling of Ultrasonic Shot Peening Process using Multiple Pin Impacts

Authors: Chao-xun Liu, Shi-hong Lu

Abstract:

In spite of its importance to the aerospace and automobile industries, little or no attention has been devoted to the accurate modeling of the ultrasonic shot peening (USP) process. It is therefore the purpose of this study to conduct finite element analysis of the process using a realistic multiple pin impacts model with the explicit solver of ABAQUS. In this paper, we research the effect of several key parameters on the residual stress distribution within the target, including impact velocity, incident angle, friction coefficient between pins and target and impact number of times were investigated. The results reveal that the impact velocity and impact number of times have obvious effect and impacting vertically could produce the most perfect residual stress distribution. Then we compare the results with the date in USP experiment and verify the exactness of the model. The analysis of the multiple pin impacts date reveal the relationships between peening process parameters and peening quality, which are useful for identifying the parameters which need to be controlled and regulated in order to produce a more beneficial compressive residual stress distribution within the target.

Keywords: ultrasonic shot peening, finite element, multiple pins, residual stress, numerical simulation

Procedia PDF Downloads 422
30387 Retail Strategy to Reduce Waste Keeping High Profit Utilizing Taylor's Law in Point-of-Sales Data

Authors: Gen Sakoda, Hideki Takayasu, Misako Takayasu

Abstract:

Waste reduction is a fundamental problem for sustainability. Methods for waste reduction with point-of-sales (POS) data are proposed, utilizing the knowledge of a recent econophysics study on a statistical property of POS data. Concretely, the non-stationary time series analysis method based on the Particle Filter is developed, which considers abnormal fluctuation scaling known as Taylor's law. This method is extended for handling incomplete sales data because of stock-outs by introducing maximum likelihood estimation for censored data. The way for optimal stock determination with pricing the cost of waste reduction is also proposed. This study focuses on the examination of the methods for large sales numbers where Taylor's law is obvious. Numerical analysis using aggregated POS data shows the effectiveness of the methods to reduce food waste maintaining a high profit for large sales numbers. Moreover, the way of pricing the cost of waste reduction reveals that a small profit loss realizes substantial waste reduction, especially in the case that the proportionality constant  of Taylor’s law is small. Specifically, around 1% profit loss realizes half disposal at =0.12, which is the actual  value of processed food items used in this research. The methods provide practical and effective solutions for waste reduction keeping a high profit, especially with large sales numbers.

Keywords: food waste reduction, particle filter, point-of-sales, sustainable development goals, Taylor's law, time series analysis

Procedia PDF Downloads 102
30386 Sensitivity Analysis for 14 Bus Systems in a Distribution Network with Distribution Generators

Authors: Lakshya Bhat, Anubhav Shrivastava, Shivarudraswamy

Abstract:

There has been a formidable interest in the area of Distributed Generation in recent times. A wide number of loads are addressed by Distributed Generators and have better efficiency too. The major disadvantage in Distributed Generation is voltage control- is highlighted in this paper. The paper addresses voltage control at buses in IEEE 14 Bus system by regulating reactive power. An analysis is carried out by selecting the most optimum location in placing the Distributed Generators through load flow analysis and seeing where the voltage profile rises. Matlab programming is used for simulation of voltage profile in the respective buses after introduction of DG’s. A tolerance limit of +/-5% of the base value has to be maintained.To maintain the tolerance limit , 3 methods are used. Sensitivity analysis of 3 methods for voltage control is carried out to determine the priority among the methods.

Keywords: distributed generators, distributed system, reactive power, voltage control, sensitivity analysis

Procedia PDF Downloads 556
30385 Establishment of the Regression Uncertainty of the Critical Heat Flux Power Correlation for an Advanced Fuel Bundle

Authors: L. Q. Yuan, J. Yang, A. Siddiqui

Abstract:

A new regression uncertainty analysis methodology was applied to determine the uncertainties of the critical heat flux (CHF) power correlation for an advanced 43-element bundle design, which was developed by Canadian Nuclear Laboratories (CNL) to achieve improved economics, resource utilization and energy sustainability. The new methodology is considered more appropriate than the traditional methodology in the assessment of the experimental uncertainty associated with regressions. The methodology was first assessed using both the Monte Carlo Method (MCM) and the Taylor Series Method (TSM) for a simple linear regression model, and then extended successfully to a non-linear CHF power regression model (CHF power as a function of inlet temperature, outlet pressure and mass flow rate). The regression uncertainty assessed by MCM agrees well with that by TSM. An equation to evaluate the CHF power regression uncertainty was developed and expressed as a function of independent variables that determine the CHF power.

Keywords: CHF experiment, CHF correlation, regression uncertainty, Monte Carlo Method, Taylor Series Method

Procedia PDF Downloads 387
30384 Time Series Forecasting (TSF) Using Various Deep Learning Models

Authors: Jimeng Shi, Mahek Jain, Giri Narasimhan

Abstract:

Time Series Forecasting (TSF) is used to predict the target variables at a future time point based on the learning from previous time points. To keep the problem tractable, learning methods use data from a fixed-length window in the past as an explicit input. In this paper, we study how the performance of predictive models changes as a function of different look-back window sizes and different amounts of time to predict the future. We also consider the performance of the recent attention-based Transformer models, which have had good success in the image processing and natural language processing domains. In all, we compare four different deep learning methods (RNN, LSTM, GRU, and Transformer) along with a baseline method. The dataset (hourly) we used is the Beijing Air Quality Dataset from the UCI website, which includes a multivariate time series of many factors measured on an hourly basis for a period of 5 years (2010-14). For each model, we also report on the relationship between the performance and the look-back window sizes and the number of predicted time points into the future. Our experiments suggest that Transformer models have the best performance with the lowest Mean Average Errors (MAE = 14.599, 23.273) and Root Mean Square Errors (RSME = 23.573, 38.131) for most of our single-step and multi-steps predictions. The best size for the look-back window to predict 1 hour into the future appears to be one day, while 2 or 4 days perform the best to predict 3 hours into the future.

Keywords: air quality prediction, deep learning algorithms, time series forecasting, look-back window

Procedia PDF Downloads 124
30383 Relationship among the Air Pollution and Atopic Dermatitis Using Meta-Analysis

Authors: Chaebong Kim, Yongmin Cho, Minkyung Han, Mooyoung Kim, KooSang Kim

Abstract:

Background: Air pollution from global warming has a considerable influence on respiratory disease and atopic dermatitis (AD). Present studies base on a hypothesis about correlation between air pollutant and AD, and the results are analyzed from various points of view. Objectives: This study aimed to integrate the relevant researches for air pollutant and AD, and to perform the systematic literature review and meta-analysis to provide the basis of air pollutant control. Methods: Research materials were collected from original articles published in English academic journals including medicine, nursing and health science from August 1 to 31, 2016. We collected the materials from Pubmed, Medline, Embase, Cochrane Central database with Prisma (Preferred Reporting Items for Systematic Reviews and Meta-Analysis) based on the Cochrane Systematic Review Manual, and performed the evaluation and analysis for selected materials. We got the research results for risk of bias using Rev-Man ver. 5.2, and meta analyses using STATA. Results: The prevalence of infantile atopic dermatitis were 1.05 times higher than other groups who were exposed to air pollution, and exposure to NO2 (1.08, 95% CI: 1.02 – 1.14), O3 (1.09, 95% CI: 1.04 – 1.15), SO2 (1.07, 95% CI: 1.02 – 1.12) in subgroup air pollutant was considerably associated with infantile atopic dermatitis. The prevalence of infantile atopic dermatitis was 1.03 times higher than other groups who were exposed to PM2.5, but the results were not statistically similar. Conclusion: Health effect from environmental pollution risen people’s interest in environmental diseases. Air pollutant was associated with AD in this study, but selected literature was based on non-RCT (Randomized Controlled Trial) study. Therefore, there was a limit in study method including control, matching, and correction of confounding variables. For clear conclusion, it is necessary to develop the appropriate tool for object of study and clear standard to measure of air pollutant.

Keywords: air pollution, atopic dermatitis, children, meta-analysis

Procedia PDF Downloads 219
30382 Quantifying Automation in the Architectural Design Process via a Framework Based on Task Breakdown Systems and Recursive Analysis: An Exploratory Study

Authors: D. M. Samartsev, A. G. Copping

Abstract:

As with all industries, architects are using increasing amounts of automation within practice, with approaches such as generative design and use of AI becoming more commonplace. However, the discourse on the rate at which the architectural design process is being automated is often personal and lacking in objective figures and measurements. This results in confusion between people and barriers to effective discourse on the subject, in turn limiting the ability of architects, policy makers, and members of the public in making informed decisions in the area of design automation. This paper proposes the use of a framework to quantify the progress of automation within the design process. The use of a reductionist analysis of the design process allows it to be quantified in a manner that enables direct comparison across different times, as well as locations and projects. The methodology is informed by the design of this framework – taking on the aspects of a systematic review but compressed in time to allow for an initial set of data to verify the validity of the framework. The use of such a framework of quantification enables various practical uses such as predicting the future of the architectural industry with regards to which tasks will be automated, as well as making more informed decisions on the subject of automation on multiple levels ranging from individual decisions to policy making from governing bodies such as the RIBA. This is achieved by analyzing the design process as a generic task that needs to be performed, then using principles of work breakdown systems to split the task of designing an entire building into smaller tasks, which can then be recursively split further as required. Each task is then assigned a series of milestones that allow for the objective analysis of its automation progress. By combining these two approaches it is possible to create a data structure that describes how much various parts of the architectural design process are automated. The data gathered in the paper serves the dual purposes of providing the framework with validation, as well as giving insights into the current situation of automation within the architectural design process. The framework can be interrogated in many ways and preliminary analysis shows that almost 40% of the architectural design process has been automated in some practical fashion at the time of writing, with the rate at which progress is made slowly increasing over the years, with the majority of tasks in the design process reaching a new milestone in automation in less than 6 years. Additionally, a further 15% of the design process is currently being automated in some way, with various products in development but not yet released to the industry. Lastly, various limitations of the framework are examined in this paper as well as further areas of study.

Keywords: analysis, architecture, automation, design process, technology

Procedia PDF Downloads 73
30381 Reconstructing the Segmental System of Proto-Graeco-Phrygian: a Bottom-Up Approach

Authors: Aljoša Šorgo

Abstract:

Recent scholarship on Phrygian has begun to more closely examine the long-held belief that Greek and Phrygian are two very closely related languages. It is now clear that Graeco-Phrygian can be firmly postulated as a subclade of the Indo-European languages. The present paper will focus on the reconstruction of the phonological and phonetic segments of Proto-Graeco-Phrygian (= PGPh.) by providing relevant correspondence sets and reconstructing the classes of segments. The PGPh. basic vowel system consisted of ten phonemic oral vowels: */a e o ā ē ī ō ū/. The correspondences of the vowels are clear and leave little open to ambiguity. There were four resonants and two semi-vowels in PGPh.: */r l m n i̯ u̯/, which could appear in both a consonantal and a syllabic function, with the distribution between the two still being phonotactically predictable. Of note is the fact that the segments *m and *n seem to have merged when their phonotactic position would see them used in a syllabic function. Whether the segment resulting from this merger was a nasalized vowel (most likely *[ã]) or a syllabic nasal *[N̥] (underspecified for place of articulation) cannot be determined at this stage. There were three fricatives in PGPh.: */s h ç/. *s and *h are easily identifiable. The existence of *ç, which may seem unexpected, is postulated on the basis of the correspondence Gr. ὄς ~ Phr. yos/ιος. It is of note that Bozzone has previously proposed the existence of *ç ( < PIE *h₁i̯-) in an early stage of Greek even without taking into account Phrygian data. Finally, the system of stops in PGPh. distinguished four places of articulation (labial, dental, velar, and labiovelar) and three phonation types. The question of which three phonation types were actually present in PGPh. is one of great importance for the ongoing debate on the realization of the three series in PIE. Since the matter is still very much in dispute, we ought to, at this stage, endeavour to reconstruct the PGPh. system without recourse to the other IE languages. The three series of correspondences are: 1. Gr. T (= tenuis) ~ Phr. T; 2. Gr. D (= media) ~ Phr. T; 3. Gr. TA (= tenuis aspirata) ~ Phr. M. The first series must clearly be reconstructed as composed of voiceless stops. The second and third series are more problematic. With a bottom-up approach, neither the second nor the third series of correspondences are compatible with simple modal voicing, and the reflexes differ greatly in voice onset time. Rather, the defining feature distinguishing the two series was [±spread glottis], with ancillary vibration of the vocal cords. In PGPh. the second series was undergoing further spreading of the glottis. As the two languages split, this process would continue, but be affected by dissimilar changes in VOT, which was ultimately phonemicized in both languages as the defining feature distinguishing between their series of stops.

Keywords: bottom-up reconstruction, Proto-Graeco-Phrygian, spread glottis, syllabic resonant

Procedia PDF Downloads 10
30380 Biomolecular Interaction of Ruthenium(II) Polypyridyl Complexes

Authors: S. N. Harun, H. Ahmad

Abstract:

A series of ruthenium(II) complexes, including two novel compounds [Ru(dppz)2(L)]2+ where dppz = dipyrido-[3,2-a:2’,3’-c]phenazine, and L = 2-phenylimidazo[4,5-f][1,10]phenanthroline (PIP) or 2-(4-hydroxyphenyl)imidazo[4,5-f][1,10]phenanthroline (p-HPIP) have been synthesized and characterized. The previously reported complexes [Ru(bpy)2L]2+ and [Ru(phen)2L]2+ were also prepared. All complexes were characterized by elemental analysis, 1H-NMR spectroscopy, ESI-Mass spectroscopy and FT-IR spectroscopy. The photophysical properties were analyzed by UV-Visible spectroscopy and fluorescence spectroscopy. [Ru(dppz)2(PIP)]2+ and [Ru(dppz)2(p-HPIP)]2+ displayed ‘molecular light-switch’ effect as they have high emission in acetonitrile but no emission in water. The cytotoxicity of all complexes against cancer cell lines Hela and MCF-7 were investigated through standard MTT assay. [Ru(dppz)2(PIP)]2+ showed moderate toxicity on both MCF-7 and Hela with IC50 of 37.64 µM and 28.02 µM, respectively. Interestingly, [Ru(dppz)2(p-HPIP)]2+ exhibited remarkable cytotoxicity results with IC50 of 13.52 µM on Hela and 11.63 µM on MCF-7 cell lines which are comparable to the infamous anti-cancer drug, cisplatin. The cytotoxicity of this complex series increased as the ligands size extended in order of [Ru(bpy)2(L)]2+ < [Ru(phen)2(L)]2+ < [Ru(dppz)2(L)]2+.

Keywords: ruthenium, cytotoxicity, molecular light-switch, anticancer

Procedia PDF Downloads 271
30379 Using Discrete Event Simulation Approach to Reduce Waiting Times in Computed Tomography Radiology Department

Authors: Mwafak Shakoor

Abstract:

The purpose of this study was to reduce patient waiting times, improve system throughput and improve resources utilization in radiology department. A discrete event simulation model was developed using Arena simulation software to investigate different alternatives to improve the overall system delivery based on adding resource scenarios due to the linkage between patient waiting times and resource availability. The study revealed that there is no addition investment need to procure additional scanner but hospital management deploy managerial tactics to enhance machine utilization and reduce the long waiting time in the department.

Keywords: discrete event simulation, radiology department, arena, waiting time, healthcare modeling, computed tomography

Procedia PDF Downloads 552
30378 Impact of Varying Malting and Fermentation Durations on Specific Chemical, Functional Properties, and Microstructural Behaviour of Pearl Millet and Sorghum Flour Using Response Surface Methodology

Authors: G. Olamiti; TK. Takalani; D. Beswa, AIO Jideani

Abstract:

The study investigated the effects of malting and fermentation times on some chemical, functional properties and microstructural behaviour of Agrigreen, Babala pearl millet cultivars and sorghum flours using response surface methodology (RSM). Central Composite Rotatable Design (CCRD) was performed on two independent variables: malting and fermentation times (h), at intervals of 24, 48, and 72, respectively. The results of dependent parameters such as pH, titratable acidity (TTA), Water absorption capacity (WAC), Oil absorption capacity (OAC), bulk density (BD), dispersibility and microstructural behaviour of the flours studied showed a significant difference in p < 0.05 upon malting and fermentation time. Babala flour exhibited a higher pH value at 4.78 at 48 h malted and 81.9 fermentation times. Agrigreen flour showed a higher TTA value at 0.159% at 81.94 h malted and 48 h fermentation times. WAC content was also higher in malted and fermented Babala flour at 2.37 ml g-1 for 81.94 h malted and 48 h fermentation time. Sorghum flour exhibited the least OAC content at 1.67 ml g-1 at 14 h malted and 48 h fermentation times. Agrigreen flour recorded the least bulk density, at 0.53 g ml-1 for 72 h malted and 24 h fermentation time. Sorghum flour exhibited a higher content of dispersibility, at 56.34%, after 24 h malted and 72 h fermented time. The response surface plots showed that increased malting and fermentation time influenced the dependent parameters. The microstructure behaviour of malting and fermentation times of pearl millet varieties and sorghum flours showed isolated, oval, spherical, or polygonal to smooth surfaces. The optimal processing conditions, such as malting and fermentation time for Agrigreen, were 32.24 h and 63.32 h; 35.18 h and 34.58 h for Babala; and 36.75 h and 47.88 h for sorghum with high desirability of 1.00. The validation of the optimum processing malting and fermentation times (h) on the dependent improved the experimented values. Food processing companies can use the study's findings to improve food processing and quality.

Keywords: Pearl millet, malting, fermentation, microstructural behaviour

Procedia PDF Downloads 38
30377 The Relationships between Carbon Dioxide (CO2) Emissions, Energy Consumption and GDP for Iran: Time Series Analysis, 1980-2010

Authors: Jinhoa Lee

Abstract:

The relationships between environmental quality, energy use and economic output have created growing attention over the past decades among researchers and policy makers. Focusing on the empirical aspects of the role of carbon dioxide (CO2) emissions and energy use in affecting the economic output, this paper is an effort to fulfill the gap in a comprehensive case study at a country level using modern econometric techniques. To achieve the goal, this country-specific study examines the short-run and long-run relationships among energy consumption (using disaggregated energy sources: Crude oil, coal, natural gas, and electricity), CO2 emissions and gross domestic product (GDP) for Iran using time series analysis from the year 1980-2010. To investigate the relationships between the variables, this paper employs the Augmented Dickey-Fuller (ADF) test for stationarity, Johansen’s maximum likelihood method for cointegration and a Vector Error Correction Model (VECM) for both short- and long-run causality among the research variables for the sample. All the variables in this study show very strong significant effects on GDP in the country for the long term. The long-run equilibrium in VECM suggests that all energy consumption variables in this study have significant impacts on GDP in the long term. The consumption of petroleum products and the direct combustion of crude oil and natural gas decrease GDP, while the coal and electricity use enhanced the GDP between 1980-2010 in Iran. In the short term, only electricity use enhances the GDP as well as its long-run effects. All variables of this study, except the CO2 emissions, show significant effects on the GDP in the country for the long term. The long-run equilibrium in VECM suggests that the consumption of petroleum products and the direct combustion of crude oil and natural gas use have positive impacts on the GDP while the consumptions of electricity and coal have adverse impacts on the GDP in the long term. In the short run, electricity use enhances the GDP over period of 1980-2010 in Iran. Overall, the results partly support arguments that there are relationships between energy use and economic output, but the associations can be differed by the sources of energy in the case of Iran over period of 1980-2010. However, there is no significant relationship between the CO2 emissions and the GDP and between the CO2 emissions and the energy use both in the short term and long term.

Keywords: CO2 emissions, energy consumption, GDP, Iran, time series analysis

Procedia PDF Downloads 558
30376 Empirical Roughness Progression Models of Heavy Duty Rural Pavements

Authors: Nahla H. Alaswadko, Rayya A. Hassan, Bayar N. Mohammed

Abstract:

Empirical deterministic models have been developed to predict roughness progression of heavy duty spray sealed pavements for a dataset representing rural arterial roads. The dataset provides a good representation of the relevant network and covers a wide range of operating and environmental conditions. A sample with a large size of historical time series data for many pavement sections has been collected and prepared for use in multilevel regression analysis. The modelling parameters include road roughness as performance parameter and traffic loading, time, initial pavement strength, reactivity level of subgrade soil, climate condition, and condition of drainage system as predictor parameters. The purpose of this paper is to report the approaches adopted for models development and validation. The study presents multilevel models that can account for the correlation among time series data of the same section and to capture the effect of unobserved variables. Study results show that the models fit the data very well. The contribution and significance of relevant influencing factors in predicting roughness progression are presented and explained. The paper concludes that the analysis approach used for developing the models confirmed their accuracy and reliability by well-fitting to the validation data.

Keywords: roughness progression, empirical model, pavement performance, heavy duty pavement

Procedia PDF Downloads 132
30375 Migration and Provision of Support to Left-Behind Parents in Rural Cambodia

Authors: Benjamas Penboon, Zachary Zimmer, Aree Jampaklay

Abstract:

Cambodia is a country where labor migration has been consistently high. Coupled with advancing labor opportunities in urban areas, a function partly of globalization, this is resulting in massive migration out of rural areas. This is particularly true in Cambodia where there are high migration and a very large proportion of adult children living some distant from their parents. This paper explores characteristics associated with migrant providing support to parents in rural Cambodia. With reference to perspectives of family altruism and solidarity, this analysis particularly focusses on how a series of variables representing family integration and residential location associates with intergenerational monetary and instrumental support from migrants. The study hypothesizes that migrants are more likely to provide support when parents are in need, and there are no alternative means of support. Data come from The Rural Household Survey (N=3,713), part of the 2011 Cambodian Rural Urban Migration Project (CRUMP). Multilevel multinomial models indicate international migrants are likely to give money, while internal migrants are likely to provide both money and instrumental support, especially when migrants have no sibling and their parent in poor health status. In addition, employed migrants are two times providing monetary compared to those unemployed. Findings elucidate the decision to which and why support occurs more often when no other source of support exists and also depends on the ability to provide of migrants themselves.

Keywords: migration, left-behind parent, intergenerational relations, support, rural, Cambodia

Procedia PDF Downloads 139
30374 On a Single Server Queue with Arrivals in Batches of Variable Size, Generalized Coxian-2 Service and Compulsory Server Vacations

Authors: Kailash C. Madan

Abstract:

We study the steady state behaviour of a batch arrival single server queue in which the first service with general service times is compulsory and the second service with general service times is optional. We term such a two phase service as generalized Coxian-2 service. Just after completion of a service the server must take a vacation of random length of time with general vacation times. We obtain steady state probability generating functions for the queue size as well as the steady state mean queue size at a random epoch of time in explicit and closed forms. Some particular cases of interest including some known results have been derived.

Keywords: batch arrivals, compound Poisson process, generalized Coxian-2 service, steady state

Procedia PDF Downloads 426
30373 A Case Study of Business Analytic Use in European Football: Analysis and Implications

Authors: M. C. Schloesser

Abstract:

The purpose of this paper is to explore the use and impact of business analytics in European football. Despite good evidence from other major sports leagues, research on this topic in Europe is currently very scarce. This research relies on expert interviews on the use and objective of business analytics. Along with revenue data over 16 seasons spanning from 2004/05 to 2019/20 from Manchester City FC, we conducted a time series analysis to detect a structural breakpoint on the different revenue streams, i.e., sponsorship and ticketing, after analytical tools have been implemented. We not only find that business analytics have indeed been applied at Manchester City FC and revenue increase is the main objective of their utilization but also that business analytics is indeed a good means to increase revenues if applied sufficiently. We can thereby support findings from other sports leagues. Consequently, professional sports organizations are advised to apply business analytics if they aim to increase revenues. This research has shown that analytical practices do, in fact, support revenue growth and help to work more efficiently. As the knowledge of analytical practices is very confidential and not publicly available, we had to select one club as a case study which can be considered a research limitation. Other practitioners should explore other clubs or leagues. Further, there are other factors that can lead to increased revenues that need to be considered. Additionally, sports organizations need resources to be able to apply and utilize business analytics. Consequently, findings might only apply to the top teams of the European football leagues. Nonetheless, this paper combines insights and results on usage, objectives, and impact of business analytics in European professional football and thereby fills a current research gap.

Keywords: business analytics, expert interviews, revenue management, time series analysis

Procedia PDF Downloads 44
30372 Comparison of Different Machine Learning Models for Time-Series Based Load Forecasting of Electric Vehicle Charging Stations

Authors: H. J. Joshi, Satyajeet Patil, Parth Dandavate, Mihir Kulkarni, Harshita Agrawal

Abstract:

As the world looks towards a sustainable future, electric vehicles have become increasingly popular. Millions worldwide are looking to switch to Electric cars over the previously favored combustion engine-powered cars. This demand has seen an increase in Electric Vehicle Charging Stations. The big challenge is that the randomness of electrical energy makes it tough for these charging stations to provide an adequate amount of energy over a specific amount of time. Thus, it has become increasingly crucial to model these patterns and forecast the energy needs of power stations. This paper aims to analyze how different machine learning models perform on Electric Vehicle charging time-series data. The data set consists of authentic Electric Vehicle Data from the Netherlands. It has an overview of ten thousand transactions from public stations operated by EVnetNL.

Keywords: forecasting, smart grid, electric vehicle load forecasting, machine learning, time series forecasting

Procedia PDF Downloads 72
30371 Effects of Feeding Time on Survival Rates, Growth Performance and Feeding Behavior of Juvenile Catfish

Authors: Abdullahi Ibrahim

Abstract:

The culture of Clarias gariepinus for fish production is becoming increasingly essential as the fish is contributing to the food abundance and nutritional benefit to family health, income generation, and employment opportunities. The effect of feeding frequency was investigated over a period of ten (10) weeks; the experiment was conducted to monitor survival rates, growth performance, and feeding behavior of juvenile catfish. The experimental fish were randomly assigned to five treatment groups; (i.e., with different feeding frequency intervals) of 100 fish each. Each treatment was replicated twice with 50 fish per replicate. All the groups were fed with floating fish feed (blue crown®). The five treatments (feeding frequency) were T1- once a day feeding of night hours only, T2- twice a day feeding time of morning and night hours, T3- trice a day feeding time of morning, evening and night hours, T-4 four times a day feeding of morning, afternoon, evening, and night hours, T-5 five times a day feeding at four hours interval. There were significant differences (p > 0.05) among treatments. Feed intake and weight gain improved significantly (p < 0.05) in T-4 and T-3. The best of the feeding time on weight gain, survival rate, and feed conversion ratio were obtained at three times a day feeding (T-3) compared to other treatments, especially those fed once and five times feeding a regiment. This might be attributed to the high level of dissolve oxygen and less stress. Feeding fish three times a day is therefore recommended for efficient catfish production to maximize profits as the feed represents more than 50% of aquaculture inputs, particularly in intensive farming systems.

Keywords: catfish, floating fish feed, dissolve oxygen, juvenile

Procedia PDF Downloads 114
30370 Algorithms Minimizing Total Tardiness

Authors: Harun Aydilek, Asiye Aydilek, Ali Allahverdi

Abstract:

The total tardiness is a widely used performance measure in the scheduling literature. This performance measure is particularly important in situations where there is a cost to complete a job beyond its due date. The cost of scheduling increases as the gap between a job's due date and its completion time increases. Such costs may also be penalty costs in contracts, loss of goodwill. This performance measure is important as the fulfillment of due dates of customers has to be taken into account while making scheduling decisions. The problem is addressed in the literature, however, it has been assumed zero setup times. Even though this assumption may be valid for some environments, it is not valid for some other scheduling environments. When setup times are treated as separate from processing times, it is possible to increase machine utilization and to reduce total tardiness. Therefore, non-zero setup times need to be considered as separate. A dominance relation is developed and several algorithms are proposed. The developed dominance relation is utilized in the proposed algorithms. Extensive computational experiments are conducted for the evaluation of the algorithms. The experiments indicated that the developed algorithms perform much better than the existing algorithms in the literature. More specifically, one of the newly proposed algorithms reduces the error of the best existing algorithm in the literature by 40 percent.

Keywords: algorithm, assembly flowshop, dominance relation, total tardiness

Procedia PDF Downloads 325
30369 Rainfall Analysis in the Contest of Climate Change for Jeddah Area, Western Saudi Arabia

Authors: Ali M. Subyani

Abstract:

The increase in the greenhouse gas emission has had a severe impact on global climate change and is bound to affect the weather patterns worldwide. This climate change impacts are among the future significant effects on any society. Rainfall levels are drastically increasing with flash floods in some places and long periods of droughts in others, especially in arid regions. These extreme events are causes of interactions concerning environmental, socio-economic and cultural life and their implementation. This paper presents the detailed features of dry and wet spell durations and rainfall intensity series available (1971-2012) on daily basis for the Jeddah area, Western, Saudi Arabia. It also presents significant articles for combating the climate change impacts on this area. Results show trend changes in dry and wet spell durations and rainfall amount on daily, monthly and annual time series. Three rain seasons were proposed in this investigation: high rain, low rain, and dry seasons. It shows that the overall average dry spell durations is about 80 continuous days while the average wet spell durations is 1.39 days with an average rainfall intensity of 8.2 mm/day. Annual and seasonal autorun analyses confirm that the rainy seasons are tending to have more intense rainfall while the seasons are becoming drier. This study would help decision makers in future for water resources management and flood risk analysis.

Keywords: climate change, daily rainfall, dry and wet spill, Jeddah, Saudi Arabia

Procedia PDF Downloads 302
30368 Economic Analysis of Rainwater Harvesting Systems for Dairy Cattle

Authors: Sandra Cecilia Muhirirwe, Bart Van Der Bruggen, Violet Kisakye

Abstract:

Economic analysis of Rainwater harvesting (RWH) systems is vital in search of a cost-effective solution to water unreliability, especially in low-income countries. There is little literature focusing on the financial aspects of RWH for dairy farmers. The main purpose was to assess the economic viability of rainwater harvesting for diary framers in the Rwenzori region. The study focused on the use of rainwater harvesting systems from the rooftop and collection in above surface tanks. Daily rainfall time series for 12 years was obtained across nine gauging stations. The daily water balance equation was used for optimal sizing of the tank. Economic analysis of the investment was carried out based on the life cycle costs and the accruing benefits for the period of 15 years. Roof areas were varied from 75m2 as the minimum required area to 500m2 while maintaining the same number of cattle and keeping the daily water demand constant. The results show that the required rainwater tank sizes are very large and may be impractical to install due to the strongly varying terrain and the initial cost of investment. In all districts, there is a significant reduction of the volume of the required tank with an increasing collection area. The results further show that increasing the collection area has a minor effect on reducing the required tank size. Generally, for all rainfall areas, the reliability increases with an increase in the roof area. The results indicate that 100% reliability can only be realized with very large collection areas that are impractical to install. The estimated benefits outweigh the cost of investment. The Present Net Value shows that the investment is economically viable and investment with a short payback of a maximum of 3 years for all the time series in the study area.

Keywords: dairy cattle, optimisation, rainwater harvesting, economic analysis

Procedia PDF Downloads 168
30367 Performance Analysis of Transformerless DC-DC Boost Converter

Authors: Nidhi Vijay, A. K. Sharma

Abstract:

Many industrial applications require power from dc source. DC-DC boost converters are now being used all over the world for rapid transit system. Although these provide high efficiency, smooth control, fast response and regeneration, conventional DC-DC boost converters are unable to provide high step up voltage gain due to effect of power switches, rectifier diodes and equivalent series resistance of inductor and capacitor. This paper proposes new transformerless dc-dc converters to achieve high step up voltage gain as compared to the conventional converter without an extremely high duty ratio. Only one power stage is used in this converter. Steady-state analysis of voltage gain is discussed in brief. Finally, a comparative analysis is given in order to verify the results.

Keywords: MATLAB, DC-DC boost converter, voltage gain, voltage stress

Procedia PDF Downloads 399
30366 Finite Element Analysis of Hollow Structural Shape (HSS) Steel Brace with Infill Reinforcement under Cyclic Loading

Authors: Chui-Hsin Chen, Yu-Ting Chen

Abstract:

Special concentrically braced frames is one of the seismic load resisting systems, which dissipates seismic energy when bracing members within the frames undergo yielding and buckling while sustaining their axial tension and compression load capacities. Most of the inelastic deformation of a buckling bracing member concentrates in the mid-length region. While experiencing cyclic loading, the region dissipates most of the seismic energy being input into the frame. Such a concentration makes the braces vulnerable to failure modes associated with low-cycle fatigue. In this research, a strategy to improve the cyclic behavior of the conventional steel bracing member is proposed by filling the Hollow Structural Shape (HSS) member with reinforcement. It prevents the local section from concentrating large plastic deformation caused by cyclic loading. The infill helps spread over the plastic hinge region into a wider area hence postpone the initiation of local buckling or even the rupture of the braces. The finite element method is introduced to simulate the complicated bracing member behavior and member-versus-infill interaction under cyclic loading. Fifteen 3-D-element-based models are built by ABAQUS software. The verification of the FEM model is done with unreinforced (UR) HSS bracing members’ cyclic test data and aluminum honeycomb plates’ bending test data. Numerical models include UR and filled HSS bracing members with various compactness ratios based on the specification of AISC-2016 and AISC-1989. The primary variables to be investigated include the relative bending stiffness and the material of the filling reinforcement. The distributions of von Mises stress and equivalent plastic strain (PEEQ) are used as indices to tell the strengths and shortcomings of each model. The result indicates that the change of relative bending stiffness of the infill is much more influential than the change of material in use to increase the energy dissipation capacity. Strengthen the relative bending stiffness of the reinforcement results in additional energy dissipation capacity to the extent of 24% and 46% in model based on AISC-2016 (16-series) and AISC-1989 (89-series), respectively. HSS members with infill show growth in 𝜂Local Buckling, normalized energy cumulated until the happening of local buckling, comparing to UR bracing members. The 89-series infill-reinforced members have more energy dissipation capacity than unreinforced 16-series members by 117% to 166%. The flexural rigidity of infills should be less than 29% and 13% of the member section itself for 16-series and 89-series bracing members accordingly, thereby guaranteeing the spread over of the plastic hinge and the happening of it within the reinforced section. If the parameters are properly configured, the ductility, energy dissipation capacity, and fatigue-life of HSS SCBF bracing members can be improved prominently by the infill-reinforced method.

Keywords: special concentrically braced frames, HSS, cyclic loading, infill reinforcement, finite element analysis, PEEQ

Procedia PDF Downloads 72
30365 Method for Improving Antidepressants Adherence in Patients with Depressive Disorder: Systemic Review and Meta-Analysis

Authors: Juntip Kanjanasilp, Ratree Sawangjit, Kanokporn Meelap, Kwanchanok Kruthakool

Abstract:

Depression is a common mental health disorder. Antidepressants are effective pharmacological treatments, but most patients have low medication adherence. This study aims to systematic review and meta-analysis what method increase the antidepressants adherence efficiently and improve clinical outcome. Systematic review of articles of randomized controlled trials obtained by a computerized literature search of The Cochrane, Library, Pubmed, Embase, PsycINFO, CINAHL, Education search, Web of Science and ThaiLIS (28 December 2017). Twenty-three studies were included and assessed the quality of research by ROB 2.0. The results reported that printing media improved in number of people who had medication adherence statistical significantly (p= 0.018), but education, phone call, and program utilization were no different (p=0.172, p=0.127, p=0.659). There was no significant difference in pharmacist’s group, health care team’s group and physician’s group (p=0.329, p=0.070, p=0.040). Times of intervention at 1 month and 6 months improved medication adherence significantly (p= 0.0001, p=0.013). There was significantly improved adherence in single intervention (p=0.027) but no different in multiple interventions (p=0.154). When we analyzed medication adherence with the mean score, no improved adherence was found, not relevant with who gives the intervention and times to intervention. However, the multiple interventions group was statistically significant improved medication adherence (p=0.040). Phone call and the physician’s group were statistically significant improved clinical outcomes in number of improved patients (0.025 and 0.020, respectively). But in the pharmacist’s group and physician’s group were not found difference in the mean score of clinical outcomes (p=0.993, p=0.120, respectively). Times to intervention and number of intervention were not significant difference than usual care. The overall intervention can increase antidepressant adherence, especially the printing media, and the appropriate timing of the intervention is at least 6 months. For effective treatment, the provider should have experience and expert in caring for patients with depressive disorders, such as a psychiatrist. Medical personnel should have knowledge in caring for these patients also.

Keywords: depression, medication adherence, clinical outcomes, systematic review, meta-analysis

Procedia PDF Downloads 108
30364 Exploring Reading into Writing: A Corpus-Based Analysis of Postgraduate Students’ Literature Review Essays

Authors: Tanzeela Anbreen, Ammara Maqsood

Abstract:

Reading into writing is one of university students' most required academic skills. The current study explored postgraduate university students’ writing quality using a corpus-based approach. Twelve postgraduate students’ literature review essays were chosen for the corpus-based analysis. These essays were chosen because students had to incorporate multiple reading sources in these essays, which was a new writing exercise for them. The students were provided feedback at least two times which comprised of the written comments by the tutor highlighting the areas of improvement and also by using the ‘track changes’ function. This exercise was repeated two times, and students submitted two drafts. This investigation included only the finally submitted work of the students. A corpus-based approach was adopted to analyse the essays because it promotes autonomous discovery and personalised learning. The aim of this analysis was to understand the existing level of students’ writing before the start of their postgraduate thesis. Text Inspector was used to analyse the quality of essays. With the help of the Text Inspector tool, the vocabulary used in the essays was compared to the English Vocabulary Profile (EVP), which describes what learners know and can do at each Common European Framework of Reference (CEFR) level. Writing quality was also measured for the Flesch reading ease score, which is a standard to describe the ease of understanding the writing content. The results reflected that students found writing essays using multiple sources challenging. In most essays, the vocabulary level achieved was between B1-B2 of the CEFL level. The study recommends that students need extensive training in developing academic writing skills, particularly in writing the literature review type assignment, which requires multiple sources citations.

Keywords: literature review essays, postgraduate students, corpus-based analysis, vocabulary proficiency

Procedia PDF Downloads 32
30363 Comparison of Two Maintenance Policies for a Two-Unit Series System Considering General Repair

Authors: Seyedvahid Najafi, Viliam Makis

Abstract:

In recent years, maintenance optimization has attracted special attention due to the growth of industrial systems complexity. Maintenance costs are high for many systems, and preventive maintenance is effective when it increases operations' reliability and safety at a reduced cost. The novelty of this research is to consider general repair in the modeling of multi-unit series systems and solve the maintenance problem for such systems using the semi-Markov decision process (SMDP) framework. We propose an opportunistic maintenance policy for a series system composed of two main units. Unit 1, which is more expensive than unit 2, is subjected to condition monitoring, and its deterioration is modeled using a gamma process. Unit 1 hazard rate is estimated by the proportional hazards model (PHM), and two hazard rate control limits are considered as the thresholds of maintenance interventions for unit 1. Maintenance is performed on unit 2, considering an age control limit. The objective is to find the optimal control limits and minimize the long-run expected average cost per unit time. The proposed algorithm is applied to a numerical example to compare the effectiveness of the proposed policy (policy Ⅰ) with policy Ⅱ, which is similar to policy Ⅰ, but instead of general repair, replacement is performed. Results show that policy Ⅰ leads to lower average cost compared with policy Ⅱ. 

Keywords: condition-based maintenance, proportional hazards model, semi-Markov decision process, two-unit series systems

Procedia PDF Downloads 93