Search results for: RR interval time series
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20160

Search results for: RR interval time series

19950 Optimizing Approach for Sifting Process to Solve a Common Type of Empirical Mode Decomposition Mode Mixing

Authors: Saad Al-Baddai, Karema Al-Subari, Elmar Lang, Bernd Ludwig

Abstract:

Empirical mode decomposition (EMD), a new data-driven of time-series decomposition, has the advantage of supposing that a time series is non-linear or non-stationary, as is implicitly achieved in Fourier decomposition. However, the EMD suffers of mode mixing problem in some cases. The aim of this paper is to present a solution for a common type of signals causing of EMD mode mixing problem, in case a signal suffers of an intermittency. By an artificial example, the solution shows superior performance in terms of cope EMD mode mixing problem comparing with the conventional EMD and Ensemble Empirical Mode decomposition (EEMD). Furthermore, the over-sifting problem is also completely avoided; and computation load is reduced roughly six times compared with EEMD, an ensemble number of 50.

Keywords: empirical mode decomposition (EMD), mode mixing, sifting process, over-sifting

Procedia PDF Downloads 390
19949 Times Series Analysis of Depositing in Industrial Design in Brazil between 1996 and 2013

Authors: Jonas Pedro Fabris, Alberth Almeida Amorim Souza, Maria Emilia Camargo, Suzana Leitão Russo

Abstract:

With the law Nº. 9279, of May 14, 1996, the Brazilian government regulates rights and obligations relating to industrial property considering the economic development of the country as granting patents, trademark registration, registration of industrial designs and other forms of protection copyright. In this study, we show the application of the methodology of Box and Jenkins in the series of deposits of industrial design at the National Institute of Industrial Property for the period from May 1996 to April 2013. First, a graphical analysis of the data was done by observing the behavior of the data and the autocorrelation function. The best model found, based on the analysis of charts and statistical tests suggested by Box and Jenkins methodology, it was possible to determine the model number for the deposit of industrial design, SARIMA (2,1,0)(2,0,0), with an equal to 9.88% MAPE.

Keywords: ARIMA models, autocorrelation, Box and Jenkins Models, industrial design, MAPE, time series

Procedia PDF Downloads 541
19948 Synthetic Daily Flow Duration Curves for the Çoruh River Basin, Turkey

Authors: Ibrahim Can, Fatih Tosunoğlu

Abstract:

The flow duration curve (FDC) is an informative method that represents the flow regime’s properties for a river basin. Therefore, the FDC is widely used for water resource projects such as hydropower, water supply, irrigation and water quality management. The primary purpose of this study is to obtain synthetic daily flow duration curves for Çoruh Basin, Turkey. For this aim, we firstly developed univariate auto-regressive moving average (ARMA) models for daily flows of 9 stations located in Çoruh basin and then these models were used to generate 100 synthetic flow series each having same size as historical series. Secondly, flow duration curves of each synthetic series were drawn and the flow values exceeded 10, 50 and 95 % of the time and 95% confidence limit of these flows were calculated. As a result, flood, mean and low flows potential of Çoruh basin will comprehensively be represented.

Keywords: ARMA models, Çoruh basin, flow duration curve, Turkey

Procedia PDF Downloads 399
19947 Localization of Geospatial Events and Hoax Prediction in the UFO Database

Authors: Harish Krishnamurthy, Anna Lafontant, Ren Yi

Abstract:

Unidentified Flying Objects (UFOs) have been an interesting topic for most enthusiasts and hence people all over the United States report such findings online at the National UFO Report Center (NUFORC). Some of these reports are a hoax and among those that seem legitimate, our task is not to establish that these events confirm that they indeed are events related to flying objects from aliens in outer space. Rather, we intend to identify if the report was a hoax as was identified by the UFO database team with their existing curation criterion. However, the database provides a wealth of information that can be exploited to provide various analyses and insights such as social reporting, identifying real-time spatial events and much more. We perform analysis to localize these time-series geospatial events and correlate with known real-time events. This paper does not confirm any legitimacy of alien activity, but rather attempts to gather information from likely legitimate reports of UFOs by studying the online reports. These events happen in geospatial clusters and also are time-based. We look at cluster density and data visualization to search the space of various cluster realizations to decide best probable clusters that provide us information about the proximity of such activity. A random forest classifier is also presented that is used to identify true events and hoax events, using the best possible features available such as region, week, time-period and duration. Lastly, we show the performance of the scheme on various days and correlate with real-time events where one of the UFO reports strongly correlates to a missile test conducted in the United States.

Keywords: time-series clustering, feature extraction, hoax prediction, geospatial events

Procedia PDF Downloads 372
19946 Reconsidering Taylor’s Law with Chaotic Population Dynamical Systems

Authors: Yuzuru Mitsui, Takashi Ikegami

Abstract:

The exponents of Taylor’s law in deterministic chaotic systems are computed, and their meanings are intensively discussed. Taylor’s law is the scaling relationship between the mean and variance (in both space and time) of population abundance, and this law is known to hold in a variety of ecological time series. The exponents found in the temporal Taylor’s law are different from those of the spatial Taylor’s law. The temporal Taylor’s law is calculated on the time series from the same locations (or the same initial states) of different temporal phases. However, with the spatial Taylor’s law, the mean and variance are calculated from the same temporal phase sampled from different places. Most previous studies were done with stochastic models, but we computed the temporal and spatial Taylor’s law in deterministic systems. The temporal Taylor’s law evaluated using the same initial state, and the spatial Taylor’s law was evaluated using the ensemble average and variance. There were two main discoveries from this work. First, it is often stated that deterministic systems tend to have the value two for Taylor’s exponent. However, most of the calculated exponents here were not two. Second, we investigated the relationships between chaotic features measured by the Lyapunov exponent, the correlation dimension, and other indexes with Taylor’s exponents. No strong correlations were found; however, there is some relationship in the same model, but with different parameter values, and we will discuss the meaning of those results at the end of this paper.

Keywords: chaos, density effect, population dynamics, Taylor’s law

Procedia PDF Downloads 171
19945 Enhancing Sell-In and Sell-Out Forecasting Using Ensemble Machine Learning Method

Authors: Vishal Das, Tianyi Mao, Zhicheng Geng, Carmen Flores, Diego Pelloso, Fang Wang

Abstract:

Accurate sell-in and sell-out forecasting is a ubiquitous problem in the retail industry. It is an important element of any demand planning activity. As a global food and beverage company, Nestlé has hundreds of products in each geographical location that they operate in. Each product has its sell-in and sell-out time series data, which are forecasted on a weekly and monthly scale for demand and financial planning. To address this challenge, Nestlé Chilein collaboration with Amazon Machine Learning Solutions Labhas developed their in-house solution of using machine learning models for forecasting. Similar products are combined together such that there is one model for each product category. In this way, the models learn from a larger set of data, and there are fewer models to maintain. The solution is scalable to all product categories and is developed to be flexible enough to include any new product or eliminate any existing product in a product category based on requirements. We show how we can use the machine learning development environment on Amazon Web Services (AWS) to explore a set of forecasting models and create business intelligence dashboards that can be used with the existing demand planning tools in Nestlé. We explored recent deep learning networks (DNN), which show promising results for a variety of time series forecasting problems. Specifically, we used a DeepAR autoregressive model that can group similar time series together and provide robust predictions. To further enhance the accuracy of the predictions and include domain-specific knowledge, we designed an ensemble approach using DeepAR and XGBoost regression model. As part of the ensemble approach, we interlinked the sell-out and sell-in information to ensure that a future sell-out influences the current sell-in predictions. Our approach outperforms the benchmark statistical models by more than 50%. The machine learning (ML) pipeline implemented in the cloud is currently being extended for other product categories and is getting adopted by other geomarkets.

Keywords: sell-in and sell-out forecasting, demand planning, DeepAR, retail, ensemble machine learning, time-series

Procedia PDF Downloads 264
19944 Temporal Delays along the Neurosurgical Care Continuum for Traumatic Brain Injury Patients in Mulago Hospital in Kampala Uganda

Authors: Silvia D. Vaca, Benjamin J. Kuo, Joao Ricardo N. Vissoci, Catherine A. Staton, Linda W. Xu, Michael Muhumuza, Hussein Ssenyonjo, John Mukasa, Joel Kiryabwire, Henry E. Rice, Gerald A. Grant, Michael M. Haglund

Abstract:

Background: While delays to care exist in resource rich settings, greater delays are seen along the care continuum in low- and middle-income countries (LMICs) largely due to limited healthcare capacity to address the disproportional rates of traumatic brain injury (TBI) in Sub Saharan Africa (SSA). While many LMICs have government subsidized systems to offset surgical costs, the burden of securing funds by the patients for medications, supplies, and CT diagnostics poses a significant challenge to timely surgical interventions. In Kampala Uganda, the challenge of obtaining timely CT scans is twofold. First, due to a lack of a functional CT scanner at the tertiary hospital, patients need to arrange their own transportation to the nearby private facility for CT scans. Second, self-financing for the private CT scans ranges from $80 - $130, which is near the average monthly income in Kampala. These bottlenecks contribute significantly to the care continuum delays and are associated with poor TBI outcomes. Objective: The objectives of this study are to 1) describe the temporal delays through a modified three delays model that fits the context of neurosurgical interventions for TBI patients in Kampala and 2) investigate the association between delays and mortality. Methods: Prospective data were collected for 563 TBI patients presenting to a tertiary hospital in Kampala from 1 June – 30 November 2016. Four time intervals were constructed along five time points: injury, hospital arrival, neurosurgical evaluation, CT results, and definitive surgery. Time interval differences among mild, moderate and severe TBI and their association with mortality were analyzed. Results: The mortality rate of all TBI patients presenting to MNRH was 9.6%, which ranged from 4.7% for mild and moderate TBI patients receiving surgery to 81.8% for severe TBI patients who failed to receive surgery. The duration from injury to surgery varied considerably across TBI severity with the largest gap seen between mild TBI (174 hours) and severe TBI (69 hours) patients. Further analysis revealed care continuum differences for interval 3 (neurosurgical evaluation to CT result) and 4 (CT result to surgery) between severe TBI patients (7 hours for interval 3 and 24 hours for interval 4) and mild TBI patients (19 hours for interval 3, and 96 hours for interval 4). These post-arrival delays were associated with mortality for mild (p=0.05) and moderate TBI (p=0.03) patients. Conclusions: To our knowledge, this is the first analysis using a modified ‘three delays’ framework to analyze the care continuum of TBI patients in Uganda from injury to surgery. We found significant associations between delays and mortality for mild and moderate TBI patients. As it currently stands, poorer outcomes were observed for these mild and moderate TBI patients who were managed non-operatively or failed to receive surgery while surgical services were shunted to more severely ill patients. While well intentioned, high mortality rates were still observed for the severe TBI patients managed surgically. These results suggest the need for future research to optimize triage practices, understand delay contributors, and improve pre-hospital logistical referral systems.

Keywords: care continuum, global neurosurgery, Kampala Uganda, LMIC, Mulago, prospective registry, traumatic brain injury

Procedia PDF Downloads 341
19943 Spatial Analysis of Park and Ride Users’ Dynamic Accessibility to Train Station: A Case Study in Perth

Authors: Ting (Grace) Lin, Jianhong (Cecilia) Xia, Todd Robinson

Abstract:

Accessibility analysis, examining people’s ability to access facilities and destinations, is a fundamental assessment for transport planning, policy making, and social exclusion research. Dynamic accessibility which measures accessibility in real-time traffic environment has been an advanced accessibility indicator in transport research. It is also a useful indicator to help travelers to understand travel time daily variability, assists traffic engineers to monitor traffic congestions, and finally develop effective strategies in order to mitigate traffic congestions. This research involved real-time traffic information by collecting travel time data with 15-minute interval via the TomTom® API. A framework for measuring dynamic accessibility was then developed based on the gravity theory and accessibility dichotomy theory through space and time interpolation. Finally, the dynamic accessibility can be derived at any given time and location under dynamic accessibility spatial analysis framework.

Keywords: dynamic accessibility, hot spot, transport research, TomTom® API

Procedia PDF Downloads 382
19942 Times2D: A Time-Frequency Method for Time Series Forecasting

Authors: Reza Nematirad, Anil Pahwa, Balasubramaniam Natarajan

Abstract:

Time series data consist of successive data points collected over a period of time. Accurate prediction of future values is essential for informed decision-making in several real-world applications, including electricity load demand forecasting, lifetime estimation of industrial machinery, traffic planning, weather prediction, and the stock market. Due to their critical relevance and wide application, there has been considerable interest in time series forecasting in recent years. However, the proliferation of sensors and IoT devices, real-time monitoring systems, and high-frequency trading data introduce significant intricate temporal variations, rapid changes, noise, and non-linearities, making time series forecasting more challenging. Classical methods such as Autoregressive integrated moving average (ARIMA) and Exponential Smoothing aim to extract pre-defined temporal variations, such as trends and seasonality. While these methods are effective for capturing well-defined seasonal patterns and trends, they often struggle with more complex, non-linear patterns present in real-world time series data. In recent years, deep learning has made significant contributions to time series forecasting. Recurrent Neural Networks (RNNs) and their variants, such as Long short-term memory (LSTMs) and Gated Recurrent Units (GRUs), have been widely adopted for modeling sequential data. However, they often suffer from the locality, making it difficult to capture local trends and rapid fluctuations. Convolutional Neural Networks (CNNs), particularly Temporal Convolutional Networks (TCNs), leverage convolutional layers to capture temporal dependencies by applying convolutional filters along the temporal dimension. Despite their advantages, TCNs struggle with capturing relationships between distant time points due to the locality of one-dimensional convolution kernels. Transformers have revolutionized time series forecasting with their powerful attention mechanisms, effectively capturing long-term dependencies and relationships between distant time points. However, the attention mechanism may struggle to discern dependencies directly from scattered time points due to intricate temporal patterns. Lastly, Multi-Layer Perceptrons (MLPs) have also been employed, with models like N-BEATS and LightTS demonstrating success. Despite this, MLPs often face high volatility and computational complexity challenges in long-horizon forecasting. To address intricate temporal variations in time series data, this study introduces Times2D, a novel framework that parallelly integrates 2D spectrogram and derivative heatmap techniques. The spectrogram focuses on the frequency domain, capturing periodicity, while the derivative patterns emphasize the time domain, highlighting sharp fluctuations and turning points. This 2D transformation enables the utilization of powerful computer vision techniques to capture various intricate temporal variations. To evaluate the performance of Times2D, extensive experiments were conducted on standard time series datasets and compared with various state-of-the-art algorithms, including DLinear (2023), TimesNet (2023), Non-stationary Transformer (2022), PatchTST (2023), N-HiTS (2023), Crossformer (2023), MICN (2023), LightTS (2022), FEDformer (2022), FiLM (2022), SCINet (2022a), Autoformer (2021), and Informer (2021) under the same modeling conditions. The initial results demonstrated that Times2D achieves consistent state-of-the-art performance in both short-term and long-term forecasting tasks. Furthermore, the generality of the Times2D framework allows it to be applied to various tasks such as time series imputation, clustering, classification, and anomaly detection, offering potential benefits in any domain that involves sequential data analysis.

Keywords: derivative patterns, spectrogram, time series forecasting, times2D, 2D representation

Procedia PDF Downloads 37
19941 Development of Materials Based on Phosphates of NaZr2(PO4)3 with Low Thermal Expansion

Authors: V. Yu. Volgutov, A. I. Orlova, S. A. Khainakov

Abstract:

NaZr2(PO4)3 (NZP) and their structural analogues are characterized by a peculiar behaviors on heating – they have different expansion and contraction along different crystallographic directions due to specific arrangements of crystal structure in these compounds. An important feature of such structures is the ability to incorporate into their structural analogues wide variety of metal cations having different size and oxidation states, with different combinations and concentrations. These cations are located in different crystallographic non-equivalent positions of octahedral tetrahedral crystal framework as well as in inter-framework cavities. Through, due to iso- and hetero-valent isomorphism of the cations (and the anions) in NZP, it becomes possible to tuning the compositions and to obtain the compounds with ‘on a plan’ properties. For the design of compounds with low and ultra-low thermal expansion including those with tailored thermal expansion properties, the following crystallochemical principles it seems are promising: 1) Insertion into crystal M1 position the cations having different sizes and, 2) the variation in the composition of compounds, providing different occupation of crystal M1 position. Following these principles we have designed and synthesized the next NZP-type phosphates series: a) where radii of the cations in the M1 crystal position was varied: Zr1/4Zr2(PO4)3 - Th1/4Zr2(PO4)3 (series I); R1/3Zr2(PO4)3 where R= Nd, Eu, Er (series II), b) where the occupation of M1 crystal position was varied: Zr1/4Zr2(PO4)3-Er1/3Zr2(PO4)3 (series III) and Zr1/4Zr2(PO4)3-Sr1/2Zr2(PO4)3 (series IV). The thermal expansion parameters were determined over the range of 25-800ºC. For each series the minimum axial coefficient of thermal expansion αa = αb, αc and their anisotropy Δα = Iαa - αcI, 10-6 K-1 was found as next: -1.51, 1.07, 2.58 for Th1/4Zr2(PO4)3 (series I); -0.72, 0.10, 0.81 for Nd1/3Zr2(PO4)3 (series II); -2.78, 1.35, 4.12 for Er1/6Zr1/8Zr2(PO4)3 (series III); 2.23, 1.32, 0.91 for Sr1/2Zr2(PO4)3 (series IV). The measured tendencies of the thermal expansion of crystals were in good agreement with predicted ones. For one of the members from the studied phosphates namely Th1/16Zr3/16Zr2(PO4)3 structural refinement have been carried out at 25, 200, 600, and 800°C. The dependencies of the structural parameters with the temperature have been determined.

Keywords: high-temperature crystallography, NaZr2(PO4)3, (NZP) analogs, structural-chemical principles, tuning thermal expansion

Procedia PDF Downloads 230
19940 Impact of Workers’ Remittances on Poverty in Pakistan: A Time Series Analysis by Ardl

Authors: Syed Aziz Rasool, Ayesha Zaman

Abstract:

Poverty is one of the most important problems for any developing nation. Workers’ remittances and investment plays a crucial role in development of any country by reducing the poverty level in Pakistan. This research studies the relationship between workers’ remittances and poverty alleviation. It also focused the significant effect on poverty reduction. This study uses time series data for the period of 1972-2013. Autoregressive Distributed Lag (ARDL)Model and Error Correction (ECM)Model has been used in order to find out the long run and short run relationship between the worker’s remittances and poverty level respectively. Thus, inflow of remittances showed the significant and negative impact on poverty level. Moreover, coefficient of error correction model explains the adjustment towards convergence and it has highly significant and negative value. According to this research, Policy makers should strongly focus on positive and effective policies to attract more remittances. JELCODE: JEL: J61

Keywords: ECM, ARDL, AIC, SC

Procedia PDF Downloads 280
19939 Analysis of Brain Signals Using Neural Networks Optimized by Co-Evolution Algorithms

Authors: Zahra Abdolkarimi, Naser Zourikalatehsamad,

Abstract:

Up to 40 years ago, after recognition of epilepsy, it was generally believed that these attacks occurred randomly and suddenly. However, thanks to the advance of mathematics and engineering, such attacks can be predicted within a few minutes or hours. In this way, various algorithms for long-term prediction of the time and frequency of the first attack are presented. In this paper, by considering the nonlinear nature of brain signals and dynamic recorded brain signals, ANFIS model is presented to predict the brain signals, since according to physiologic structure of the onset of attacks, more complex neural structures can better model the signal during attacks. Contribution of this work is the co-evolution algorithm for optimization of ANFIS network parameters. Our objective is to predict brain signals based on time series obtained from brain signals of the people suffering from epilepsy using ANFIS. Results reveal that compared to other methods, this method has less sensitivity to uncertainties such as presence of noise and interruption in recorded signals of the brain as well as more accuracy. Long-term prediction capacity of the model illustrates the usage of planted systems for warning medication and preventing brain signals.

Keywords: co-evolution algorithms, brain signals, time series, neural networks, ANFIS model, physiologic structure, time prediction, epilepsy suffering, illustrates model

Procedia PDF Downloads 274
19938 A Prediction Method for Large-Size Event Occurrences in the Sandpile Model

Authors: S. Channgam, A. Sae-Tang, T. Termsaithong

Abstract:

In this research, the occurrences of large size events in various system sizes of the Bak-Tang-Wiesenfeld sandpile model are considered. The system sizes (square lattice) of model considered here are 25×25, 50×50, 75×75 and 100×100. The cross-correlation between the ratio of sites containing 3 grain time series and the large size event time series for these 4 system sizes are also analyzed. Moreover, a prediction method of the large-size event for the 50×50 system size is also introduced. Lastly, it can be shown that this prediction method provides a slightly higher efficiency than random predictions.

Keywords: Bak-Tang-Wiesenfeld sandpile model, cross-correlation, avalanches, prediction method

Procedia PDF Downloads 377
19937 Control of Biofilm Formation and Inorganic Particle Accumulation on Reverse Osmosis Membrane by Hypochlorite Washing

Authors: Masaki Ohno, Cervinia Manalo, Tetsuji Okuda, Satoshi Nakai, Wataru Nishijima

Abstract:

Reverse osmosis (RO) membranes have been widely used for desalination to purify water for drinking and other purposes. Although at present most RO membranes have no resistance to chlorine, chlorine-resistant membranes are being developed. Therefore, direct chlorine treatment or chlorine washing will be an option in preventing biofouling on chlorine-resistant membranes. Furthermore, if particle accumulation control is possible by using chlorine washing, expensive pretreatment for particle removal can be removed or simplified. The objective of this study was to determine the effective hypochlorite washing condition required for controlling biofilm formation and inorganic particle accumulation on RO membrane in a continuous flow channel with RO membrane and spacer. In this study, direct chlorine washing was done by soaking fouled RO membranes in hypochlorite solution and fluorescence intensity was used to quantify biofilm on the membrane surface. After 48 h of soaking the membranes in high fouling potential waters, the fluorescence intensity decreased to 0 from 470 using the following washing conditions: 10 mg/L chlorine concentration, 2 times/d washing interval, and 30 min washing time. The chlorine concentration required to control biofilm formation decreased as the chlorine concentration (0.5–10 mg/L), the washing interval (1–4 times/d), or the washing time (1–30 min) increased. For the sample solutions used in the study, 10 mg/L chlorine concentration with 2 times/d interval, and 5 min washing time was required for biofilm control. The optimum chlorine washing conditions obtained from soaking experiments proved to be applicable also in controlling biofilm formation in continuous flow experiments. Moreover, chlorine washing employed in controlling biofilm with suspended particles resulted in lower amounts of organic (0.03 mg/cm2) and inorganic (0.14 mg/cm2) deposits on the membrane than that for sample water without chlorine washing (0.14 mg/cm2 and 0.33 mg/cm2, respectively). The amount of biofilm formed was 79% controlled by continuous washing with 10 mg/L of free chlorine concentration, and the inorganic accumulation amount decreased by 58% to levels similar to that of pure water with kaolin (0.17 mg/cm2) as feed water. These results confirmed the acceleration of particle accumulation due to biofilm formation, and that the inhibition of biofilm growth can almost completely reduce further particle accumulation. In addition, effective hypochlorite washing condition which can control both biofilm formation and particle accumulation could be achieved.

Keywords: reverse osmosis, washing condition optimization, hypochlorous acid, biofouling control

Procedia PDF Downloads 344
19936 Application of Forensic Entomology to Estimate the Post Mortem Interval

Authors: Meriem Taleb, Ghania Tail, Fatma Zohra Kara, Brahim Djedouani, T. Moussa

Abstract:

Forensic entomology has grown immensely as a discipline in the past thirty years. The main purpose of forensic entomology is to establish the post mortem interval or PMI. Three days after the death, insect evidence is often the most accurate and sometimes the only method of determining elapsed time since death. This work presents the estimation of the PMI in an experiment to test the reliability of the accumulated degree days (ADD) method and the application of this method in a real case. The study was conducted at the Laboratory of Entomology at the National Institute for Criminalistics and Criminology of the National Gendarmerie, Algeria. The domestic rabbit Oryctolagus cuniculus L. was selected as the animal model. On 08th July 2012, the animal was killed. Larvae were collected and raised to adulthood. Estimation of oviposition time was calculated by summing up average daily temperatures minus minimum development temperature (also specific to each species). When the sum is reached, it corresponds to the oviposition day. Weather data were obtained from the nearest meteorological station. After rearing was accomplished, three species emerged: Lucilia sericata, Chrysomya albiceps, and Sarcophaga africa. For Chrysomya albiceps species, a cumulation of 186°C is necessary. The emergence of adults occured on 22nd July 2012. A value of 193.4°C is reached on 9th August 2012. Lucilia sericata species require a cumulation of 207°C. The emergence of adults occurred on 23rd, July 2012. A value of 211.35°C is reached on 9th August 2012. We should also consider that oviposition may occur more than 12 hours after death. Thus, the obtained PMI is in agreement with the actual time of death. We illustrate the use of this method during the investigation of a case of a decaying human body found on 03rd March 2015 in Bechar, South West of Algerian desert. Maggots were collected and sent to the Laboratory of Entomology. Lucilia sericata adults were identified on 24th March 2015 after emergence. A sum of 211.6°C was reached on 1st March 2015 which corresponds to the estimated day of oviposition. Therefore, the estimated date of death is 1st March 2015 ± 24 hours. The estimated PMI by accumulated degree days (ADD) method seems to be very precise. Entomological evidence should always be used in homicide investigations when the time of death cannot be determined by other methods.

Keywords: forensic entomology, accumulated degree days, postmortem interval, diptera, Algeria

Procedia PDF Downloads 290
19935 Evaluation of Non-Pharmacological Method-Transcervical Foley Catheter and Misoprostol to Intravaginal Misoprostol for Preinduction Cervical Ripening

Authors: Krishna Dahiya, Esha Charaya

Abstract:

Induction of labour is a common obstetrical intervention. Around 1 in every 4 patient undergo induction of labour for different indications Purpose: To study the efficacy of the combination of Foley bulb and vaginal misoprostol in comparison to vaginal misoprostol alone for cervical ripening and induction of labour. Methods: A prospective randomised study was conducted on 150 patients with term singleton pregnancy admitted for induction of labour. Seventy-five patients were induced with both Foley bulb, and vaginal misoprostol and another 75 were given vaginal misoprostol alone for induction of labour. Both groups were then compared with respect to change in Bishop score, induction to the active phase of labour interval, induction delivery interval, duration of labour, maternal complications and neonatal outcomes. Data was analysed using statistical software SPSS version 11.5. Tests with P,.05 were considered significant. Results: The two groups were comparable with respect to maternal age, parity, gestational age, indication for induction, and initial Bishop scores. Both groups had a significant change in Bishop score (2.99 ± 1.72 and 2.17 ± 1.48 respectively with statistically significant difference (p=0.001 S, 95% C.I. -0.1978 to 0.8378). Mean induction to delivery interval was significantly lower in the combination group (11.76 ± 5.89 hours) than misoprostol group (14.54 ± 7.32 hours). Difference was of 2.78 hours (p=0.018,S, 95% CI -5.1042 to -0.4558). Induction to delivery interval was significantly lower in nulliparous women of combination group (13.64 ± 5.75 hours) than misoprostol group (18.4±7.09 hours), and the difference was of 4.76 hours (p=0.002, S, 95% CI 1.0465 to 14.7335). There was no difference between the groups in the mode of delivery, infant weight, Apgar score and intrapartum complications. Conclusion: From the present study it was concluded that addition of Foley catheter to vaginal misoprostol have the synergistic effect and results in early cervical ripening and delivery. These results suggest that the combination may be used to achieve timely and safe delivery in the presence of an unfavorable cervix. A combination of the Foley bulb and vaginal misoprostol resulted in a shorter induction-to-delivery time when compared with vaginal misoprostol alone without increasing labor complications.

Keywords: Bishop score, Foley catheter, induction of labor, misoprostol

Procedia PDF Downloads 304
19934 Effects of Financial Development on Economic Growth in South Asia

Authors: Anupam Das

Abstract:

Although financial liberalization has been one of the most important policy prescriptions of international organizations like the World Bank and the IMF, the effect of financial liberalization on economic growth in developing countries is far from unanimous. Since the '80s, South Asian countries made a significant development in liberalization the financial sector. However, due to unavailability of a sufficient number of time series observations, the relationship between economic growth and financial development has not been investigated adequately. We aim to fill this gap by examining time series data of five developing countries from the South Asian region: Bangladesh, India, Pakistan, Sri Lanka, and Nepal. Applying the cointegration tests and Granger causality within the vector error correction model (VECM), we do not find unanimous evidence of financial development on positive economic growth. These results are helpful for developing countries which have been trying to liberalize the financial sector in recent decades.

Keywords: economic growth, financial development, Granger causality, South Asia

Procedia PDF Downloads 369
19933 On Coverage Probability of Confidence Intervals for the Normal Mean with Known Coefficient of Variation

Authors: Suparat Niwitpong, Sa-aat Niwitpong

Abstract:

Statistical inference of normal mean with known coefficient of variation has been investigated recently. This phenomenon occurs normally in environment and agriculture experiments when the scientist knows the coefficient of variation of their experiments. In this paper, we constructed new confidence intervals for the normal population mean with known coefficient of variation. We also derived analytic expressions for the coverage probability of each confidence interval. To confirm our theoretical results, Monte Carlo simulation will be used to assess the performance of these intervals based on their coverage probabilities.

Keywords: confidence interval, coverage probability, expected length, known coefficient of variation

Procedia PDF Downloads 387
19932 Increasing the Apparent Time Resolution of Tc-99m Diethylenetriamine Pentaacetic Acid Galactosyl Human Serum Albumin Dynamic SPECT by Use of an 180-Degree Interpolation Method

Authors: Yasuyuki Takahashi, Maya Yamashita, Kyoko Saito

Abstract:

In general, dynamic SPECT data acquisition needs a few minutes for one rotation. Thus, the time-activity curve (TAC) derived from the dynamic SPECT is relatively coarse. In order to effectively shorten the interval, between data points, we adopted a 180-degree interpolation method. This method is already used for reconstruction of the X-ray CT data. In this study, we applied this 180-degree interpolation method to SPECT and investigated its effectiveness.To briefly describe the 180-degree interpolation method: the 180-degree data in the second half of one rotation are combined with the 180-degree data in the first half of the next rotation to generate a 360-degree data set appropriate for the time halfway between the first and second rotations. In both a phantom and a patient study, the data points from the interpolated images fell in good agreement with the data points tracking the accumulation of 99mTc activity over time for appropriate region of interest. We conclude that data derived from interpolated images improves the apparent time resolution of dynamic SPECT.

Keywords: dynamic SPECT, time resolution, 180-degree interpolation method, 99mTc-GSA.

Procedia PDF Downloads 491
19931 Power Quality Improvement Using Interval Type-2 Fuzzy Logic Controller for Five-Level Shunt Active Power Filter

Authors: Yousfi Abdelkader, Chaker Abdelkader, Bot Youcef

Abstract:

This article proposes a five-level shunt active power filter for power quality improvement using a interval type-2 fuzzy logic controller (IT2 FLC). The reference compensating current is extracted using the P-Q theory. The majority of works previously reported are based on two-level inverters with a conventional Proportional integral (PI) controller, which requires rigorous mathematical modeling of the system. In this paper, a IT2 FLC controlled five-level active power filter is proposed to overcome the problem associated with PI controller. The IT2 FLC algorithm is applied for controlling the DC-side capacitor voltage as well as the harmonic currents of the five-level active power filter. The active power filter with a IT2 FLC is simulated in MATLAB Simulink environment. The simulated response shows that the proposed shunt active power filter controller has produced a sinusoidal supply current with low harmonic distortion and in phase with the source voltage.

Keywords: power quality, shunt active power filter, interval type-2 fuzzy logic controller (T2FL), multilevel inverter

Procedia PDF Downloads 173
19930 Forecasting the Fluctuation of Currency Exchange Rate Using Random Forest

Authors: Lule Basha, Eralda Gjika

Abstract:

The exchange rate is one of the most important economic variables, especially for a small, open economy such as Albania. Its effect is noticeable in one country's competitiveness, trade and current account, inflation, wages, domestic economic activity, and bank stability. This study investigates the fluctuation of Albania’s exchange rates using monthly average foreign currency, Euro (Eur) to Albanian Lek (ALL) exchange rate with a time span from January 2008 to June 2021, and the macroeconomic factors that have a significant effect on the exchange rate. Initially, the Random Forest Regression algorithm is constructed to understand the impact of economic variables on the behavior of monthly average foreign currencies exchange rates. Then the forecast of macro-economic indicators for 12 months was performed using time series models. The predicted values received are placed in the random forest model in order to obtain the average monthly forecast of the Euro to Albanian Lek (ALL) exchange rate for the period July 2021 to June 2022.

Keywords: exchange rate, random forest, time series, machine learning, prediction

Procedia PDF Downloads 99
19929 A Review of Paleo-Depositional Environment and Thermal Alteration Index of Carboniferous, Permian, and Triassic of A1-9 Well, NW Libya

Authors: M. A. Alrabib, Y. Sherif, A. K. Mohamed, E. A. Elfandi, E. I. Fandi

Abstract:

This paper introduces a paleo-environmental and hydrocarbon show in this well was identified in the interval of Dembaba formation to the Hassaona Formation was poor to very poor oil show. And from palaeo-environmental analysis there is neither particularly good reservoir nor source rock have been developed in the area. Recent palaeo-environment work undertakes that the sedimentary succession in this area comprises the Upper Paleozoic rock of the Carboniferous and Permian and the Mesozoic (Triassic) sedimentary sequences. No early Paleozoic rocks have been found in this area, these rocks were eroding during the Late Carboniferous and Early Permian time. During Latest Permian and earliest Triassic time evidence for major marine transgression has occurred. From depths 5930-5940 feet, to 10800-10810 feet, the TAI of the Al Guidr, the Bir Al Jaja Al Uotia, Hebilia and the top varies between 3+ to 4-(mature-dry gas). This interval corporate the rest part of the Dembaba Formation. From depth 10800- 10810 feet, until total sediment depth (11944 feet Log) which corporate the rest of the Dembaba and underlying equivalents of the Assedjefar and M Rar Formations and the underlying Indeterminate unit (Hassouna Formation) the TAI varies between 4 and 5 (dry gas-black and deformed).

Keywords: paleoenvironmental, thermal alteration index, north western Libya, hydrocarbon

Procedia PDF Downloads 465
19928 Experimental Characterization of Flowable Cement Pastes Made with Marble Waste

Authors: F. Messaoudi, O. Haddad, R. Bouras, S. Kaci

Abstract:

The development of self-compacting concrete (SCC) marks a huge step towards improved efficiency and working conditions on construction sites and in the precast industry. SCC flows easily into more complex shapes and through reinforcement bars, reduces the manpower required for the placement; no vibration is required to ensure correct compaction of concrete. This concrete contains a high volume of binder which is controlled by their rheological behavior. The paste consists of binders (Portland cement with or without supplementary cementitious materials), water, chemical admixtures and fillers. In this study, two series of tests were performed on self-compacting cement pastes made with marble waste additions as the mineral addition. The first series of this investigation was to determine the flow time of paste using Marsh cone, the second series was to determine the rheological parameters of the same paste namely yield stress and plastic viscosity using the rheometer Haake RheoStress 1. The results of this investigation allowed us to study the evolution of the yield stress, viscosity and the flow time Marsh cone paste as a function of the composition of the paste. A correlation between the results obtained on the flow test Marsh cone and those of the plastic viscosity on the mottled different cement pastes is proposed.

Keywords: adjuvant, rheological parameter, self-compacting cement pastes, waste marble

Procedia PDF Downloads 271
19927 The Prevalence and Associated Factors of Frailty and Its Relationship with Falls in Patients with Schizophrenia

Authors: Bo-Jian Wu, Si-Heng Wu

Abstract:

Objectives: Frailty is a condition of a person who has chronic health problems complicated by a loss of physiological reserve and deteriorating functional abilities. The frailty syndrome was defined by Fried and colleagues, i.e., weight loss, fatigue, decreased grip strength, slow gait speed, and low physical activity. However, to our best knowledge, there have been rare studies exploring the prevalence of frailty and its association with falls in patients with schizophrenia. Methods: A total of 559 hospitalized patients were recruited from a public psychiatric hospital in 2013. The majority of the subjects were males (361, 64.6%). The average age was 53.5 years. All patients received the assessment of frailty status defined by Fried and colleagues. The status of a fall within one year after the assessment of frailty, clinical and demographic data was collected from medical records. Logistic regression was used to calculate the odds ratio of associated factors. Results : A total of 9.2% of the participants met the criteria of frailty. The percentage of patients having a fall was 7.2%. Age were significantly associated with frailty (odds ratio = 1.057, 95% confidence interval = 1.025-1.091); however, sex was not associated with frailty (p = 0.17). After adjustment for age and sex, frailty status was associated with a fall (odds ratio = 3.62, 95% confidence interval = 1.58-8.28). Concerning the components of frailty, decreased grip strength (odds ratio = 2.44, 95% confidence interval = 1.16-5.14), slow gait speed (odds ratio = 2.82, 95% confidence interval = 1.21-6.53), and low physical activity (odds ratio = 2.64, 95% confidence interval = 1.21-5.78) were found to be associated with a fall. Conclusions: Our findings suggest the prevalence of frailty was about 10% in hospitalized patients with chronic patients with schizophrenia, and frailty status was significant with a fall in this group. By using the status of frailty, it may be beneficial to potential target candidates having fallen in the future as early as possible. The effective intervention of prevention of further falls may be given in advance. Our results bridge this gap and open a potential avenue for the prevention of falls in patients with schizophrenia. Frailty is certainly an important factor for maintaining wellbeing among these patients.

Keywords: fall, frailty, schizophrenia, Taiwan

Procedia PDF Downloads 161
19926 Forecasting Container Throughput: Using Aggregate or Terminal-Specific Data?

Authors: Gu Pang, Bartosz Gebka

Abstract:

We forecast the demand of total container throughput at the Indonesia’s largest seaport, Tanjung Priok Port. We propose four univariate forecasting models, including SARIMA, the additive Seasonal Holt-Winters, the multiplicative Seasonal Holt-Winters and the Vector Error Correction Model. Our aim is to provide insights into whether forecasting the total container throughput obtained by historical aggregated port throughput time series is superior to the forecasts of the total throughput obtained by summing up the best individual terminal forecasts. We test the monthly port/individual terminal container throughput time series between 2003 and 2013. The performance of forecasting models is evaluated based on Mean Absolute Error and Root Mean Squared Error. Our results show that the multiplicative Seasonal Holt-Winters model produces the most accurate forecasts of total container throughput, whereas SARIMA generates the worst in-sample model fit. The Vector Error Correction Model provides the best model fits and forecasts for individual terminals. Our results report that the total container throughput forecasts based on modelling the total throughput time series are consistently better than those obtained by combining those forecasts generated by terminal-specific models. The forecasts of total throughput until the end of 2018 provide an essential insight into the strategic decision-making on the expansion of port's capacity and construction of new container terminals at Tanjung Priok Port.

Keywords: SARIMA, Seasonal Holt-Winters, Vector Error Correction Model, container throughput

Procedia PDF Downloads 500
19925 Modeling of Diurnal Pattern of Air Temperature in a Tropical Environment: Ile-Ife and Ibadan, Nigeria

Authors: Rufus Temidayo Akinnubi, M. O. Adeniyi

Abstract:

Existing diurnal air temperature models simulate night time air temperature over Nigeria with high biases. An improved parameterization is presented for modeling the diurnal pattern of air temperature (Ta) which is applicable in the calculation of turbulent heat fluxes in Global climate models, based on Nigeria Micrometeorological Experimental site (NIMEX) surface layer observations. Five diurnal Ta models for estimating hourly Ta from daily maximum, daily minimum, and daily mean air temperature were validated using root-mean-square error (RMSE), Mean Error Bias (MBE) and scatter graphs. The original Fourier series model showed better performance for unstable air temperature parameterizations while the stable Ta was strongly overestimated with a large error. The model was improved with the inclusion of the atmospheric cooling rate that accounts for the temperature inversion that occurs during the nocturnal boundary layer condition. The MBE and RMSE estimated by the modified Fourier series model reduced by 4.45 oC and 3.12 oC during the transitional period from dry to wet stable atmospheric conditions. The modified Fourier series model gave good estimation of the diurnal weather patterns of Ta when compared with other existing models for a tropical environment.

Keywords: air temperature, mean bias error, Fourier series analysis, surface energy balance,

Procedia PDF Downloads 226
19924 The Relationships between Carbon Dioxide (CO2) Emissions, Energy Consumption and GDP per capita for Oman: Time Series Analysis, 1980–2010

Authors: Jinhoa Lee

Abstract:

The relationships between environmental quality, energy use and economic output have created growing attention over the past decades among researchers and policy makers. Focusing on the empirical aspects of the role of CO2 emissions and energy use in affecting the economic output, this paper is an effort to fulfil the gap in a comprehensive case study at a country level using modern econometric techniques. To achieve the goal, this country-specific study examines the short-run and long-run relationships among energy consumption, carbon dioxide (CO2) emissions and gross domestic product (GDP) for Oman using time series analysis from the year 1980-2010. To investigate the relationships between the variables, this paper employs the Augmented Dickey Fuller (ADF) test for stationary, Johansen maximum likelihood method for co-integration and a Vector Error Correction Model (VECM) for both short- and long-run causality among the research variables for the sample. All the variables in this study show very strong significant effects on GDP in the country for the long term. The long-run equilibrium in the VECM suggests positive long-run causalities from CO2 emissions to GDP. Conversely, negative impacts of energy consumption on GDP are found to be significant in Oman during the period. In the short run, there exist negative unidirectional causalities among GDP, CO2 emissions and energy consumption running from GDP to CO2 emissions and from energy consumption to CO2 emissions. Overall, the results support arguments that there are relationships among environmental quality, energy use and economic output in Oman over of period 1980-2010.

Keywords: CO2 emissions, energy consumption, GDP, Oman, time series analysis

Procedia PDF Downloads 458
19923 Kýklos Dimensional Geometry: Entity Specific Core Measurement System

Authors: Steven D. P Moore

Abstract:

A novel method referred to asKýklos(Ky) dimensional geometry is proposed as an entity specific core geometric dimensional measurement system. Ky geometric measures can constructscaled multi-dimensionalmodels using regular and irregular sets in IRn. This entity specific-derived geometric measurement system shares similar fractal methods in which a ‘fractal transformation operator’ is applied to a set S to produce a union of N copies. The Kýklos’ inputs use 1D geometry as a core measure. One-dimensional inputs include the radius interval of a circle/sphere or the semiminor/semimajor axes intervals of an ellipse or spheroid. These geometric inputs have finite values that can be measured by SI distance units. The outputs for each interval are divided and subdivided 1D subcomponents with a union equal to the interval geometry/length. Setting a limit of subdivision iterations creates a finite value for each 1Dsubcomponent. The uniqueness of this method is captured by allowing the simplest 1D inputs to define entity specific subclass geometric core measurements that can also be used to derive length measures. Current methodologies for celestial based measurement of time, as defined within SI units, fits within this methodology, thus combining spatial and temporal features into geometric core measures. The novel Ky method discussed here offers geometric measures to construct scaled multi-dimensional structures, even models. Ky classes proposed for consideration include celestial even subatomic. The application of this offers incredible possibilities, for example, geometric architecture that can represent scaled celestial models that incorporates planets (spheroids) and celestial motion (elliptical orbits).

Keywords: Kyklos, geometry, measurement, celestial, dimension

Procedia PDF Downloads 162
19922 Effect of High Intensity Interval Training and Moderate Interval Continuous Training on Cardiovascular Endurance In young Healthy Female

Authors: Sidra Majeed, Irum Ali, Aroosa Ishfaq, Munazzah Parveen

Abstract:

Objectives: The objective is to compare the effects of high-intensity interval training VS moderate moderate-intensity continuous training on cardiovascular endurance in young healthy females. Method: 30 young, healthy females were collected and randomly assigned into two training groups, HIIT and MICT, each group having a sample size of (n=15). There will be three parameters to be tested, including (VO2max, Resting heart rate, and Rate perceived exertion). Each group will be tested at three different times, e.g. (at Baseline measurement, after two weeks and after four weeks). For the first two weeks, the HIIT group has to perform at 70%HRR and for the third and fourth weeks, at 75%HRR for two minutes, followed by an active resting interval at 30%HRR for two minutes (1:1) with warm-up and cool-down period (2 minutes each period ) on the treadmill. For the first two weeks, the MICT group has to perform at 40%HRR and for the third and fourth weeks at 50% HRR for fifteen minutes continuously on the treadmill, including warm up and cool down period (2 minutes each period). Result: The final assessment of HIIT and MICT groups had shown p values for VO2max (p=.000), RHR (p=.323) and for RPE (p=.085). These values indicating significant improvement in these three parameters in both groups. Conclusion: This study showed that there were significant improvements in both groups but there were more improvements in VO2max in HIIT group so, it is proved that HIIT is more beneficial than MICT in improving cardiovascular endurance.

Keywords: HIIT, MICT, RPE, RHR

Procedia PDF Downloads 41
19921 Forecasting Stock Indexes Using Bayesian Additive Regression Tree

Authors: Darren Zou

Abstract:

Forecasting the stock market is a very challenging task. Various economic indicators such as GDP, exchange rates, interest rates, and unemployment have a substantial impact on the stock market. Time series models are the traditional methods used to predict stock market changes. In this paper, a machine learning method, Bayesian Additive Regression Tree (BART) is used in predicting stock market indexes based on multiple economic indicators. BART can be used to model heterogeneous treatment effects, and thereby works well when models are misspecified. It also has the capability to handle non-linear main effects and multi-way interactions without much input from financial analysts. In this research, BART is proposed to provide a reliable prediction on day-to-day stock market activities. By comparing the analysis results from BART and with time series method, BART can perform well and has better prediction capability than the traditional methods.

Keywords: BART, Bayesian, predict, stock

Procedia PDF Downloads 125