Search results for: time series data mining
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 38105

Search results for: time series data mining

35045 Diagnosis of Induction Machine Faults by DWT

Authors: Hamidreza Akbari

Abstract:

In this paper, for detection of inclined eccentricity in an induction motor, time–frequency analysis of the stator startup current is carried out. For this purpose, the discrete wavelet transform is used. Data are obtained from simulations, using winding function approach. The results show the validity of the approach for detecting the fault and discriminating with respect to other faults.

Keywords: induction machine, fault, DWT, electric

Procedia PDF Downloads 336
35044 Investigation of the Fading Time Effects on Microstructure and Mechanical Properties in Vermicular Cast Iron

Authors: Mehmet Ekici

Abstract:

In this study, the fading time affecting the mechanical properties and microstructures of vermicular cast iron were studied. Pig iron and steel scrap weighing about 12 kg were charged into the high-frequency induction furnace crucible and completely melted for production of vermicular cast iron. The slag was skimmed using a common flux. After fading time was set at 1. 3 and 5 minutes. In this way, three vermicular cast iron was produced that same composition but different phase structures. The microstructure of specimens was investigated, and uni-axial tensile test and the Charpy impact test were performed, and their micro-hardness measurements were done in order to characterize the mechanical behaviours of vermicular cast iron.

Keywords: vermicular cast iron, fading time, hardness, tensile test and impact test

Procedia PDF Downloads 326
35043 Algorithm and Software Based on Multilayer Perceptron Neural Networks for Estimating Channel Use in the Spectral Decision Stage in Cognitive Radio Networks

Authors: Danilo López, Johana Hernández, Edwin Rivas

Abstract:

The use of the Multilayer Perceptron Neural Networks (MLPNN) technique is presented to estimate the future state of use of a licensed channel by primary users (PUs); this will be useful at the spectral decision stage in cognitive radio networks (CRN) to determine approximately in which time instants of future may secondary users (SUs) opportunistically use the spectral bandwidth to send data through the primary wireless network. To validate the results, sequences of occupancy data of channel were generated by simulation. The results show that the prediction percentage is greater than 60% in some of the tests carried out.

Keywords: cognitive radio, neural network, prediction, primary user

Procedia PDF Downloads 353
35042 Case-Based Reasoning for Build Order in Real-Time Strategy Games

Authors: Ben G. Weber, Michael Mateas

Abstract:

We present a case-based reasoning technique for selecting build orders in a real-time strategy game. The case retrieval process generalizes features of the game state and selects cases using domain-specific recall methods, which perform exact matching on a subset of the case features. We demonstrate the performance of the technique by implementing it as a component of the integrated agent framework of McCoy and Mateas. Our results demonstrate that the technique outperforms nearest-neighbor retrieval when imperfect information is enforced in a real-time strategy game.

Keywords: case based reasoning, real time strategy systems, requirements elicitation, requirement analyst, artificial intelligence

Procedia PDF Downloads 423
35041 Modern Agriculture and Employment Generation in Nigeria: A Recursive Model Approach

Authors: Ese Urhie, Olabisi Popoola, Obindah Gershon

Abstract:

Several policies and programs initiated to address the challenge of unemployment in Nigeria seem to be inadequate. The desired structural transformation which is expected to absorb the excess labour in the economy is yet to be achieved. The agricultural sector accounts for almost half of the labour force with very low productivity. This could partly explain why the much anticipated structural transformation has not been achieved. A major reason for the low productivity is the fact that the production process is predominantly based on the use of traditional tools. In view of the underdeveloped nature of the agricultural sector, Nigeria still has huge potentials for productivity enhancement through modern technology. Aside from productivity enhancement, modern agriculture also stimulates both backward and forward linkages that promote investment and thus generate employment. Contrary to the apprehension usually expressed by many stake-holders about the adoption of modern technology by labour-abundant less-developed countries, this study showed that though there will be job loss initially, the reverse will be the case in the long-run. The outcome of this study will enhance the understanding of all stakeholders in the sector and also encourage them to adopt modern techniques of farming. It will also aid policy formulation at both sectoral and national levels. The recursive model and analysis adopted in the study is useful because it exhibits a unilateral cause-and-effect relationship which most simultaneous equation models do not. It enables the structural equations to be ordered in such a way that the first equation includes only predetermined variables on the right-hand side, while the solution for the final endogenous variable is completely determined by all equations of the system. The study examines the transmission channels and effect of modern agriculture on agricultural productivity and employment growth in Nigeria, via its forward and backward linkages. Using time series data spanning 1980 to 2014, the result of the analyses shows that: (i) a significant and positive relationship between agricultural productivity growth and modern agriculture; (ii) a significant and negative relationship between export price index and agricultural productivity growth; (iii) a significant and positive relationship between export and investment; and (iv) a significant and positive relationship between investment and employment growth. The unbalanced growth theory will be a good strategy to adopt by developing countries such as Nigeria.

Keywords: employment, modern agriculture, productivity, recursive model

Procedia PDF Downloads 243
35040 Principal Components Analysis of the Causes of High Blood Pressure at Komfo Anokye Teaching Hospital, Ghana

Authors: Joseph K. A. Johnson

Abstract:

Hypertension affects 20 percent of the people within the ages 55 upward in Ghana. Of these, almost one-third are unaware of their condition. Also at the age of 55, more men turned to have hypertension than women. After that age, the condition becomes more prevalent with women. Hypertension is significantly more common in African Americans of both sexes than the racial or ethnic groups. This study was conducted to determine the causes of high blood pressure in Ashanti Region, Ghana. The study employed One Hundred and Seventy (170) respondents. The sample population for the study was all the available respondents at the time of the data collection. The research was conducted using primary data where convenience sampling was used to locate the respondents. A set of questionnaire were used to gather the data for the study. The gathered data was analysed using principal component analysis. The study revealed that, personal description, lifestyle behavior and risk awareness as some of the causes of high blood pressure in Ashanti Region. The study therefore recommend that people must be advice to see to their personal characteristics that may contribute to high blood pressure such as controlling of their temper and how to react perfectly to stressful situations. They must be educated on the factors that may increase the level of their blood pressure such as the essence of seeing a medical doctor before taking in any drug. People must also be made known by the public health officers to those lifestyles behaviour such as smoking and drinking of alcohol which are major contributors of high blood pressure.

Keywords: high blood pressure, principal component analysis, hypertension, public health

Procedia PDF Downloads 467
35039 Raising Multilingual Awareness towards Plurilingual Competence Development: Through Which Approach and Which Pedagogical Material-A Case Study in the Greek Primary Education

Authors: Eftychia Damaskou

Abstract:

This article intends to place the question of the adequate approach for teaching multilingualism within the public education. Linguistic education, as it is defined by the Common European Framework of Reference for the Languages, is no longer the proficiency in one or two languages. It’s about the development of a linguistic repertoire, where all linguistic skills find their place. In fact, the linguistic theories that frame the development of plurilingual competence point out the affective and intercultural aspect of such a process, insisting on an awareness of linguistic diversification, rather than an acquisition of communicative competence in many languages. In this spirit, our article attempts to go beyond a mere plurilingual awareness, present a research based on an experience in class, within 115 pupils, aiming at the development of plurilingual competence in five unknown foreign languages. This experience was held through a teaching unit personally conceived and applied, and consisted of a series of 6 activities based on a cross-linguistic content approach. The data analysis proves to be very interesting, as it reveals the development of plurilingual competences, as well as positive attitudes towards less common languages by the majority of our sample.

Keywords: multilingual awareness, multilingual teaching material, plurilingual competence

Procedia PDF Downloads 436
35038 Estimating Destinations of Bus Passengers Using Smart Card Data

Authors: Hasik Lee, Seung-Young Kho

Abstract:

Nowadays, automatic fare collection (AFC) system is widely used in many countries. However, smart card data from many of cities does not contain alighting information which is necessary to build OD matrices. Therefore, in order to utilize smart card data, destinations of passengers should be estimated. In this paper, kernel density estimation was used to forecast probabilities of alighting stations of bus passengers and applied to smart card data in Seoul, Korea which contains boarding and alighting information. This method was also validated with actual data. In some cases, stochastic method was more accurate than deterministic method. Therefore, it is sufficiently accurate to be used to build OD matrices.

Keywords: destination estimation, Kernel density estimation, smart card data, validation

Procedia PDF Downloads 335
35037 Examining Coping Resources and Ways of Strategic Coping for Individuals with Spinal Cord Injury During the COVID-19 Crisis

Authors: Se-Hyuk Park, Hee-Jung Seo

Abstract:

Previous studies have investigated effective coping strategies for excessive stress, positive adaptation, resilience, mental health, and personal growth. However, to the best of the authors' knowledge, little research has been conducted to investigate how Koreans with physical disabilities deal with the COVID-19 pandemic. The purpose of this study was to identify coping strategies and coping resources that Koreans with physical disabilities utilized during the COVID-19 crisis. This study used semi-structured, in-depth interviews with 15 participants. Data were qualitatively analyzed using the constant comparative method with content mapping and content mining questions. We identified three salient themes that were used by participants as coping strategies to deal with various COVID-related challenges: (a) engagement in meaningful activities, (b) improvement of social and emotional support, and (c) experience of resilience. The findings of the present study highlighted that Korean adults with SCI actively engaged in various leisure activities, maintained and developed closer social relationships, and experienced resilience to face COVID-19-related stressors. These coping strategies were noted as a catalyst for physical health as well as psychological well-being of individuals with SCI.

Keywords: spinal cord injury, covid-19 pandemic, coping strategies, coping resources, leisure

Procedia PDF Downloads 20
35036 On the Accuracy of Basic Modal Displacement Method Considering Various Earthquakes

Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar

Abstract:

Time history seismic analysis is supposed to be the most accurate method to predict the seismic demand of structures. On the other hand, the required computational time of this method toward achieving the result is its main deficiency. While being applied in optimization process, in which the structure must be analyzed thousands of time, reducing the required computational time of seismic analysis of structures makes the optimization algorithms more practical. Apparently, the invented approximate methods produce some amount of errors in comparison with exact time history analysis but the recently proposed method namely, Complete Quadratic Combination (CQC) and Sum Root of the Sum of Squares (SRSS) drastically reduces the computational time by combination of peak responses in each mode. In the present research, the Basic Modal Displacement (BMD) method is introduced and applied towards estimation of seismic demand of main structure. Seismic demand of sampled structure is estimated by calculation of modal displacement of basic structure (in which the modal displacement has been calculated). Shear steel sampled structures are selected as case studies. The error applying the introduced method is calculated by comparison of the estimated seismic demands with exact time history dynamic analysis. The efficiency of the proposed method is demonstrated by application of three types of earthquakes (in view of time of peak ground acceleration).

Keywords: time history dynamic analysis, basic modal displacement, earthquake-induced demands, shear steel structures

Procedia PDF Downloads 344
35035 Evaluating Effectiveness of Training and Development Corporate Programs: The Russian Agribusiness Context

Authors: Ekaterina Tikhonova

Abstract:

This research is aimed to evaluate the effectiveness of T&D (Training and Development) on the example of two T&D programs for the Executive TOP Management run in 2012, 2015-2016 in Komos Group. This study is commissioned to research the effectiveness of two similar corporate T&D programs (within one company) in two periods of time (2012, 2015-2016) through evaluating the programs’ effectiveness using the four-level Kirkpatrick’s model of evaluating T&D programs and calculating ROI as an instrument for T&D program measuring by Phillips’ formula. The research investigates the correlation of two figures: the ROI calculated and the rating percentage scale per the ROI implementation (Wagle’s scale). The study includes an assessment of feedback 360 (Kirkpatrick's model) and Phillips’ ROI Methodology that provides a step-by-step process for collecting data, summarizing and processing the collected information. The data is collected from the company accounting data, the HR budgets, MCFO and the company annual reports for the research periods. All analyzed data and reports are organized and presented in forms of tables, charts, and graphs. The paper also gives a brief description of some constrains of the research considered. After ROI calculation, the study reveals that ROI ranges between the average implementation (65% to 75%) by Wagle’s scale that can be considered as a positive outcome. The paper also gives some recommendations how to use ROI in practice and describes main benefits of ROI implementation.

Keywords: ROI, organizational performance, efficacy of T&D program, employee performance

Procedia PDF Downloads 237
35034 Solution Approaches for Some Scheduling Problems with Learning Effect and Job Dependent Delivery Times

Authors: M. Duran Toksari, Berrin Ucarkus

Abstract:

In this paper, we propose two algorithms to optimally solve makespan and total completion time scheduling problems with learning effect and job dependent delivery times in a single machine environment. The delivery time is the extra time to eliminate adverse effect between the main processing and delivery to the customer. In this paper, we introduce the job dependent delivery times for some single machine scheduling problems with position dependent learning effect, which are makespan are total completion. The results with respect to two algorithms proposed for solving of the each problem are compared with LINGO solutions for 50-jobs, 100-jobs and 150-jobs problems. The proposed algorithms can find the same results in shorter time.

Keywords: delivery Times, learning effect, makespan, scheduling, total completion time

Procedia PDF Downloads 459
35033 The Impact of Two Factors on EFL Learners' Fluency

Authors: Alireza Behfar, Mohammad Mahdavi

Abstract:

Nowadays, in the light of progress in the world of science, technology and communications, mastery of learning international languages is a sure and needful matter. In learning any language as a second language, progress and achieving a desirable level in speaking is indeed important for approximately all learners. In this research, we find out how preparation can influence L2 learners' oral fluency with respect to individual differences in working memory capacity. The participants consisted of sixty-one advanced L2 learners including MA students of TEFL at Isfahan University as well as instructors teaching English at Sadr Institute in Isfahan. The data collection consisted of two phases: A working memory test (reading span test) and a picture description task, with a one-month interval between the two tasks. Speaking was elicited through speech generation task in which the individuals were asked to discuss four topics emerging in two pairs. The two pairs included one simple and one complex topic and was accompanied by planning time and without any planning time respectively. Each topic was accompanied by several relevant pictures. L2 fluency was assessed based on preparation. The data were then analyzed in terms of the number of syllables, the number of silent pauses, and the mean length of pauses produced per minute. The study offers implications for strategies to improve learners’ both fluency and working memory.

Keywords: two factors, fluency, working memory capacity, preparation, L2 speech production reading span test picture description

Procedia PDF Downloads 213
35032 Boundary Alert System for Powered Wheelchair in Confined Area Training

Authors: Tsoi Kim Ming, Yu King Pong

Abstract:

Background: With powered wheelchair, patients can travel more easily and conveniently. However, some patients suffer from other difficulties, such as visual impairment, cognitive disorder, or psychological issues, which make them unable to control powered wheelchair safely. Purpose: Therefore, those patients are required to complete a comprehensive driving training by therapists on confined area, which simulates narrow paths in daily live. During the training, therapists will give series of driving instruction to patients, which may be unaware of patients crossing out the boundary of area. To facilitate the training, it is needed to develop a device to provide warning to patients during training Method: We adopt LIDAR for distance sensing started from center of confined area. Then, we program the LIDAR with linear geometry to remember each side of the area. The LIDAR will sense the location of wheelchair continuously. Once the wheelchair is driven out of the boundary, audio alert will be given to patient. Result: Patients can pay their attention to the particular driving situation followed by audio alert during driving training, which can learn how to avoid out of boundary in similar situation next time. Conclusion: Instead of only instructed by therapist, the LIDAR can facilitate the powered wheelchair training by patients actively pay their attention to driving situation. After training, they are able to control the powered wheelchair safely when facing difficult and narrow path in real life.

Keywords: PWC, training, rehab, AT

Procedia PDF Downloads 84
35031 Experimental Studies and CFD Predictions on Hydrodynamics of Gas-Solid Flow in an ICFB with a Draft Tube

Authors: Ravi Gujjula, Chinna Eranna, Narasimha Mangadoddy

Abstract:

Hydrodynamic study of gas and solid flow in an internally circulating fluidized bed with draft tube is made in this paper using high speed camera and pressure probes for the laboratory ICFB test rig 3.0 m X 2.7 m column having a draft tube located in the center of ICFB. Experiments were conducted using different sized sand particles with varying particle size distribution. At each experimental run the standard pressure-flow curves for both draft tube and annular region beds measured and the same time downward particles velocity in the annular bed region were also measured. The effect of superficial gas velocity, static bed height (40, 50 & 60 cm) and the draft tube gap height (10.5 & 14.5 cm) on pressure drop profiles, solid circulation pattern, and gas bypassing dynamics for the ICFB investigated extensively. The mechanism of governing solid recirculation and the pressure losses in an ICFB has been eluded based on gas and solid dynamics obtained from the experimental data. 3D ICFB CFD simulation runs conducted and extracted data validated with ICFB experimental data.

Keywords: icfb, cfd, pressure drop, solids recirculation, bed height, draft tube

Procedia PDF Downloads 507
35030 Time of Death Determination in Medicolegal Death Investigations

Authors: Michelle Rippy

Abstract:

Medicolegal death investigation historically is a field that does not receive much research attention or advancement, as all of the subjects are deceased. Public health threats, drug epidemics and contagious diseases are typically recognized in decedents first, with thorough and accurate death investigations able to assist in epidemiology research and prevention programs. One vital component of medicolegal death investigation is determining the decedent’s time of death. An accurate time of death can assist in corroborating alibies, determining sequence of death in multiple casualty circumstances and provide vital facts in civil situations. Popular television portrays an unrealistic forensic ability to provide the exact time of death to the minute for someone found deceased with no witnesses present. The actuality of unattended decedent time of death determination can generally only be narrowed to a 4-6 hour window. In the mid- to late-20th century, liver temperatures were an invasive action taken by death investigators to determine the decedent’s core temperature. The core temperature was programmed into an equation to determine an approximate time of death. Due to many inconsistencies with the placement of the thermometer and other variables, the accuracy of the liver temperatures was dispelled and this once common place action lost scientific support. Currently, medicolegal death investigators utilize three major after death or post-mortem changes at a death scene. Many factors are considered in the subjective determination as to the time of death, including the cooling of the decedent, stiffness of the muscles, release of blood internally, clothing, ambient temperature, disease and recent exercise. Current research is utilizing non-invasive hospital grade tympanic thermometers to measure the temperature in the each of the decedent’s ears. This tool can be used at the scene and in conjunction with scene indicators may provide a more accurate time of death. The research is significant and important to investigations and can provide an area of accuracy to a historically inaccurate area, considerably improving criminal and civil death investigations. The goal of the research is to provide a scientific basis to unwitnessed deaths, instead of the art that the determination currently is. The research is currently in progress with expected termination in December 2018. There are currently 15 completed case studies with vital information including the ambient temperature, decedent height/weight/sex/age, layers of clothing, found position, if medical intervention occurred and if the death was witnessed. This data will be analyzed with the multiple variables studied and available for presentation in January 2019.

Keywords: algor mortis, forensic pathology, investigations, medicolegal, time of death, tympanic

Procedia PDF Downloads 103
35029 An Evaluation of the Use of Telematics for Improving the Driving Behaviours of Young People

Authors: James Boylan, Denny Meyer, Won Sun Chen

Abstract:

Background: Globally, there is an increasing trend of road traffic deaths, reaching 1.35 million in 2016 in comparison to 1.3 million a decade ago, and overall, road traffic injuries are ranked as the eighth leading cause of death for all age groups. The reported death rate for younger drivers aged 16-19 years is almost twice the rate reported for older drivers aged 25 and above, with a rate of 3.5 road traffic fatalities per annum for every 10,000 licenses held. Telematics refers to a system with the ability to capture real-time data about vehicle usage. The data collected from telematics can be used to better assess a driver's risk. It is typically used to measure acceleration, turn, braking, and speed, as well as to provide locational information. With the Australian government creating the National Telematics Framework, there has been an increase in the government's focus on using telematics data to improve road safety outcomes. The purpose of this study is to test the hypothesis that improvements in telematics measured driving behaviour to relate to improvements in road safety attitudes measured by the Driving Behaviour Questionnaire (DBQ). Methodology: 28 participants were recruited and given a telematics device to insert into their vehicles for the duration of the study. The participant's driving behaviour over the course of the first month will be compared to their driving behaviour in the second month to determine whether feedback from telematics devices improves driving behaviour. Participants completed the DBQ, evaluated using a 6-point Likert scale (0 = never, 5 = nearly all the time) at the beginning, after the first month, and after the second month of the study. This is a well-established instrument used worldwide. Trends in the telematics data will be captured and correlated with the changes in the DBQ using regression models in SAS. Results: The DBQ has provided a reliable measure (alpha = .823) of driving behaviour based on a sample of 23 participants, with an average of 50.5 and a standard deviation of 11.36, and a range of 29 to 76, with higher scores, indicating worse driving behaviours. This initial sample is well stratified in terms of gender and age (range 19-27). It is expected that in the next six weeks, a larger sample of around 40 will have completed the DBQ after experiencing in-vehicle telematics for 30 days, allowing a comparison with baseline levels. The trends in the telematics data over the first 30 days will be compared with the changes observed in the DBQ. Conclusions: It is expected that there will be a significant relationship between the improvements in the DBQ and the trends in reduced telematics measured aggressive driving behaviours supporting the hypothesis.

Keywords: telematics, driving behavior, young drivers, driving behaviour questionnaire

Procedia PDF Downloads 92
35028 Utilization of Cloud-Based Learning Platform for the Enhancement of IT Onboarding System

Authors: Christian Luarca

Abstract:

The study aims to define the efficiency of e-Trainings by the use of cloud platform as part of the onboarding process for IT support engineers. Traditional lecture based trainings involves human resource to guide and assist new hires as part of onboarding which takes time and effort. The use of electronic medium as a platform for training provides a two-way basic communication that can be done in a repetitive manner. The study focuses on determining the most efficient manner of learning the basic knowledge on IT support in the shortest time possible. This was determined by conducting the same set of knowledge transfer categories in two different approaches, one being the e-Training and the other using the traditional method. Performance assessment will be done by the use of Service Tracker Assessment (STA) Tool and Service Manager. Data gathered from this ongoing study will promote the utilization of e-Trainings in the IT onboarding process.

Keywords: cloud platform, e-Training, efficiency, onboarding

Procedia PDF Downloads 139
35027 Multi-Temporal Mapping of Built-up Areas Using Daytime and Nighttime Satellite Images Based on Google Earth Engine Platform

Authors: S. Hutasavi, D. Chen

Abstract:

The built-up area is a significant proxy to measure regional economic growth and reflects the Gross Provincial Product (GPP). However, an up-to-date and reliable database of built-up areas is not always available, especially in developing countries. The cloud-based geospatial analysis platform such as Google Earth Engine (GEE) provides an opportunity with accessibility and computational power for those countries to generate the built-up data. Therefore, this study aims to extract the built-up areas in Eastern Economic Corridor (EEC), Thailand using day and nighttime satellite imagery based on GEE facilities. The normalized indices were generated from Landsat 8 surface reflectance dataset, including Normalized Difference Built-up Index (NDBI), Built-up Index (BUI), and Modified Built-up Index (MBUI). These indices were applied to identify built-up areas in EEC. The result shows that MBUI performs better than BUI and NDBI, with the highest accuracy of 0.85 and Kappa of 0.82. Moreover, the overall accuracy of classification was improved from 79% to 90%, and error of total built-up area was decreased from 29% to 0.7%, after night-time light data from the Visible and Infrared Imaging Suite (VIIRS) Day Night Band (DNB). The results suggest that MBUI with night-time light imagery is appropriate for built-up area extraction and be utilize for further study of socioeconomic impacts of regional development policy over the EEC region.

Keywords: built-up area extraction, google earth engine, adaptive thresholding method, rapid mapping

Procedia PDF Downloads 110
35026 Investigating the Factors Affecting on One Time Passwords Technology Acceptance: A Case Study in Banking Environment

Authors: Sajad Shokohuyar, Mahsa Zomorrodi Anbaji, Saghar Pouyan Shad

Abstract:

According to fast technology growth, modern banking tries to decrease going to banks’ branches and increase customers’ consent. One of the problems which banks face is securing customer’s password. The banks’ solution is one time password creation system. In this research by adapting from acceptance of technology model theory, assesses factors that are effective on banking in Iran especially in using one time password machine by one of the private banks of Iran customers. The statistical population is all of this bank’s customers who use electronic banking service and one time password technology and the questionnaires were distributed among members of statistical population in 5 selected groups of north, south, center, east and west of Tehran. Findings show that confidential preservation, education, ease of utilization and advertising and informing has positive relations and distinct hardware and age has negative relations.

Keywords: security, electronic banking, one time password, information technology

Procedia PDF Downloads 434
35025 Evaluated Nuclear Data Based Photon Induced Nuclear Reaction Model of GEANT4

Authors: Jae Won Shin

Abstract:

We develop an evaluated nuclear data based photonuclear reaction model of GEANT4 for a more accurate simulation of photon-induced neutron production. The evaluated photonuclear data libraries from the ENDF/B-VII.1 are taken as input. Incident photon energies up to 140 MeV which is the threshold energy for the pion production are considered. For checking the validity of the use of the data-based model, we calculate the photoneutron production cross-sections and yields and compared them with experimental data. The results obtained from the developed model are found to be in good agreement with the experimental data for (γ,xn) reactions.

Keywords: ENDF/B-VII.1, GEANT4, photoneutron, photonuclear reaction

Procedia PDF Downloads 264
35024 Reduction in Hot Metal Silicon through Statistical Analysis at G-Blast Furnace, Tata Steel Jamshedpur

Authors: Shoumodip Roy, Ankit Singhania, Santanu Mallick, Abhiram Jha, M. K. Agarwal, R. V. Ramna, Uttam Singh

Abstract:

The quality of hot metal at any blast furnace is judged by the silicon content in it. Lower hot metal silicon not only enhances process efficiency at steel melting shops but also reduces hot metal costs. The Hot metal produced at G-Blast furnace Tata Steel Jamshedpur has a significantly higher Si content than Benchmark Blast furnaces. The higher content of hot metal Si is mainly due to inferior raw material quality than those used in benchmark blast furnaces. With minimum control over raw material quality, the only option left to control hot metal Si is via optimizing the furnace parameters. Therefore, in order to identify the levers to reduce hot metal Si, Data mining was carried out, and multiple regression models were developed. The statistical analysis revealed that Slag B3{(CaO+MgO)/SiO2}, Slag Alumina and Hot metal temperature are key controllable parameters affecting hot metal silicon. Contour Plots were used to determine the optimum range of levels identified through statistical analysis. A trial plan was formulated to operate relevant parameters, at G blast furnace, in the identified range to reduce hot metal silicon. This paper details out the process followed and subsequent reduction in hot metal silicon by 15% at G blast furnace.

Keywords: blast furnace, optimization, silicon, statistical tools

Procedia PDF Downloads 206
35023 Comparison of Different Reanalysis Products for Predicting Extreme Precipitation in the Southern Coast of the Caspian Sea

Authors: Parvin Ghafarian, Mohammadreza Mohammadpur Panchah, Mehri Fallahi

Abstract:

Synoptic patterns from surface up to tropopause are very important for forecasting the weather and atmospheric conditions. There are many tools to prepare and analyze these maps. Reanalysis data and the outputs of numerical weather prediction models, satellite images, meteorological radar, and weather station data are used in world forecasting centers to predict the weather. The forecasting extreme precipitating on the southern coast of the Caspian Sea (CS) is the main issue due to complex topography. Also, there are different types of climate in these areas. In this research, we used two reanalysis data such as ECMWF Reanalysis 5th Generation Description (ERA5) and National Centers for Environmental Prediction /National Center for Atmospheric Research (NCEP/NCAR) for verification of the numerical model. ERA5 is the latest version of ECMWF. The temporal resolution of ERA5 is hourly, and the NCEP/NCAR is every six hours. Some atmospheric parameters such as mean sea level pressure, geopotential height, relative humidity, wind speed and direction, sea surface temperature, etc. were selected and analyzed. Some different type of precipitation (rain and snow) was selected. The results showed that the NCEP/NCAR has more ability to demonstrate the intensity of the atmospheric system. The ERA5 is suitable for extract the value of parameters for specific point. Also, ERA5 is appropriate to analyze the snowfall events over CS (snow cover and snow depth). Sea surface temperature has the main role to generate instability over CS, especially when the cold air pass from the CS. Sea surface temperature of NCEP/NCAR product has low resolution near coast. However, both data were able to detect meteorological synoptic patterns that led to heavy rainfall over CS. However, due to the time lag, they are not suitable for forecast centers. The application of these two data is for research and verification of meteorological models. Finally, ERA5 has a better resolution, respect to NCEP/NCAR reanalysis data, but NCEP/NCAR data is available from 1948 and appropriate for long term research.

Keywords: synoptic patterns, heavy precipitation, reanalysis data, snow

Procedia PDF Downloads 103
35022 Predicting Daily Patient Hospital Visits Using Machine Learning

Authors: Shreya Goyal

Abstract:

The study aims to build user-friendly software to understand patient arrival patterns and compute the number of potential patients who will visit a particular health facility for a given period by using a machine learning algorithm. The underlying machine learning algorithm used in this study is the Support Vector Machine (SVM). Accurate prediction of patient arrival allows hospitals to operate more effectively, providing timely and efficient care while optimizing resources and improving patient experience. It allows for better allocation of staff, equipment, and other resources. If there's a projected surge in patients, additional staff or resources can be allocated to handle the influx, preventing bottlenecks or delays in care. Understanding patient arrival patterns can also help streamline processes to minimize waiting times for patients and ensure timely access to care for patients in need. Another big advantage of using this software is adhering to strict data protection regulations such as the Health Insurance Portability and Accountability Act (HIPAA) in the United States as the hospital will not have to share the data with any third party or upload it to the cloud because the software can read data locally from the machine. The data needs to be arranged in. a particular format and the software will be able to read the data and provide meaningful output. Using software that operates locally can facilitate compliance with these regulations by minimizing data exposure. Keeping patient data within the hospital's local systems reduces the risk of unauthorized access or breaches associated with transmitting data over networks or storing it in external servers. This can help maintain the confidentiality and integrity of sensitive patient information. Historical patient data is used in this study. The input variables used to train the model include patient age, time of day, day of the week, seasonal variations, and local events. The algorithm uses a Supervised learning method to optimize the objective function and find the global minima. The algorithm stores the values of the local minima after each iteration and at the end compares all the local minima to find the global minima. The strength of this study is the transfer function used to calculate the number of patients. The model has an output accuracy of >95%. The method proposed in this study could be used for better management planning of personnel and medical resources.

Keywords: machine learning, SVM, HIPAA, data

Procedia PDF Downloads 53
35021 A Semi-Automated GIS-Based Implementation of Slope Angle Design Reconciliation Process at Debswana Jwaneng Mine, Botswana

Authors: K. Mokatse, O. M. Barei, K. Gabanakgosi, P. Matlhabaphiri

Abstract:

The mining of pit slopes is often associated with some level of deviation from design recommendations, and this may translate to associated changes in the stability of the excavated pit slopes. Therefore slope angle design reconciliations are essential for assessing and monitoring compliance of excavated pit slopes to accepted slope designs. These associated changes in slope stability may be reflected by changes in the calculated factors of safety and/or probabilities of failure. Reconciliations of as-mined and slope design profiles are conducted periodically to assess the implications of these deviations on pit slope stability. Currently, the slope design reconciliation process being implemented in Jwaneng Mine involves the measurement of as-mined and design slope angles along vertical sections cut along the established geotechnical design section lines on the GEOVIA GEMS™ software. Bench retentions are calculated as a percentage of the available catchment area, less over-mined and under-mined areas, to that of the designed catchment area. This process has proven to be both tedious and requires a lot of manual effort and time to execute. Consequently, a new semi-automated mine-to-design reconciliation approach that utilizes laser scanning and GIS-based tools is being proposed at Jwaneng Mine. This method involves high-resolution scanning of targeted bench walls, subsequent creation of 3D surfaces from point cloud data and the derivation of slope toe lines and crest lines on the Maptek I-Site Studio software. The toe lines and crest lines are then exported to the ArcGIS software where distance offsets between the design and actual bench toe lines and crest lines are calculated. Retained bench catchment capacity is measured as distances between the toe lines and crest lines on the same bench elevations. The assessment of the performance of the inter-ramp and overall slopes entails the measurement of excavated and design slope angles along vertical sections on the ArcGIS software. Excavated and design toe-to-toe or crest-to-crest slope angles are measured for inter-ramp stack slope reconciliations. Crest-to-toe slope angles are also measured for overall slope angle design reconciliations. The proposed approach allows for a more automated, accurate, quick and easier workflow for carrying out slope angle design reconciliations. This process has proved highly effective and timeous in the assessment of slope performance in Jwaneng Mine. This paper presents a newly proposed process for assessing compliance to slope angle designs for Jwaneng Mine.

Keywords: slope angle designs, slope design recommendations, slope performance, slope stability

Procedia PDF Downloads 211
35020 Assessment of the Impacts of Climate Change on Watershed Runoff Using Soil and Water Assessment Tool Model in Southeast Nigeria

Authors: Samuel Emeka Anarah, Kingsley Nnaemeka Ogbu, Obasi Arinze

Abstract:

Quantifying the hydrological response due to changes in climate change is imperative for proper management of water resources within a watershed. The impact of climate change on the hydrology of the Upper Ebony River (UER) watershed, South East Nigeria, was studied using the Soil and Water Assessment Tool (SWAT) hydrological model. A climatological time series analysis from 1985 - 2014 using non-parametric test showed significant negative trends in precipitation and relative humidity trend while minimum and maximum temperature, solar radiation and wind speed showed significant positive trends. Future hypothetical land-use change scenarios (Scenarios 1, 2, 3 and 4) representing urbanization and conversion of forest to agricultural land were combined with future downscaled climate model (CSIRO-Mk3-6-0) and simulated in SWAT model. Relative to the Baseline scenario (2005 - 2014), the results showed a decrease in streamflow by 10.29%, 26.20%, 11.80% and 26.72% for Scenarios 1, 2, 3, and 4 respectively. Model results suggest development of adaptation strategies to cope with the predicted hydrological conditions under future climate change in the watershed.

Keywords: climate change, hydrology, runoff, SWAT model

Procedia PDF Downloads 125
35019 Proposal of Commutation Protocol in Hybrid Sensors and Vehicular Networks for Intelligent Transport Systems

Authors: Taha Bensiradj, Samira Moussaoui

Abstract:

Hybrid Sensors and Vehicular Networks (HSVN), represent a hybrid network, which uses several generations of Ad-Hoc networks. It is used especially in Intelligent Transport Systems (ITS). The HSVN allows making collaboration between the Wireless Sensors Network (WSN) deployed on the border of the road and the Vehicular Network (VANET). This collaboration is defined by messages exchanged between the two networks for the purpose to inform the drivers about the state of the road, provide road safety information and more information about traffic on the road. Moreover, this collaboration created by HSVN, also allows the use of a network and the advantage of improving another network. For example, the dissemination of information between the sensors quickly decreases its energy, and therefore, we can use vehicles that do not have energy constraint to disseminate the information between sensors. On the other hand, to solve the disconnection problem in VANET, the sensors can be used as gateways that allow sending the messages received by one vehicle to another. However, because of the short communication range of the sensor and its low capacity of storage and processing of data, it is difficult to ensure the exchange of road messages between it and the vehicle, which can be moving at high speed at the time of exchange. This represents the time where the vehicle is in communication range with the sensor. This work is the proposition of a communication protocol between the sensors and the vehicle used in HSVN. The latter has as the purpose to ensure the exchange of road messages in the available time of exchange.

Keywords: HSVN, ITS, VANET, WSN

Procedia PDF Downloads 347
35018 The Determinants of Trade Flow and Potential between Ethiopia and Group of Twenty

Authors: Terefe Alemu

Abstract:

This study is intended to examine Ethiopia’s trade flow determinants and trade potential with G20 countries whether it was overtraded or there is/are trade potential by using trade gravity model. The sources of panel data used were IMF, WDI, United Nations population division, The Heritage Foundation, Washington's No. 1 think tank online website database, online distance calculator, and others for the duration of 2010 to 2019 for 10 consecutive years. The empirical data analyzing tool used was Random effect model (REM), which is effective in estimation of time-invariant data. The empirical data analyzed using STATA software result indicates that Ethiopia has a trade potential with seven countries of G20, whereas Ethiopia overtrade with 12 countries and EU region. The Ethiopia’s and G20 countries/region bilateral trade flow statistically significant/ p<0.05/determinants were the population of G20 countries, growth domestic products of G20 countries, growth domestic products of Ethiopia, geographical distance between Ethiopia and G20 countries. The top five G20 countries exported to Ethiopia were china, United State of America, European Union, India, and South Africa, whereas the top five G20 countries imported from Ethiopia were EU, China, United State of America, Saudi Arabia, and Germany, respectively. Finally, the policy implication were Ethiopia has to Keep the consistence of trade flow with overtraded countries and improve with under traded countries through trade policy revision, and secondly, focusing on the trade determinants to improve trade flow is recommended.

Keywords: trade gravity model, trade determinants, G20, international trade, trade potential

Procedia PDF Downloads 189
35017 Asymmetric of the Segregation-Enhanced Brazil Nut Effect

Authors: Panupat Chaiworn, Soraya lama

Abstract:

We study the motion of particles in cylinders which are subjected to a sinusoidal vertical vibration. We measure the rising time of a large intruder from the bottom of the container to free surface of the bed particles and find that the rising time as a function of intruder density increases to a maximum and then decreases monotonically. The result is qualitatively accord to the previous findings in experiments using relative humidity of the bed particles and found speed convection of the bed particles containers it moving slowly, and the rising time of the intruder where a minimal instead of maximal rising time in the small density region was found. Our experimental results suggest that the topology of the container plays an important role in the Brazil nut effect.

Keywords: granular particles, Brazil nut effect, cylinder container, vertical vibration, convection

Procedia PDF Downloads 511
35016 Impact of Macroeconomic Variables on Indian Mutual Funds: A Time Series Analysis

Authors: Sonali Agarwal

Abstract:

The investor perception about investment avenues is affected to a great degree by the current happenings, within the country, and on the global stage. The influencing events can range from government policies, bilateral trade agreements, election agendas, to changing exchange rates, appreciation and depreciation of currency, recessions, meltdowns, bankruptcies etc. The current research attempts to discover and unravel the effect of various macroeconomic variables (crude oil price, gold price, silver price and USD exchange rate) on the Indian mutual fund industry in general and the chosen funds (Axis Gold Fund, BSL Gold Fund, Kotak Gold Fund & SBI gold fund) in particular. Cointegration tests and Vector error correction equations prove that the chosen variables have strong effect on the NAVs (net asset values) of the mutual funds. However, the greatest influence is felt from the fund’s own past and current information and it is found that when an innovation of fund’s own lagged NAVs is given, variance caused is high that changes the current NAVs markedly. The study helps to highlight the interplay of macroeconomic variables and their repercussion on mutual fund industry.

Keywords: cointegration, Granger causality, impulse response, macroeconomic variables, mutual funds, stationarity, unit root test, variance decomposition, VECM

Procedia PDF Downloads 230