Search results for: running time
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18031

Search results for: running time

17941 Improving Activity Recognition Classification of Repetitious Beginner Swimming Using a 2-Step Peak/Valley Segmentation Method with Smoothing and Resampling for Machine Learning

Authors: Larry Powell, Seth Polsley, Drew Casey, Tracy Hammond

Abstract:

Human activity recognition (HAR) systems have shown positive performance when recognizing repetitive activities like walking, running, and sleeping. Water-based activities are a reasonably new area for activity recognition. However, water-based activity recognition has largely focused on supporting the elite and competitive swimming population, which already has amazing coordination and proper form. Beginner swimmers are not perfect, and activity recognition needs to support the individual motions to help beginners. Activity recognition algorithms are traditionally built around short segments of timed sensor data. Using a time window input can cause performance issues in the machine learning model. The window’s size can be too small or large, requiring careful tuning and precise data segmentation. In this work, we present a method that uses a time window as the initial segmentation, then separates the data based on the change in the sensor value. Our system uses a multi-phase segmentation method that pulls all peaks and valleys for each axis of an accelerometer placed on the swimmer’s lower back. This results in high recognition performance using leave-one-subject-out validation on our study with 20 beginner swimmers, with our model optimized from our final dataset resulting in an F-Score of 0.95.

Keywords: time window, peak/valley segmentation, feature extraction, beginner swimming, activity recognition

Procedia PDF Downloads 85
17940 Optimization of Assay Parameters of L-Glutaminase from Bacillus cereus MTCC1305 Using Artificial Neural Network

Authors: P. Singh, R. M. Banik

Abstract:

Artificial neural network (ANN) was employed to optimize assay parameters viz., time, temperature, pH of reaction mixture, enzyme volume and substrate concentration of L-glutaminase from Bacillus cereus MTCC 1305. ANN model showed high value of coefficient of determination (0.9999), low value of root mean square error (0.6697) and low value of absolute average deviation. A multilayer perceptron neural network trained with an error back-propagation algorithm was incorporated for developing a predictive model and its topology was obtained as 5-3-1 after applying Levenberg Marquardt (LM) training algorithm. The predicted activity of L-glutaminase was obtained as 633.7349 U/l by considering optimum assay parameters, viz., pH of reaction mixture (7.5), reaction time (20 minutes), incubation temperature (35˚C), substrate concentration (40mM), and enzyme volume (0.5ml). The predicted data was verified by running experiment at simulated optimum assay condition and activity was obtained as 634.00 U/l. The application of ANN model for optimization of assay conditions improved the activity of L-glutaminase by 1.499 fold.

Keywords: Bacillus cereus, L-glutaminase, assay parameters, artificial neural network

Procedia PDF Downloads 399
17939 Real-Time Network Anomaly Detection Systems Based on Machine-Learning Algorithms

Authors: Zahra Ramezanpanah, Joachim Carvallo, Aurelien Rodriguez

Abstract:

This paper aims to detect anomalies in streaming data using machine learning algorithms. In this regard, we designed two separate pipelines and evaluated the effectiveness of each separately. The first pipeline, based on supervised machine learning methods, consists of two phases. In the first phase, we trained several supervised models using the UNSW-NB15 data-set. We measured the efficiency of each using different performance metrics and selected the best model for the second phase. At the beginning of the second phase, we first, using Argus Server, sniffed a local area network. Several types of attacks were simulated and then sent the sniffed data to a running algorithm at short intervals. This algorithm can display the results of each packet of received data in real-time using the trained model. The second pipeline presented in this paper is based on unsupervised algorithms, in which a Temporal Graph Network (TGN) is used to monitor a local network. The TGN is trained to predict the probability of future states of the network based on its past behavior. Our contribution in this section is introducing an indicator to identify anomalies from these predicted probabilities.

Keywords: temporal graph network, anomaly detection, cyber security, IDS

Procedia PDF Downloads 74
17938 Improved Particle Swarm Optimization with Cellular Automata and Fuzzy Cellular Automata

Authors: Ramin Javadzadeh

Abstract:

The particle swarm optimization are Meta heuristic optimization method, which are used for clustering and pattern recognition applications are abundantly. These algorithms in multimodal optimization problems are more efficient than genetic algorithms. A major drawback in these algorithms is their slow convergence to global optimum and their weak stability can be considered in various running of these algorithms. In this paper, improved Particle swarm optimization is introduced for the first time to overcome its problems. The fuzzy cellular automata is used for improving the algorithm efficiently. The credibility of the proposed approach is evaluated by simulations, and it is shown that the proposed approach achieves better results can be achieved compared to the Particle swarm optimization algorithms.

Keywords: cellular automata, cellular learning automata, local search, optimization, particle swarm optimization

Procedia PDF Downloads 571
17937 Load Comparison between Different Positions during Elite Male Basketball Games: A Sport Metabolomics Approach

Authors: Kayvan Khoramipour, Abbas Ali Gaeini, Elham Shirzad, Øyvind Sandbakk

Abstract:

Basketball has different positions with individual movement profiles, which may influence metabolic demands. Accordingly, the present study aimed to compare the movement and metabolic load between different positions during elite male basketball games. Five main players of 14 teams (n = 70), who participated in the 2017-18 Iranian national basketball leagues, were selected as participants. The players were defined as backcourt (Posts 1-3) and frontcourt (Posts 4-5). Video based time motion analysis (VBTMA) was performed based on players’ individual running and shuffling speed using Dartfish software. Movements were classified into high and low intensity running with and without having the ball, as well as high and low-intensity shuffling and static movements. Mean frequency, duration, and distance were calculated for each class, except for static movements where only frequency was calculated. Saliva samples were collected from each player before and after 40-minute basketball games and analyzed using metabolomics. Principal component analysis (PCA) and Partial least square discriminant analysis (PLSDA) (for metabolomics data) and independent T-tests (for VBTMA) were used as statistical tests. Movement frequency, duration, and distance were higher in backcourt players (all p ≤ 0.05), while static movement frequency did not differ. Saliva samples showed that the levels of Taurine, Succinic acid, Citric acid, Pyruvate, Glycerol, Acetoacetic acid, Acetone, and Hypoxanthine were all higher in backcourt players, whereas Lactate, Alanine, 3-Metyl Histidine, and Methionine were higher in frontcourt players Based on metabolomics, we demonstrate that backcourt and frontcourt players have different metabolic profiles during games, where backcourt players move clearly more during games and therefore rely more on aerobic energy, whereas frontcourt players rely more on anaerobic energy systems in line with less dynamic but more static movement patterns.

Keywords: basketball, metabolomics, saliva, sport loadomics

Procedia PDF Downloads 85
17936 A Methodology for Optimisation of Water Containment Systems

Authors: Amir Hedjripour

Abstract:

The required dewatering configuration for a contaminated sediment dam is discussed to meet no-spill criteria for a defined Average Recurrence Interval (ARI). There is an option for the sediment dam to pump the contaminated water to another storage facility before its capacity is exceeded. The system is subjected to a range of storm durations belonging to the design ARI with concurrent dewatering to the other storage facility. The model is set up in 1-minute time intervals and temporal patterns of storm events are used to de-segregate the total storm depth into partial durations. By running the model for selected storm durations, the maximum water volume in the dam is recorded as the critical volume, which indicates the required storage capacity for that storm duration. Runoff from upstream catchment and the direct rainfall over the dam open area are calculated by taking into account the time of concentration for the catchment. Total 99 different storm durations from 5 minutes to 72 hours were modelled together with five dewatering scenarios from 50 l/s to 500 l/s. The optimised dam/pump configuration is selected by plotting critical points for all cases and storage-dewatering envelopes. A simple economic analysis is also presented in the paper using Present-Value (PV) analysis to assist with the financial evaluation of each configuration and selection of the best alternative.

Keywords: contaminated water, optimisation, pump, sediment dam

Procedia PDF Downloads 329
17935 MCERTL: Mutation-Based Correction Engine for Register-Transfer Level Designs

Authors: Khaled Salah

Abstract:

In this paper, we present MCERTL (mutation-based correction engine for RTL designs) as an automatic error correction technique based on mutation analysis. A mutation-based correction methodology is proposed to automatically fix the erroneous RTL designs. The proposed strategy combines the processes of mutation and assertion-based localization. The erroneous statements are mutated to produce possible fixes for the failed RTL code. A concurrent mutation engine is proposed to mitigate the computational cost of running sequential mutants operators. The proposed methodology is evaluated against some benchmarks. The experimental results demonstrate that our proposed method enables us to automatically locate and correct multiple bugs at reasonable time.

Keywords: bug localization, error correction, mutation, mutants

Procedia PDF Downloads 248
17934 Design of Low Latency Multiport Network Router on Chip

Authors: P. G. Kaviya, B. Muthupandian, R. Ganesan

Abstract:

On-chip routers typically have buffers are used input or output ports for temporarily storing packets. The buffers are consuming some router area and power. The multiple queues in parallel as in VC router. While running a traffic trace, not all input ports have incoming packets needed to be transferred. Therefore large numbers of queues are empty and others are busy in the network. So the time consumption should be high for the high traffic. Therefore using a RoShaQ, minimize the buffer area and time The RoShaQ architecture was send the input packets are travel through the shared queues at low traffic. At high load traffic the input packets are bypasses the shared queues. So the power and area consumption was reduced. A parallel cross bar architecture is proposed in this project in order to reduce the power consumption. Also a new adaptive weighted routing algorithm for 8-port router architecture is proposed in order to decrease the delay of the network on chip router. The proposed system is simulated using Modelsim and synthesized using Xilinx Project Navigator.

Keywords: buffer, RoShaQ architecture, shared queue, VC router, weighted routing algorithm

Procedia PDF Downloads 521
17933 Springback Prediction for Sheet Metal Cold Stamping Using Convolutional Neural Networks

Authors: Lei Zhu, Nan Li

Abstract:

Cold stamping has been widely applied in the automotive industry for the mass production of a great range of automotive panels. Predicting the springback to ensure the dimensional accuracy of the cold-stamped components is a critical step. The main approaches for the prediction and compensation of springback in cold stamping include running Finite Element (FE) simulations and conducting experiments, which require forming process expertise and can be time-consuming and expensive for the design of cold stamping tools. Machine learning technologies have been proven and successfully applied in learning complex system behaviours using presentative samples. These technologies exhibit the promising potential to be used as supporting design tools for metal forming technologies. This study, for the first time, presents a novel application of a Convolutional Neural Network (CNN) based surrogate model to predict the springback fields for variable U-shape cold bending geometries. A dataset is created based on the U-shape cold bending geometries and the corresponding FE simulations results. The dataset is then applied to train the CNN surrogate model. The result shows that the surrogate model can achieve near indistinguishable full-field predictions in real-time when compared with the FE simulation results. The application of CNN in efficient springback prediction can be adopted in industrial settings to aid both conceptual and final component designs for designers without having manufacturing knowledge.

Keywords: springback, cold stamping, convolutional neural networks, machine learning

Procedia PDF Downloads 121
17932 Using High Performance Computing for Online Flood Monitoring and Prediction

Authors: Stepan Kuchar, Martin Golasowski, Radim Vavrik, Michal Podhoranyi, Boris Sir, Jan Martinovic

Abstract:

The main goal of this article is to describe the online flood monitoring and prediction system Floreon+ primarily developed for the Moravian-Silesian region in the Czech Republic and the basic process it uses for running automatic rainfall-runoff and hydrodynamic simulations along with their calibration and uncertainty modeling. It takes a long time to execute such process sequentially, which is not acceptable in the online scenario, so the use of high-performance computing environment is proposed for all parts of the process to shorten their duration. Finally, a case study on the Ostravice river catchment is presented that shows actual durations and their gain from the parallel implementation.

Keywords: flood prediction process, high performance computing, online flood prediction system, parallelization

Procedia PDF Downloads 463
17931 Runaway Girl Children and the Reasons: Qualitative Study in Government Girls Home Bangalore

Authors: Hazel Johanna J., Ntailang Mary Tariang

Abstract:

The paper “Runaway Girl Children and the Reasons: Qualitative Study In Government Girls Home Bangalore” explores the different reasons why children choose this last resort of running away rather than seeking proper help from the authorities. A qualitative study using a purposive sampling method was used to identify the participants based on the objectives. Girl runaway children between the age group of 12-18 years admitted to the Government Girls Home, Bangalore, were chosen for this study. Data was collected through in-depth interviews using semi-structured questions. Thematic analysis has been done using QDA Miner Lite. The main objectives of this study were to identify the reasons behind running away in children, to explore their childhood experiences and future dreams after they leave the Child Care Institution. The findings of this study derived five major themes that have caused the children to run away from their homes. The themes are child maltreatment and dysfunctional families, coerced into adulthood, forced work, adolescent dalliance, and aspirations. As a result, all the themes that emerged here are related to the family in one way or another. In conclusion, it is revealed that interpersonal family conflicts lead to the violation of child rights in so many ways, which in this context leads the child to run away from the comfort of their home.

Keywords: runaway children, dysfunctional family, abuse, child marriage, education

Procedia PDF Downloads 43
17930 Predicting Blockchain Technology Installation Cost in Supply Chain System through Supervised Learning

Authors: Hossein Havaeji, Tony Wong, Thien-My Dao

Abstract:

1. Research Problems and Research Objectives: Blockchain Technology-enabled Supply Chain System (BT-enabled SCS) is the system using BT to drive SCS transparency, security, durability, and process integrity as SCS data is not always visible, available, or trusted. The costs of operating BT in the SCS are a common problem in several organizations. The costs must be estimated as they can impact existing cost control strategies. To account for system and deployment costs, it is necessary to overcome the following hurdle. The problem is that the costs of developing and running a BT in SCS are not yet clear in most cases. Many industries aiming to use BT have special attention to the importance of BT installation cost which has a direct impact on the total costs of SCS. Predicting BT installation cost in SCS may help managers decide whether BT is to be an economic advantage. The purpose of the research is to identify some main BT installation cost components in SCS needed for deeper cost analysis. We then identify and categorize the main groups of cost components in more detail to utilize them in the prediction process. The second objective is to determine the suitable Supervised Learning technique in order to predict the costs of developing and running BT in SCS in a particular case study. The last aim is to investigate how the running BT cost can be involved in the total cost of SCS. 2. Work Performed: Applied successfully in various fields, Supervised Learning is a method to set the data frame, treat the data, and train/practice the method sort. It is a learning model directed to make predictions of an outcome measurement based on a set of unforeseen input data. The following steps must be conducted to search for the objectives of our subject. The first step is to make a literature review to identify the different cost components of BT installation in SCS. Based on the literature review, we should choose some Supervised Learning methods which are suitable for BT installation cost prediction in SCS. According to the literature review, some Supervised Learning algorithms which provide us with a powerful tool to classify BT installation components and predict BT installation cost are the Support Vector Regression (SVR) algorithm, Back Propagation (BP) neural network, and Artificial Neural Network (ANN). Choosing a case study to feed data into the models comes into the third step. Finally, we will propose the best predictive performance to find the minimum BT installation costs in SCS. 3. Expected Results and Conclusion: This study tends to propose a cost prediction of BT installation in SCS with the help of Supervised Learning algorithms. At first attempt, we will select a case study in the field of BT-enabled SCS, and then use some Supervised Learning algorithms to predict BT installation cost in SCS. We continue to find the best predictive performance for developing and running BT in SCS. Finally, the paper will be presented at the conference.

Keywords: blockchain technology, blockchain technology-enabled supply chain system, installation cost, supervised learning

Procedia PDF Downloads 95
17929 Memory, Self, and Time: A Bachelardian Perspective

Authors: Michael Granado

Abstract:

The French philosopher Gaston Bachelard’s philosophy of time is articulated in his two works on the subject, the Intuition of the Instant (1932) and his The Dialectic of Duration (1936). Both works present a systematic methodology predicated upon the assumption that our understanding of time has radically changed as a result of Einstein and subsequently needs to be reimagined. Bachelard makes a major distinction in his discussion of time: 1. Time as it is (physical time), 2. Time as we experience it (phenomenological time). This paper will focus on the second distinction, phenomenological time, and explore the connections between Bachelard’s work and contemporary psychology. Several aspects of Bachelard’s philosophy of time nicely complement our current understanding of memory and self and clarify how the self relates to experienced time. Two points, in particular, stand out; the first is the relative nature of subjective time, and the second is the implications of subjective time in the formation of the narrative self. Bachelard introduces two philosophical concepts to explain these points: rhythmanalysis and reverie. By exploring these concepts, it will become apparent that there is an undeniable link between memory, self, and time. Through the use of narrative self, the individual connects and links memories and time together to form a sense of personal identity.

Keywords: Gaston Bachelard, memory, self, time

Procedia PDF Downloads 134
17928 Modified Approximation Methods for Finding an Optimal Solution for the Transportation Problem

Authors: N. Guruprasad

Abstract:

This paper presents a modification of approximation method for transportation problems. The initial basic feasible solution can be computed using either Russel's or Vogel's approximation methods. Russell’s approximation method provides another excellent criterion that is still quick to implement on a computer (not manually) In most cases Russel's method yields a better initial solution, though it takes longer than Vogel's method (finding the next entering variable in Russel's method is in O(n1*n2), and in O(n1+n2) for Vogel's method). However, Russel's method normally has a lesser total running time because less pivots are required to reach the optimum for all but small problem sizes (n1+n2=~20). With this motivation behind we have incorporated a variation of the same – what we have proposed it has TMC (Total Modified Cost) to obtain fast and efficient solutions.

Keywords: computation, efficiency, modified cost, Russell’s approximation method, transportation, Vogel’s approximation method

Procedia PDF Downloads 501
17927 Adaptive Swarm Balancing Algorithms for Rare-Event Prediction in Imbalanced Healthcare Data

Authors: Jinyan Li, Simon Fong, Raymond Wong, Mohammed Sabah, Fiaidhi Jinan

Abstract:

Clinical data analysis and forecasting have make great contributions to disease control, prevention and detection. However, such data usually suffer from highly unbalanced samples in class distributions. In this paper, we target at the binary imbalanced dataset, where the positive samples take up only the minority. We investigate two different meta-heuristic algorithms, particle swarm optimization and bat-inspired algorithm, and combine both of them with the synthetic minority over-sampling technique (SMOTE) for processing the datasets. One approach is to process the full dataset as a whole. The other is to split up the dataset and adaptively process it one segment at a time. The experimental results reveal that while the performance improvements obtained by the former methods are not scalable to larger data scales, the later one, which we call Adaptive Swarm Balancing Algorithms, leads to significant efficiency and effectiveness improvements on large datasets. We also find it more consistent with the practice of the typical large imbalanced medical datasets. We further use the meta-heuristic algorithms to optimize two key parameters of SMOTE. Leading to more credible performances of the classifier, and shortening the running time compared with the brute-force method.

Keywords: Imbalanced dataset, meta-heuristic algorithm, SMOTE, big data

Procedia PDF Downloads 414
17926 The Relationships between Carbon Dioxide (CO2) Emissions, Energy Consumption and GDP per capita for Oman: Time Series Analysis, 1980–2010

Authors: Jinhoa Lee

Abstract:

The relationships between environmental quality, energy use and economic output have created growing attention over the past decades among researchers and policy makers. Focusing on the empirical aspects of the role of CO2 emissions and energy use in affecting the economic output, this paper is an effort to fulfil the gap in a comprehensive case study at a country level using modern econometric techniques. To achieve the goal, this country-specific study examines the short-run and long-run relationships among energy consumption, carbon dioxide (CO2) emissions and gross domestic product (GDP) for Oman using time series analysis from the year 1980-2010. To investigate the relationships between the variables, this paper employs the Augmented Dickey Fuller (ADF) test for stationary, Johansen maximum likelihood method for co-integration and a Vector Error Correction Model (VECM) for both short- and long-run causality among the research variables for the sample. All the variables in this study show very strong significant effects on GDP in the country for the long term. The long-run equilibrium in the VECM suggests positive long-run causalities from CO2 emissions to GDP. Conversely, negative impacts of energy consumption on GDP are found to be significant in Oman during the period. In the short run, there exist negative unidirectional causalities among GDP, CO2 emissions and energy consumption running from GDP to CO2 emissions and from energy consumption to CO2 emissions. Overall, the results support arguments that there are relationships among environmental quality, energy use and economic output in Oman over of period 1980-2010.

Keywords: CO2 emissions, energy consumption, GDP, Oman, time series analysis

Procedia PDF Downloads 434
17925 The Relationships between Energy Consumption, Carbon Dioxide (CO2) Emissions, and GDP for Turkey: Time Series Analysis, 1980-2010

Authors: Jinhoa Lee

Abstract:

The relationships between environmental quality, energy use and economic output have created growing attention over the past decades among researchers and policy makers. Focusing on the empirical aspects of the role of carbon dioxide (CO2) emissions and energy use in affecting the economic output, this paper is an effort to fulfill the gap in a comprehensive case study at a country level using modern econometric techniques. To achieve the goal, this country-specific study examines the short-run and long-run relationships among energy consumption (using disaggregated energy sources: crude oil, coal, natural gas, and electricity), CO2 emissions and gross domestic product (GDP) for Turkey using time series analysis from the year 1980-2010. To investigate the relationships between the variables, this paper employs the Augmented Dickey-Fuller (ADF) test for stationarity, Johansen’s maximum likelihood method for cointegration and a Vector Error Correction Model (VECM) for both short- and long-run causality among the research variables for the sample. The long-run equilibrium in the VECM suggests no effects of the CO2 emissions and energy use on the GDP in Turkey. There exists a short-run bidirectional relationship between the electricity and natural gas consumption, and also there is a negative unidirectional causality running from the GDP to electricity use. Overall, the results partly support arguments that there are relationships between energy use and economic output; however, the effects may differ due to the source of energy such as in the case of Turkey for the period of 1980-2010. However, there is no significant relationship between the CO2 emissions and the GDP and between the CO2 emissions and the energy use both in the short term and long term.

Keywords: CO2 emissions, energy consumption, GDP, Turkey, time series analysis

Procedia PDF Downloads 485
17924 An Experimental Study on the Optimum Installation of Fire Detector for Early Stage Fire Detecting in Rack-Type Warehouses

Authors: Ki Ok Choi, Sung Ho Hong, Dong Suck Kim, Don Mook Choi

Abstract:

Rack type warehouses are different from general buildings in the kinds, amount, and arrangement of stored goods, so the fire risk of rack type warehouses is different from those buildings. The fire pattern of rack type warehouses is different in combustion characteristic and storing condition of stored goods. The initial fire burning rate is different in the surface condition of materials, but the running time of fire is closely related with the kinds of stored materials and stored conditions. The stored goods of the warehouse are consisted of diverse combustibles, combustible liquid, and so on. Fire detection time may be delayed because the residents are less than office and commercial buildings. If fire detectors installed in rack type warehouses are inadaptable, the fire of the warehouse may be the great fire because of delaying of fire detection. In this paper, we studied what kinds of fire detectors are optimized in early detecting of rack type warehouse fire by real-scale fire tests. The fire detectors used in the tests are rate of rise type, fixed type, photo electric type, and aspirating type detectors. We considered optimum fire detecting method in rack type warehouses suggested by the response characteristic and comparative analysis of the fire detectors.

Keywords: fire detector, rack, response characteristic, warehouse

Procedia PDF Downloads 713
17923 Enhance Engineering Learning Using Cognitive Simulator

Authors: Lior Davidovitch

Abstract:

Traditional training based on static models and case studies is the backbone of most teaching and training programs of engineering education. However, project management learning is characterized by dynamics models that requires new and enhanced learning method. The results of empirical experiments evaluating the effectiveness and efficiency of using cognitive simulator as a new training technique are reported. The empirical findings are focused on the impact of keeping and reviewing learning history in a dynamic and interactive simulation environment of engineering education. The cognitive simulator for engineering project management learning had two learning history keeping modes: manual (student-controlled), automatic (simulator-controlled) and a version with no history keeping. A group of industrial engineering students performed four simulation-runs divided into three identical simple scenarios and one complicated scenario. The performances of participants running the simulation with the manual history mode were significantly better than users running the simulation with the automatic history mode. Moreover, the effects of using the undo enhanced further the learning process. The findings indicate an enhancement of engineering students’ learning and decision making when they use the record functionality of the history during their engineering training process. Furthermore, the cognitive simulator as educational innovation improves students learning and training. The practical implications of using simulators in the field of engineering education are discussed.

Keywords: cognitive simulator, decision making, engineering learning, project management

Procedia PDF Downloads 219
17922 Automatic Registration of Rail Profile Based Local Maximum Curvature Entropy

Authors: Hao Wang, Shengchun Wang, Weidong Wang

Abstract:

On the influence of train vibration and environmental noise on the measurement of track wear, we proposed a method for automatic extraction of circular arc on the inner or outer side of the rail waist and achieved the high-precision registration of rail profile. Firstly, a polynomial fitting method based on truncated residual histogram was proposed to find the optimal fitting curve of the profile and reduce the influence of noise on profile curve fitting. Then, based on the curvature distribution characteristics of the fitting curve, the interval search algorithm based on dynamic window’s maximum curvature entropy was proposed to realize the automatic segmentation of small circular arc. At last, we fit two circle centers as matching reference points based on small circular arcs on both sides and realized the alignment from the measured profile to the standard designed profile. The static experimental results show that the mean and standard deviation of the method are controlled within 0.01mm with small measurement errors and high repeatability. The dynamic test also verified the repeatability of the method in the train-running environment, and the dynamic measurement deviation of rail wear is within 0.2mm with high repeatability.

Keywords: curvature entropy, profile registration, rail wear, structured light, train-running

Procedia PDF Downloads 229
17921 Automatic Fluid-Structure Interaction Modeling and Analysis of Butterfly Valve Using Python Script

Authors: N. Guru Prasath, Sangjin Ma, Chang-Wan Kim

Abstract:

A butterfly valve is a quarter turn valve which is used to control the flow of a fluid through a section of pipe. Generally, butterfly valve is used in wide range of applications such as water distribution, sewage, oil and gas plants. In particular, butterfly valve with larger diameter finds its immense applications in hydro power plants to control the fluid flow. In-lieu with the constraints in cost and size to run laboratory setup, analysis of large diameter values will be mostly studied by computational method which is the best and inexpensive solution. For fluid and structural analysis, CFD and FEM software is used to perform large scale valve analyses, respectively. In order to perform above analysis in butterfly valve, the CAD model has to recreate and perform mesh in conventional software’s for various dimensions of valve. Therefore, its limitation is time consuming process. In-order to overcome that issue, python code was created to outcome complete pre-processing setup automatically in Salome software. Applying dimensions of the model clearly in the python code makes the running time comparatively lower and easier way to perform analysis of the valve. Hence, in this paper, an attempt was made to study the fluid-structure interaction (FSI) of butterfly valves by varying the valve angles and dimensions using python code in pre-processing software, and results are produced.

Keywords: butterfly valve, flow coefficient, automatic CFD analysis, FSI analysis

Procedia PDF Downloads 211
17920 Flushing Model for Artificial Islands in the Persian Gulf

Authors: Sawsan Eissa, Momen Gharib, Omnia Kabbany

Abstract:

A flushing numerical study has been performed for intended artificial islands on the Persian Gulf coast in Abu Dhabi, UAE. The island masterplan was tested for flushing using the DELFT 3D hydrodynamic model, and it was found that its residence time exceeds the acceptable PIANC flushing Criteria. Therefore, a number of mitigation measures were applied and tested one by one using the flushing model. Namely, changing the location of the entrance opening, dredging, removing part of the mangrove existing in the near vicinity to create a channel, removing the mangrove altogether, using culverts of different numbers and locations, and pumping at selected points. The pumping option gave the best solution, but it was disregarded due to high capital and running costs. Therefore, it opted for a combination of other solutions, including removing mangroves, introducing culverts, and adjusting island boundaries and types of protection.

Keywords: hydrodynamics, flushing, delft 3d, Persian Gulf, artificial islands.

Procedia PDF Downloads 23
17919 The Accuracy of an 8-Minute Running Field Test to Estimate Lactate Threshold

Authors: Timothy Quinn, Ronald Croce, Aliaksandr Leuchanka, Justin Walker

Abstract:

Many endurance athletes train at or just below an intensity associated with their lactate threshold (LT) and often the heart rate (HR) that these athletes use for their LT are above their true LT-HR measured in a laboratory. Training above their true LT-HR may lead to overtraining and injury. Few athletes have the capability of measuring their LT in a laboratory and rely on perception to guide them, as accurate field tests to determine LT are limited. Therefore, the purpose of this study was to determine if an 8-minute field test could accurately define the HR associated with LT as measured in the laboratory. On Day 1, fifteen male runners (mean±SD; age, 27.8±4.1 years; height, 177.9±7.1 cm; body mass, 72.3±6.2 kg; body fat, 8.3±3.1%) performed a discontinuous treadmill LT/maximal oxygen consumption (LT/VO2max) test using a portable metabolic gas analyzer (Cosmed K4b2) and a lactate analyzer (Analox GL5). The LT (and associated HR) was determined using the 1/+1 method, where blood lactate increased by 1 mMol•L-1 over baseline followed by an additional 1 mMol•L-1 increase. Days 2 and 3 were randomized, and the athletes performed either an 8-minute run on the treadmill (TM) or on a 160-m indoor track (TR) in an effort to cover as much distance as possible while maintaining a high intensity throughout the entire 8 minutes. VO2, HR, ventilation (VE), and respiratory exchange ratio (RER) were measured using the Cosmed system, and rating of perceived exertion (RPE; 6-20 scale) was recorded every minute. All variables were averaged over the 8 minutes. The total distance covered over the 8 minutes was measured in both conditions. At the completion of the 8-minute runs, blood lactate was measured. Paired sample t-tests and pairwise Pearson correlations were computed to determine the relationship between variables measured in the field tests versus those obtained in the laboratory at LT. An alpha level of <0.05 was required for statistical significance. The HR (mean +SD) during the TM (167+9 bpm) and TR (172+9 bpm) tests were strongly correlated to the HR measured during the laboratory LT (169+11 bpm) test (r=0.68; p<0.03 and r=0.88; p<0.001, respectively). Blood lactate values during the TM and TR tests were not different from each other but were strongly correlated with the laboratory LT (r=0.73; p<0.04 and r=0.66; p<0.05, respectively). VE (Lmin-1) was significantly greater during the TR (134.8+11.4 Lmin-1) as compared to the TM (123.3+16.2 Lmin-1) with moderately strong correlations to the laboratory threshold values (r=0.38; p=0.27 and r=0.58; p=0.06, respectively). VO2 was higher during TR (51.4 mlkg-1min-1) compared to TM (47.4 mlkg-1min-1) with correlations of 0.33 (p=0.35) and 0.48 (p=0.13), respectively to threshold values. Total distance run was significantly greater during the TR (2331.6+180.9 m) as compared to the TM (2177.0+232.6 m), but they were strongly correlated with each other (r=0.82; p<0.002). These results suggest that an 8-minute running field test can accurately predict the HR associated with the LT and may be a simple test that athletes and coaches could implement to aid in training techniques.

Keywords: blood lactate, heart rate, running, training

Procedia PDF Downloads 226
17918 Scalable UI Test Automation for Large-scale Web Applications

Authors: Kuniaki Kudo, Raviraj Solanki, Kaushal Patel, Yash Virani

Abstract:

This research mainly concerns optimizing UI test automation for large-scale web applications. The test target application is the HHAexchange homecare management WEB application that seamlessly connects providers, state Medicaid programs, managed care organizations (MCOs), and caregivers through one platform with large-scale functionalities. This study focuses on user interface automation testing for the WEB application. The quality assurance team must execute many manual users interface test cases in the development process to confirm no regression bugs. The team automated 346 test cases; the UI automation test execution time was over 17 hours. The business requirement was reducing the execution time to release high-quality products quickly, and the quality assurance automation team modernized the test automation framework to optimize the execution time. The base of the WEB UI automation test environment is Selenium, and the test code is written in Python. Adopting a compilation language to write test code leads to an inefficient flow when introducing scalability into a traditional test automation environment. In order to efficiently introduce scalability into Test Automation, a scripting language was adopted. The scalability implementation is mainly implemented with AWS's serverless technology, an elastic container service. The definition of scalability here is the ability to automatically set up computers to test automation and increase or decrease the number of computers running those tests. This means the scalable mechanism can help test cases run parallelly. Then test execution time is dramatically decreased. Also, introducing scalable test automation is for more than just reducing test execution time. There is a possibility that some challenging bugs are detected by introducing scalable test automation, such as race conditions, Etc. since test cases can be executed at same timing. If API and Unit tests are implemented, the test strategies can be adopted more efficiently for this scalability testing. However, in WEB applications, as a practical matter, API and Unit testing cannot cover 100% functional testing since they do not reach front-end codes. This study applied a scalable UI automation testing strategy to the large-scale homecare management system. It confirmed the optimization of the test case execution time and the detection of a challenging bug. This study first describes the detailed architecture of the scalable test automation environment, then describes the actual performance reduction time and an example of challenging issue detection.

Keywords: aws, elastic container service, scalability, serverless, ui automation test

Procedia PDF Downloads 66
17917 SMEs Access to Finance in Croatia – Model Approach

Authors: Vinko Vidučić, Ljiljana Vidučić, Damir Boras

Abstract:

The goals of the research include the determination of the characteristics of SMEs finance in Croatia, as well as the determination of indirect growth rates of the information model of the entrepreneurs` perception of business environment. The research results show that cost of finance and access to finance are most important constraining factor in setting up and running the business of small entrepreneurs in Croatia. Furthermore, small entrepreneurs in Croatia are significantly dissatisfied with the administrative barriers although relatively to a lesser extent than was the case in the pre-crisis time. High collateral requirement represents the main characteristic of bank lending concerning SMEs followed by long credit elaboration process. Formulated information model has defined the individual impact of indirect growth rates of the remaining variables on the model’s specific variable.

Keywords: business environment, information model, indirect growth rates, SME finance

Procedia PDF Downloads 322
17916 Business Process Mashup

Authors: Fethia Zenak, Salima Benbernou, Linda Zaoui

Abstract:

Recently, many companies are based on process development from scratch to achieve their business goals. The process development is not trivial and the main objective of enterprise managing processes is to decrease the software development time. Several concepts have been proposed in the field of business process-based reused development, known as BP Mashup. This concept consists of reusing existing business processes which have been modeled in order to respond to a particular goal. To meet user process requirements, our contribution is to mix parts of processes as 'processes fragments' components to build a new process (i.e. process mashup). The main idea of our paper is to offer graphical framework tool for both creating and running processes mashup. Allow users to perform a mixture of fragments, using a simple interface with set of graphical mixture operators based on a proposed formal model. A process mashup and mixture behavior are described within a new specification of a high-level language, language for process mashup (BPML).

Keywords: business process, mashup, fragments, bp mashup

Procedia PDF Downloads 594
17915 A Hazard Rate Function for the Time of Ruin

Authors: Sule Sahin, Basak Bulut Karageyik

Abstract:

This paper introduces a hazard rate function for the time of ruin to calculate the conditional probability of ruin for very small intervals. We call this function the force of ruin (FoR). We obtain the expected time of ruin and conditional expected time of ruin from the exact finite time ruin probability with exponential claim amounts. Then we introduce the FoR which gives the conditional probability of ruin and the condition is that ruin has not occurred at time t. We analyse the behavior of the FoR function for different initial surpluses over a specific time interval. We also obtain FoR under the excess of loss reinsurance arrangement and examine the effect of reinsurance on the FoR.

Keywords: conditional time of ruin, finite time ruin probability, force of ruin, reinsurance

Procedia PDF Downloads 360
17914 Design of Control System Based On PLC and Kingview for Granulation Product Line

Authors: Mei-Feng, Yude-Fan, Min-Zhu

Abstract:

Based on PLC and kingview, this paper proposed a method that designed a set of the automatic control system according to the craft flow and demands for granulation product line. There were the main station and subordinate stations in PLC which were communicated by PROFIBUS network. PLC and computer were communicated by Ethernet network. The conversation function between human and machine was realized by kingview software, including actual time craft flows, historic report curves and product report forms. The construction of the control system, hardware collocation and software design were introduced. Besides these, PROFIBUS network frequency conversion control, the difficult points and configuration software design were elaborated. The running results showed that there were several advantages in the control system. They were high automatic degree, perfect function, perfect steady and convenient operation.

Keywords: PLC, PROFIBUS, configuration, frequency

Procedia PDF Downloads 369
17913 Increasing Efficiency of Own Used Fuel Gas by “LOTION” Method in Generating Systems PT. Pertamina EP Cepu Donggi Matindok Field in Central Sulawesi Province, Indonesia

Authors: Ridwan Kiay Demak, Firmansyahrullah, Muchammad Sibro Mulis, Eko Tri Wasisto, Nixon Poltak Frederic, Agung Putu Andika, Lapo Ajis Kamamu, Muhammad Sobirin, Kornelius Eppang

Abstract:

PC Prove LSM successfully improved the efficiency of Own Used Fuel Gas with the "Lotion" method in the PT Pertamina EP Cepu Donggi Matindok Generating System. The innovation of using the "LOTION" (LOAD PRIORITY SELECTION) method in the generating system is modeling that can provide a priority qualification of main and non-main equipment to keep gas processing running even though it leaves 1 GTG operating. GTG operating system has been integrated, controlled, and monitored properly through PC programs and web-based access to answer Industry 4.0 problems. The results of these improvements have succeeded in making Donggi Matindok Field Production reach 98.77 MMSCFD and become a proper EMAS candidate in 2022-2023. Additional revenue from increasing the efficiency of the use of own used gas amounting to USD USD 5.06 Million per year and reducing operational costs from maintenance efficiency (ABO) due to saving running hours GTG amounted to USD 3.26 Million per year. Continuity of fuel gas availability for the GTG generation system can maintain the operational reliability of the plant, which is 3.833333 MMSCFD. And reduced gas emissions wasted to the environment by 33,810 tons of C02 eq per year.

Keywords: LOTION method, load priority selection, fuel gas efficiency, gas turbine generator, reduce emissions

Procedia PDF Downloads 22
17912 Fast and Robust Long-term Tracking with Effective Searching Model

Authors: Thang V. Kieu, Long P. Nguyen

Abstract:

Kernelized Correlation Filter (KCF) based trackers have gained a lot of attention recently because of their accuracy and fast calculation speed. However, this algorithm is not robust in cases where the object is lost by a sudden change of direction, being obscured or going out of view. In order to improve KCF performance in long-term tracking, this paper proposes an anomaly detection method for target loss warning by analyzing the response map of each frame, and a classification algorithm for reliable target re-locating mechanism by using Random fern. Being tested with Visual Tracker Benchmark and Visual Object Tracking datasets, the experimental results indicated that the precision and success rate of the proposed algorithm were 2.92 and 2.61 times higher than that of the original KCF algorithm, respectively. Moreover, the proposed tracker handles occlusion better than many state-of-the-art long-term tracking methods while running at 60 frames per second.

Keywords: correlation filter, long-term tracking, random fern, real-time tracking

Procedia PDF Downloads 114