Search results for: accelerated failure time model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 31651

Search results for: accelerated failure time model

29971 Development of the Logistic Service Providers under the Pandemic Affects during COVID-19 in Turkey

Authors: Süleyman Günes

Abstract:

The crucial effects of the COVID-19 pandemic have on social and economic systems in Turkey as well as all over the world. It has impacted logistic providers and worldwide supply chains. Unexpected risks played a central role in creating vulnerabilities for logistics service operations during the pandemic terms. This study aims to research and design qualitative and quantitive contributions to logistic services. The COVID-19 pandemic brought unavoidable risks to the logistics industry in Turkey. The Logistic Service Providers (LSPs) have learned how to ensure uncertainties and risks triggered by main and adverse effects. The risks that LSPs encounter during the COVID-19 pandemic have been investigated and unveiled, and identified uncertainties and risks. The cause-effect structures were displayed by the qualitative and quantitive studies. The results suggest that supply chains and demand changes triggered by the COVID-19 pandemic while it influenced financial failure and forecast horizon with operational performances.

Keywords: logistic service providers, COVID-19, development, financial failure

Procedia PDF Downloads 60
29970 Flushing Model for Artificial Islands in the Persian Gulf

Authors: Sawsan Eissa, Momen Gharib, Omnia Kabbany

Abstract:

A flushing numerical study has been performed for intended artificial islands on the Persian Gulf coast in Abu Dhabi, UAE. The island masterplan was tested for flushing using the DELFT 3D hydrodynamic model, and it was found that its residence time exceeds the acceptable PIANC flushing Criteria. Therefore, a number of mitigation measures were applied and tested one by one using the flushing model. Namely, changing the location of the entrance opening, dredging, removing part of the mangrove existing in the near vicinity to create a channel, removing the mangrove altogether, using culverts of different numbers and locations, and pumping at selected points. The pumping option gave the best solution, but it was disregarded due to high capital and running costs. Therefore, it opted for a combination of other solutions, including removing mangroves, introducing culverts, and adjusting island boundaries and types of protection.

Keywords: hydrodynamics, flushing, delft 3d, Persian Gulf, artificial islands.

Procedia PDF Downloads 37
29969 A Data-Driven Compartmental Model for Dengue Forecasting and Covariate Inference

Authors: Yichao Liu, Peter Fransson, Julian Heidecke, Jonas Wallin, Joacim Rockloev

Abstract:

Dengue, a mosquito-borne viral disease, poses a significant public health challenge in endemic tropical or subtropical countries, including Sri Lanka. To reveal insights into the complexity of the dynamics of this disease and study the drivers, a comprehensive model capable of both robust forecasting and insightful inference of drivers while capturing the co-circulating of several virus strains is essential. However, existing studies mostly focus on only one aspect at a time and do not integrate and carry insights across the siloed approach. While mechanistic models are developed to capture immunity dynamics, they are often oversimplified and lack integration of all the diverse drivers of disease transmission. On the other hand, purely data-driven methods lack constraints imposed by immuno-epidemiological processes, making them prone to overfitting and inference bias. This research presents a hybrid model that combines machine learning techniques with mechanistic modelling to overcome the limitations of existing approaches. Leveraging eight years of newly reported dengue case data, along with socioeconomic factors, such as human mobility, weekly climate data from 2011 to 2018, genetic data detecting the introduction and presence of new strains, and estimates of seropositivity for different districts in Sri Lanka, we derive a data-driven vector (SEI) to human (SEIR) model across 16 regions in Sri Lanka at the weekly time scale. By conducting ablation studies, the lag effects allowing delays up to 12 weeks of time-varying climate factors were determined. The model demonstrates superior predictive performance over a pure machine learning approach when considering lead times of 5 and 10 weeks on data withheld from model fitting. It further reveals several interesting interpretable findings of drivers while adjusting for the dynamics and influences of immunity and introduction of a new strain. The study uncovers strong influences of socioeconomic variables: population density, mobility, household income and rural vs. urban population. The study reveals substantial sensitivity to the diurnal temperature range and precipitation, while mean temperature and humidity appear less important in the study location. Additionally, the model indicated sensitivity to vegetation index, both max and average. Predictions on testing data reveal high model accuracy. Overall, this study advances the knowledge of dengue transmission in Sri Lanka and demonstrates the importance of incorporating hybrid modelling techniques to use biologically informed model structures with flexible data-driven estimates of model parameters. The findings show the potential to both inference of drivers in situations of complex disease dynamics and robust forecasting models.

Keywords: compartmental model, climate, dengue, machine learning, social-economic

Procedia PDF Downloads 57
29968 1D Velocity Model for the Gobi-Altai Region from Local Earthquakes

Authors: Dolgormaa Munkhbaatar, Munkhsaikhan Adiya, Tseedulam Khuut

Abstract:

We performed an inversion method to determine the 1D-velocity model with station corrections of the Gobi-Altai area in the southern part of Mongolia using earthquake data collected in the National Data Center during the last 10 years. In this study, the concept of the new 1D model has been employed to minimize the average RMS of a set of well-located earthquakes, recorded at permanent (between 2006 and 2016) and temporary seismic stations (between 2014 and 2016), compute solutions for the coupled hypocenter and 1D velocity model. We selected 4800 events with RMS less than 0.5 seconds and with a maximum GAP of 170 degrees and determined velocity structures. Also, we relocated all possible events located in the Gobi-Altai area using the new 1D velocity model and achieved constrained hypocentral determinations for events within this area. We concluded that the estimated new 1D velocity model is a relatively low range compared to the previous velocity model in a significant improvement intend to, and the quality of the information basis for future research center locations to determine the earthquake epicenter area with this new transmission model.

Keywords: 1D velocity model, earthquake, relocation, Velest

Procedia PDF Downloads 144
29967 Optimization of Process Parameters and Modeling of Mass Transport during Hybrid Solar Drying of Paddy

Authors: Aprajeeta Jha, Punyadarshini P. Tripathy

Abstract:

Drying is one of the most critical unit operations for prolonging the shelf-life of food grains in order to ensure global food security. Photovoltaic integrated solar dryers can be a sustainable solution for replacing energy intensive thermal dryers as it is capable of drying in off-sunshine hours and provide better control over drying conditions. But, performance and reliability of PV based solar dryers depend hugely on climatic conditions thereby, drastically affecting process parameters. Therefore, to ensure quality and prolonged shelf-life of paddy, optimization of process parameters for solar dryers is critical. Proper moisture distribution within the grains is most detrimental factor to enhance the shelf-life of paddy therefore; modeling of mass transport can help in providing a better insight of moisture migration. Hence, present work aims at optimizing the process parameters and to develop a 3D finite element model (FEM) for predicting moisture profile in paddy during solar drying. Optimization of process parameters (power level, air velocity and moisture content) was done using box Behnken model in Design expert software. Furthermore, COMSOL Multiphysics was employed to develop a 3D finite element model for predicting moisture profile. Optimized model for drying paddy was found to be 700W, 2.75 m/s and 13% wb with optimum temperature, milling yield and drying time of 42˚C, 62%, 86 min respectively, having desirability of 0.905. Furthermore, 3D finite element model (FEM) for predicting moisture migration in single kernel for every time step has been developed. The mean absolute error (MAE), mean relative error (MRE) and standard error (SE) were found to be 0.003, 0.0531 and 0.0007, respectively, indicating close agreement of model with experimental results. Above optimized conditions can be successfully used to dry paddy in PV integrated solar dryer in order to attain maximum uniformity, quality and yield of product to achieve global food and energy security

Keywords: finite element modeling, hybrid solar drying, mass transport, paddy, process optimization

Procedia PDF Downloads 124
29966 Estimation of Soil Moisture at High Resolution through Integration of Optical and Microwave Remote Sensing and Applications in Drought Analyses

Authors: Donglian Sun, Yu Li, Paul Houser, Xiwu Zhan

Abstract:

California experienced severe drought conditions in the past years. In this study, the drought conditions in California are analyzed using soil moisture anomalies derived from integrated optical and microwave satellite observations along with auxiliary land surface data. Based on the U.S. Drought Monitor (USDM) classifications, three typical drought conditions were selected for the analysis: extreme drought conditions in 2007 and 2013, severe drought conditions in 2004 and 2009, and normal conditions in 2005 and 2006. Drought is defined as negative soil moisture anomaly. To estimate soil moisture at high spatial resolutions, three approaches are explored in this study: the universal triangle model that estimates soil moisture from Normalized Difference Vegetation Index (NDVI) and Land Surface Temperature (LST); the basic model that estimates soil moisture under different conditions with auxiliary data like precipitation, soil texture, topography, and surface types; and the refined model that uses accumulated precipitation and its lagging effects. It is found that the basic model shows better agreements with the USDM classifications than the universal triangle model, while the refined model using precipitation accumulated from the previous summer to current time demonstrated the closest agreements with the USDM patterns.

Keywords: soil moisture, high resolution, regional drought, analysis and monitoring

Procedia PDF Downloads 121
29965 Discrete-Event Modeling and Simulation Methodologies: Past, Present and Future

Authors: Gabriel Wainer

Abstract:

Modeling and Simulation methods have been used to better analyze the behavior of complex physical systems, and it is now common to use simulation as a part of the scientific and technological discovery process. M&S advanced thanks to the improvements in computer technology, which, in many cases, resulted in the development of simulation software using ad-hoc techniques. Formal M&S appeared in order to try to improve the development task of very complex simulation systems. Some of these techniques proved to be successful in providing a sound base for the development of discrete-event simulation models, improving the ease of model definition and enhancing the application development tasks; reducing costs and favoring reuse. The DEVS formalism is one of these techniques, which proved to be successful in providing means for modeling while reducing development complexity and costs. DEVS model development is based on a sound theoretical framework. The independence of M&S tasks made possible to run DEVS models on different environments (personal computers, parallel computers, real-time equipment, and distributed simulators) and middleware. We will present a historical perspective of discrete-event M&S methodologies, showing different modeling techniques. We will introduce DEVS origins and general ideas, and compare it with some of these techniques. We will then show the current status of DEVS M&S, and we will discuss a technological perspective to solve current M&S problems (including real-time simulation, interoperability, and model-centered development techniques). We will show some examples of the current use of DEVS, including applications in different fields. We will finally show current open topics in the area, which include advanced methods for centralized, parallel or distributed simulation, the need for real-time modeling techniques, and our view in these fields.

Keywords: modeling and simulation, discrete-event simulation, hybrid systems modeling, parallel and distributed simulation

Procedia PDF Downloads 307
29964 Transformative Measures in Chemical and Petrochemical Industry Through Agile Principles and Industry 4.0 Technologies

Authors: Bahman Ghorashi

Abstract:

The immense awareness of the global climate change has compelled traditional fossil fuel companies to develop strategies to reduce their carbon footprint and simultaneously consider the production of various sources of clean energy in order to mitigate the environmental impact of their operations. Similarly, supply chain issues, the scarcity of certain raw materials, energy costs as well as market needs, and changing consumer expectations have forced the traditional chemical industry to reexamine their time-honored modes of operation. This study examines how such transformative change might occur through the applications of agile principles as well as industry 4.0 technologies. Clearly, such a transformation is complex, costly, and requires a total commitment on the part of the top leadership and the entire management structure. Factors that need to be considered include organizational speed of change, a restructuring that would lend itself toward collaboration and the selling of solutions to customers’ problems, rather than just products, integrating ‘along’ as well as ‘across’ value chains, mastering change and uncertainty as well as a recognition of the importance of concept-to-cash time, i.e., the velocity of introducing new products to market, and the leveraging of people and information. At the same time, parallel to implementing such major shifts in the ethos, and the fabric of the organization, the change leaders should remain mindful of the companies’ DNA while incorporating the necessary DNA defying shifts. Furthermore, such strategic maneuvers should inevitably incorporate the managing of the upstream and downstream operations, harnessing future opportunities, preparing and training the workforce, implementing faster decision making and quick adaptation to change, managing accelerated response times, as well as forming autonomous and cross-functional teams. Moreover, the leaders should establish the balance between high-value solutions versus high-margin products, fully implement digitization of operations and, when appropriate, incorporate the latest relevant technologies, such as: AI, IIoT, ML, and immersive technologies. This study presents a summary of the agile principles and the relevant technologies and draws lessons from some of the best practices that are already implemented within the chemical industry in order to establish a roadmap to agility. Finally, the critical role of educational institutions in preparing the future workforce for Industry 4.0 is addressed.

Keywords: agile principles, immersive technologies, industry 4.0, workforce preparation

Procedia PDF Downloads 93
29963 The Transformative Landscape of the University of the Western Cape’s Elearning Center: Institutionalizing ELearning

Authors: Paul Dankers, Juliet Stoltenkamp, Carolynne Kies

Abstract:

In May 2005, the University of the Western Cape (UWC) established an eLearning Division (ED) that, over the past 18 years, accelerated into the institutionalization of an efficient eLearning Centre. The initial objective of the ED was to incessantly align itself with emerging technologies caused by digital transformation, which progressively impacted Higher Education Institutions (HEIs) globally. In this paper, we present how the UWC eLearning Division (ED) first evolved into the eLearning Development and Support Unit (EDUS), currently called the ‘Centre for Innovative Education and Communication Technologies (CIECT). CIECT was strategically separated from the Department of Information and Communication Services (ICS) in 2009 and repositioned as an independent structure at UWC. Using a comparative research method, we highlight the transformative eLearning landscape at UWC by doing a detailed account of the shift in practices. Our research method will determine the initial vision and outcomes of institutionalizing an eLearning division. The study aims to compare across space or time the eLearning division’s rate of growth. By comparing the progressive growth of the UWCs eLearning division over the years, we will be able to document the successes and achievements of the eLearning division precisely. This study’s outcomes will act as a reference for novel research subjects on formalising eLearning. More research that delves into the effectiveness of having an eLearning division at HEIs in support of students’ teaching and learning is needed.

Keywords: eLearning, institutionalization, teaching and learning, transformation

Procedia PDF Downloads 18
29962 Investigating Climate Change Trend Based on Data Simulation and IPCC Scenario during 2010-2030 AD: Case Study of Fars Province

Authors: Leila Rashidian, Abbas Ebrahimi

Abstract:

The development of industrial activities, increase in fossil fuel consumption, vehicles, destruction of forests and grasslands, changes in land use, and population growth have caused to increase the amount of greenhouse gases especially CO2 in the atmosphere in recent decades. This has led to global warming and climate change. In the present paper, we have investigated the trend of climate change according to the data simulation during the time interval of 2010-2030 in the Fars province. In this research, the daily climatic parameters such as maximum and minimum temperature, precipitation and number of sunny hours during the 1977-2008 time interval for synoptic stations of Shiraz and Abadeh and during 1995-2008 for Lar stations and also the output of HADCM3 model in 2010-2030 time interval have been used based on the A2 propagation scenario. The results of the model show that the average temperature will increase by about 1 degree centigrade and the amount of precipitation will increase by 23.9% compared to the observational data. In conclusion, according to the temperature increase in this province, the amount of precipitation in the form of snow will be reduced and precipitations often will occur in the form of rain. This 1-degree centigrade increase during the season will reduce production by 6 to 10% because of shortening the growing period of wheat.

Keywords: climate change, Lars WG, HADCM3, Gillan province, climatic parameters, A2 scenario

Procedia PDF Downloads 197
29961 Failure Statistics Analysis of China’s Spacecraft in Full-Life

Authors: Xin-Yan Ji

Abstract:

The historical failures data of the spacecraft is very useful to improve the spacecraft design and the test philosophies and reduce the spacecraft flight risk. A study of spacecraft failures data was performed, which is the most comprehensive statistics of spacecrafts in China. 2593 on-orbit failures data and 1298 ground data that occurred on 150 spacecraft launched from 2000 to 2016 were identified and collected, which covered the navigation satellites, communication satellites, remote sensing deep space exploration manned spaceflight platforms. In this paper, the failures were analyzed to compare different spacecraft subsystem and estimate their impact on the mission, then the development of spacecraft in China was evaluated from design, software, workmanship, management, parts, and materials. Finally, the lessons learned from the past years show that electrical and mechanical failures are responsible for the largest parts, and the key solution to reduce in-orbit failures is improving design technology, enough redundancy, adequate space environment protection measures, and adequate ground testing.

Keywords: spacecraft anomalies, anomalies mechanism, failure cause, spacecraft testing

Procedia PDF Downloads 100
29960 An Elbow Biomechanical Model and Its Coefficients Adjustment

Authors: Jie Bai, Yongsheng Gao, Shengxin Wang, Jie Zhao

Abstract:

Through the establishment of the elbow biomechanical model, it can provide theoretical guide for rehabilitation therapy on the upper limb of the human body. A biomechanical model of the elbow joint can be built by the connection of muscle force model and elbow dynamics. But there are many undetermined coefficients in the model like the optimal joint angle and optimal muscle force which are usually specified as the experimental parameters of other workers. Because of the individual differences, there is a certain deviation of the final result. To this end, the RMS value of the deviation between the actual angle and calculated angle is considered. A set of coefficients which lead to the minimum RMS value will be chosen to be the optimal parameters. The direct search method and the conjugacy search method are used to get the optimal parameters, thus the model can be more accurate and mode adaptability.

Keywords: elbow biomechanical model, RMS, direct search, conjugacy search

Procedia PDF Downloads 526
29959 Logistics Hub Location and Scheduling Model for Urban Last-Mile Deliveries

Authors: Anastasios Charisis, Evangelos Kaisar, Steven Spana, Lili Du

Abstract:

Logistics play a vital role in the prosperity of today’s cities, but current urban logistics practices are proving problematic, causing negative effects such as traffic congestion and environmental impacts. This paper proposes an alternative urban logistics system, leasing hubs inside cities for designated time intervals, and using handcarts for last-mile deliveries. A mathematical model for selecting the locations of hubs and allocating customers, while also scheduling the optimal times during the day for leasing hubs is developed. The proposed model is compared to current delivery methods requiring door-to-door truck deliveries. It is shown that truck traveled distances decrease by more than 60%. In addition, analysis shows that in certain conditions the approach can be economically competitive and successfully applied to address real problems.

Keywords: hub location, last-mile, logistics, optimization

Procedia PDF Downloads 173
29958 Forecasting for Financial Stock Returns Using a Quantile Function Model

Authors: Yuzhi Cai

Abstract:

In this paper, we introduce a newly developed quantile function model that can be used for estimating conditional distributions of financial returns and for obtaining multi-step ahead out-of-sample predictive distributions of financial returns. Since we forecast the whole conditional distributions, any predictive quantity of interest about the future financial returns can be obtained simply as a by-product of the method. We also show an application of the model to the daily closing prices of Dow Jones Industrial Average (DJIA) series over the period from 2 January 2004 - 8 October 2010. We obtained the predictive distributions up to 15 days ahead for the DJIA returns, which were further compared with the actually observed returns and those predicted from an AR-GARCH model. The results show that the new model can capture the main features of financial returns and provide a better fitted model together with improved mean forecasts compared with conventional methods. We hope this talk will help audience to see that this new model has the potential to be very useful in practice.

Keywords: DJIA, financial returns, predictive distribution, quantile function model

Procedia PDF Downloads 354
29957 Deterioration Prediction of Pavement Load Bearing Capacity from FWD Data

Authors: Kotaro Sasai, Daijiro Mizutani, Kiyoyuki Kaito

Abstract:

Expressways in Japan have been built in an accelerating manner since the 1960s with the aid of rapid economic growth. About 40 percent in length of expressways in Japan is now 30 years and older and has become superannuated. Time-related deterioration has therefore reached to a degree that administrators, from a standpoint of operation and maintenance, are forced to take prompt measures on a large scale aiming at repairing inner damage deep in pavements. These measures have already been performed for bridge management in Japan and are also expected to be embodied for pavement management. Thus, planning methods for the measures are increasingly demanded. Deterioration of layers around road surface such as surface course and binder course is brought about at the early stages of whole pavement deterioration process, around 10 to 30 years after construction. These layers have been repaired primarily because inner damage usually becomes significant after outer damage, and because surveys for measuring inner damage such as Falling Weight Deflectometer (FWD) survey and open-cut survey are costly and time-consuming process, which has made it difficult for administrators to focus on inner damage as much as they have been supposed to. As expressways today have serious time-related deterioration within them deriving from the long time span since they started to be used, it is obvious the idea of repairing layers deep in pavements such as base course and subgrade must be taken into consideration when planning maintenance on a large scale. This sort of maintenance requires precisely predicting degrees of deterioration as well as grasping the present situations of pavements. Methods for predicting deterioration are determined to be either mechanical or statistical. While few mechanical models have been presented, as far as the authors know of, previous studies have presented statistical methods for predicting deterioration in pavements. One describes deterioration process by estimating Markov deterioration hazard model, while another study illustrates it by estimating Proportional deterioration hazard model. Both of the studies analyze deflection data obtained from FWD surveys and present statistical methods for predicting deterioration process of layers around road surface. However, layers of base course and subgrade remain unanalyzed. In this study, data collected from FWD surveys are analyzed to predict deterioration process of layers deep in pavements in addition to surface layers by a means of estimating a deterioration hazard model using continuous indexes. This model can prevent the loss of information of data when setting rating categories in Markov deterioration hazard model when evaluating degrees of deterioration in roadbeds and subgrades. As a result of portraying continuous indexes, the model can predict deterioration in each layer of pavements and evaluate it quantitatively. Additionally, as the model can also depict probability distribution of the indexes at an arbitrary point and establish a risk control level arbitrarily, it is expected that this study will provide knowledge like life cycle cost and informative content during decision making process referring to where to do maintenance on as well as when.

Keywords: deterioration hazard model, falling weight deflectometer, inner damage, load bearing capacity, pavement

Procedia PDF Downloads 366
29956 Surveying Energy Dissipation in Stepped Spillway Using Finite Element Modeling

Authors: Mehdi Fuladipanah

Abstract:

Stepped spillway includes several steps from the crest to the toe. The steps of stepped spillway could cause to decrease the energy with making energy distribution in the longitude mode and also to reduce the outcome speed. The aim of this study was to stimulate the stepped spillway combined with stilling basin-step using Fluent model and the turbulent superficial flow using RNG, K-ε. The free surface of the flow was monitored by VOF model. The velocity and the depth of the flow were measured by tail water depth by the numerical model and then the dissipated energy was calculated along the spillway. The results indicated that the stilling basin-step complex may cause energy dissipation increment in the stepped spillway. Also, the numerical model was suggested as an effective method to predict the circular and complicated flows in the stepped spillways.

Keywords: stepped spillway, fluent model, VOF model, K-ε model, energy distribution

Procedia PDF Downloads 355
29955 Elastic and Plastic Collision Comparison Using Finite Element Method

Authors: Gustavo Rodrigues, Hans Weber, Larissa Driemeier

Abstract:

The prevision of post-impact conditions and the behavior of the bodies during the impact have been object of several collision models. The formulation from Hertz’s theory is generally used dated from the 19th century. These models consider the repulsive force as proportional to the deformation of the bodies under contact and may consider it proportional to the rate of deformation. The objective of the present work is to analyze the behavior of the bodies during impact using the Finite Element Method (FEM) with elastic and plastic material models. The main parameters to evaluate are, the contact force, the time of contact and the deformation of the bodies. An advantage of using the FEM approach is the possibility to apply a plastic deformation to the model according to the material definition: there will be used Johnson–Cook plasticity model whose parameters are obtained through empirical tests of real materials. This model allows analyzing the permanent deformation caused by impact, phenomenon observed in real world depending on the forces applied to the body. These results are compared between them and with the model-based Hertz theory.

Keywords: collision, impact models, finite element method, Hertz Theory

Procedia PDF Downloads 158
29954 Earthquake Effect in Micro Hydro Sector: Case Study of Dulakha District, Nepal

Authors: Keshav Raj Dhakal, Jit Bahadur Rokaya Chhetri

Abstract:

The Micro Hydro (MH) is one of the successful technology in Rural Nepal. Out of 75 district, 59 districts have installed 1287 MH projects with a total capacity of 24 Mega Watt (MW). Now, the challenge is how to sustain them. Dolakha is a prominent district for sustainable endues of power to sustain the MH projects. A total of 37 MH projects have been constructed with producing 886 Kilo Watt (KW) of energy in the district. This study traces out the impact of earthquake in MH sector in Dolakha district. It shows that 59 % of projects have been affected by devastating earthquake in April and May, 2015 where 29 % are completely damaged. Most of the damages are in civil structures like Penstock, forebay, power house, Canal, Intake. Transmission and distribution line have been partially damaged. This paper analysis failure of the civil structural component of MH projects and its financial consequence to the community. This study recommends that a disaster impact assessment is essential before construction of MH projects.

Keywords: micro hydro, earthquake, structural failure, financial consequence

Procedia PDF Downloads 189
29953 Faster Pedestrian Recognition Using Deformable Part Models

Authors: Alessandro Preziosi, Antonio Prioletti, Luca Castangia

Abstract:

Deformable part models achieve high precision in pedestrian recognition, but all publicly available implementations are too slow for real-time applications. We implemented a deformable part model algorithm fast enough for real-time use by exploiting information about the camera position and orientation. This implementation is both faster and more precise than alternative DPM implementations. These results are obtained by computing convolutions in the frequency domain and using lookup tables to speed up feature computation. This approach is almost an order of magnitude faster than the reference DPM implementation, with no loss in precision. Knowing the position of the camera with respect to horizon it is also possible prune many hypotheses based on their size and location. The range of acceptable sizes and positions is set by looking at the statistical distribution of bounding boxes in labelled images. With this approach it is not needed to compute the entire feature pyramid: for example higher resolution features are only needed near the horizon. This results in an increase in mean average precision of 5% and an increase in speed by a factor of two. Furthermore, to reduce misdetections involving small pedestrians near the horizon, input images are supersampled near the horizon. Supersampling the image at 1.5 times the original scale, results in an increase in precision of about 4%. The implementation was tested against the public KITTI dataset, obtaining an 8% improvement in mean average precision over the best performing DPM-based method. By allowing for a small loss in precision computational time can be easily brought down to our target of 100ms per image, reaching a solution that is faster and still more precise than all publicly available DPM implementations.

Keywords: autonomous vehicles, deformable part model, dpm, pedestrian detection, real time

Procedia PDF Downloads 261
29952 The Comparison of Joint Simulation and Estimation Methods for the Geometallurgical Modeling

Authors: Farzaneh Khorram

Abstract:

This paper endeavors to construct a block model to assess grinding energy consumption (CCE) and pinpoint blocks with the highest potential for energy usage during the grinding process within a specified region. Leveraging geostatistical techniques, particularly joint estimation, or simulation, based on geometallurgical data from various mineral processing stages, our objective is to forecast CCE across the study area. The dataset encompasses variables obtained from 2754 drill samples and a block model comprising 4680 blocks. The initial analysis encompassed exploratory data examination, variography, multivariate analysis, and the delineation of geological and structural units. Subsequent analysis involved the assessment of contacts between these units and the estimation of CCE via cokriging, considering its correlation with SPI. The selection of blocks exhibiting maximum CCE holds paramount importance for cost estimation, production planning, and risk mitigation. The study conducted exploratory data analysis on lithology, rock type, and failure variables, revealing seamless boundaries between geometallurgical units. Simulation methods, such as Plurigaussian and Turning band, demonstrated more realistic outcomes compared to cokriging, owing to the inherent characteristics of geometallurgical data and the limitations of kriging methods.

Keywords: geometallurgy, multivariate analysis, plurigaussian, turning band method, cokriging

Procedia PDF Downloads 37
29951 AI-Based Information System for Hygiene and Safety Management of Shared Kitchens

Authors: Jongtae Rhee, Sangkwon Han, Seungbin Ji, Junhyeong Park, Byeonghun Kim, Taekyung Kim, Byeonghyeon Jeon, Jiwoo Yang

Abstract:

The shared kitchen is a concept that transfers the value of the sharing economy to the kitchen. It is a type of kitchen equipped with cooking facilities that allows multiple companies or chefs to share time and space and use it jointly. These shared kitchens provide economic benefits and convenience, such as reduced investment costs and rent, but also increase the risk of safety management, such as cross-contamination of food ingredients. Therefore, to manage the safety of food ingredients and finished products in a shared kitchen where several entities jointly use the kitchen and handle various types of food ingredients, it is critical to manage followings: the freshness of food ingredients, user hygiene and safety and cross-contamination of cooking equipment and facilities. In this study, it propose a machine learning-based system for hygiene safety and cross-contamination management, which are highly difficult to manage. User clothing management and user access management, which are most relevant to the hygiene and safety of shared kitchens, are solved through machine learning-based methodology, and cutting board usage management, which is most relevant to cross-contamination management, is implemented as an integrated safety management system based on artificial intelligence. First, to prevent cross-contamination of food ingredients, we use images collected through a real-time camera to determine whether the food ingredients match a given cutting board based on a real-time object detection model, YOLO v7. To manage the hygiene of user clothing, we use a camera-based facial recognition model to recognize the user, and real-time object detection model to determine whether a sanitary hat and mask are worn. In addition, to manage access for users qualified to enter the shared kitchen, we utilize machine learning based signature recognition module. By comparing the pairwise distance between the contract signature and the signature at the time of entrance to the shared kitchen, access permission is determined through a pre-trained signature verification model. These machine learning-based safety management tasks are integrated into a single information system, and each result is managed in an integrated database. Through this, users are warned of safety dangers through the tablet PC installed in the shared kitchen, and managers can track the cause of the sanitary and safety accidents. As a result of system integration analysis, real-time safety management services can be continuously provided by artificial intelligence, and machine learning-based methodologies are used for integrated safety management of shared kitchens that allows dynamic contracts among various users. By solving this problem, we were able to secure the feasibility and safety of the shared kitchen business.

Keywords: artificial intelligence, food safety, information system, safety management, shared kitchen

Procedia PDF Downloads 48
29950 Punching Shear Strengthening of Reinforced Concrete Flat Slabs Using Internal Square Patches of Carbon Fiber Reinforced Polymer

Authors: Malik Assi

Abstract:

This research presents a strengthening technique for enhancing the punching shear resistance of concrete flat slabs. Internal square patches of CFRP were centrally installed inside 450*450mm concrete panels during casting at a chosen distance from the tension face to produce six simply supported samples. The dimensions of those patches ranged from 50*50mm to 360*360mm. All the examined slabs contained the same amount of tensile reinforcement, had identical dimensions, were designed according to the American Concrete Institute code (ACI) and tested to failure. Compared to the control unstrengthened spacemen, all the strengthened slabs have shown an enhancement in punching capacity and stiffness. This enhancement has been found to be proportional to the area of the installed CFRP patches. In addition to the reasonably enhanced stiffness and punching shear, this strengthening technique can change the slab failure mode from shear to flexural.

Keywords: CFRP patches, Flat slabs, Flexural, Stiffness, Punching shear

Procedia PDF Downloads 246
29949 Monitoring Three-Dimensional Models of Tree and Forest by Using Digital Close-Range Photogrammetry

Authors: S. Y. Cicekli

Abstract:

In this study, tree-dimensional model of tree was created by using terrestrial close range photogrammetry. For this close range photos were taken. Photomodeler Pro 5 software was used for camera calibration and create three-dimensional model of trees. In first test, three-dimensional model of a tree was created, in the second test three-dimensional model of three trees were created. This study aim is creating three-dimensional model of trees and indicate the use of close-range photogrammetry in forestry. At the end of the study, three-dimensional model of tree and three trees were created. This study showed that usability of close-range photogrammetry for monitoring tree and forests three-dimensional model.

Keywords: close- range photogrammetry, forest, tree, three-dimensional model

Procedia PDF Downloads 376
29948 A Mathematical-Based Formulation of EEG Fluctuations

Authors: Razi Khalafi

Abstract:

Brain is the information processing center of the human body. Stimuli in form of information are transferred to the brain and then brain makes the decision on how to respond to them. In this research we propose a new partial differential equation which analyses the EEG signals and make a relationship between the incoming stimuli and the brain response to them. In order to test the proposed model, a set of external stimuli applied to the model and the model’s outputs were checked versus the real EEG data. The results show that this model can model the EEG signal well. The proposed model is useful not only for modeling of the EEG signal in case external stimuli but it can be used for the modeling of brain response in case of internal stimuli.

Keywords: Brain, stimuli, partial differential equation, response, eeg signal

Procedia PDF Downloads 413
29947 Solid Particles Transport and Deposition Prediction in a Turbulent Impinging Jet Using the Lattice Boltzmann Method and a Probabilistic Model on GPU

Authors: Ali Abdul Kadhim, Fue Lien

Abstract:

Solid particle distribution on an impingement surface has been simulated utilizing a graphical processing unit (GPU). In-house computational fluid dynamics (CFD) code has been developed to investigate a 3D turbulent impinging jet using the lattice Boltzmann method (LBM) in conjunction with large eddy simulation (LES) and the multiple relaxation time (MRT) models. This paper proposed an improvement in the LBM-cellular automata (LBM-CA) probabilistic method. In the current model, the fluid flow utilizes the D3Q19 lattice, while the particle model employs the D3Q27 lattice. The particle numbers are defined at the same regular LBM nodes, and transport of particles from one node to its neighboring nodes are determined in accordance with the particle bulk density and velocity by considering all the external forces. The previous models distribute particles at each time step without considering the local velocity and the number of particles at each node. The present model overcomes the deficiencies of the previous LBM-CA models and, therefore, can better capture the dynamic interaction between particles and the surrounding turbulent flow field. Despite the increasing popularity of LBM-MRT-CA model in simulating complex multiphase fluid flows, this approach is still expensive in term of memory size and computational time required to perform 3D simulations. To improve the throughput of each simulation, a single GeForce GTX TITAN X GPU is used in the present work. The CUDA parallel programming platform and the CuRAND library are utilized to form an efficient LBM-CA algorithm. The methodology was first validated against a benchmark test case involving particle deposition on a square cylinder confined in a duct. The flow was unsteady and laminar at Re=200 (Re is the Reynolds number), and simulations were conducted for different Stokes numbers. The present LBM solutions agree well with other results available in the open literature. The GPU code was then used to simulate the particle transport and deposition in a turbulent impinging jet at Re=10,000. The simulations were conducted for L/D=2,4 and 6, where L is the nozzle-to-surface distance and D is the jet diameter. The effect of changing the Stokes number on the particle deposition profile was studied at different L/D ratios. For comparative studies, another in-house serial CPU code was also developed, coupling LBM with the classical Lagrangian particle dispersion model. Agreement between results obtained with LBM-CA and LBM-Lagrangian models and the experimental data is generally good. The present GPU approach achieves a speedup ratio of about 350 against the serial code running on a single CPU.

Keywords: CUDA, GPU parallel programming, LES, lattice Boltzmann method, MRT, multi-phase flow, probabilistic model

Procedia PDF Downloads 189
29946 A Statistical Model for the Dynamics of Single Cathode Spot in Vacuum Cylindrical Cathode

Authors: Po-Wen Chen, Jin-Yu Wu, Md. Manirul Ali, Yang Peng, Chen-Te Chang, Der-Jun Jan

Abstract:

Dynamics of cathode spot has become a major part of vacuum arc discharge with its high academic interest and wide application potential. In this article, using a three-dimensional statistical model, we simulate the distribution of the ignition probability of a new cathode spot occurring in different magnetic pressure on old cathode spot surface and at different arcing time. This model for the ignition probability of a new cathode spot was proposed in two typical situations, one by the pure isotropic random walk in the absence of an external magnetic field, other by the retrograde motion in external magnetic field, in parallel with the cathode surface. We mainly focus on developed relationship between the ignition probability density distribution of a new cathode spot and the external magnetic field.

Keywords: cathode spot, vacuum arc discharge, transverse magnetic field, random walk

Procedia PDF Downloads 417
29945 Autonomic Sonar Sensor Fault Manager for Mobile Robots

Authors: Martin Doran, Roy Sterritt, George Wilkie

Abstract:

NASA, ESA, and NSSC space agencies have plans to put planetary rovers on Mars in 2020. For these future planetary rovers to succeed, they will heavily depend on sensors to detect obstacles. This will also become of vital importance in the future, if rovers become less dependent on commands received from earth-based control and more dependent on self-configuration and self-decision making. These planetary rovers will face harsh environments and the possibility of hardware failure is high, as seen in missions from the past. In this paper, we focus on using Autonomic principles where self-healing, self-optimization, and self-adaption are explored using the MAPE-K model and expanding this model to encapsulate the attributes such as Awareness, Analysis, and Adjustment (AAA-3). In the experimentation, a Pioneer P3-DX research robot is used to simulate a planetary rover. The sonar sensors on the P3-DX robot are used to simulate the sensors on a planetary rover (even though in reality, sonar sensors cannot operate in a vacuum). Experiments using the P3-DX robot focus on how our software system can be adapted with the loss of sonar sensor functionality. The autonomic manager system is responsible for the decision making on how to make use of remaining ‘enabled’ sonars sensors to compensate for those sonar sensors that are ‘disabled’. The key to this research is that the robot can still detect objects even with reduced sonar sensor capability.

Keywords: autonomic, self-adaption, self-healing, self-optimization

Procedia PDF Downloads 330
29944 A Mathematical Equation to Calculate Stock Price of Different Growth Model

Authors: Weiping Liu

Abstract:

This paper presents an equation to calculate stock prices of different growth model. This equation is mathematically derived by using discounted cash flow method. It has the advantages of being very easy to use and very accurate. It can still be used even when the first stage is lengthy. This equation is more generalized because it can be used for all the three popular stock price models. It can be programmed into financial calculator or electronic spreadsheets. In addition, it can be extended to a multistage model. It is more versatile and efficient than the traditional methods.

Keywords: stock price, multistage model, different growth model, discounted cash flow method

Procedia PDF Downloads 387
29943 Evaluation of a Piecewise Linear Mixed-Effects Model in the Analysis of Randomized Cross-over Trial

Authors: Moses Mwangi, Geert Verbeke, Geert Molenberghs

Abstract:

Cross-over designs are commonly used in randomized clinical trials to estimate efficacy of a new treatment with respect to a reference treatment (placebo or standard). The main advantage of using cross-over design over conventional parallel design is its flexibility, where every subject become its own control, thereby reducing confounding effect. Jones & Kenward, discuss in detail more recent developments in the analysis of cross-over trials. We revisit the simple piecewise linear mixed-effects model, proposed by Mwangi et. al, (in press) for its first application in the analysis of cross-over trials. We compared performance of the proposed piecewise linear mixed-effects model with two commonly cited statistical models namely, (1) Grizzle model; and (2) Jones & Kenward model, used in estimation of the treatment effect, in the analysis of randomized cross-over trial. We estimate two performance measurements (mean square error (MSE) and coverage probability) for the three methods, using data simulated from the proposed piecewise linear mixed-effects model. Piecewise linear mixed-effects model yielded lowest MSE estimates compared to Grizzle and Jones & Kenward models for both small (Nobs=20) and large (Nobs=600) sample sizes. It’s coverage probability were highest compared to Grizzle and Jones & Kenward models for both small and large sample sizes. A piecewise linear mixed-effects model is a better estimator of treatment effect than its two competing estimators (Grizzle and Jones & Kenward models) in the analysis of cross-over trials. The data generating mechanism used in this paper captures two time periods for a simple 2-Treatments x 2-Periods cross-over design. Its application is extendible to more complex cross-over designs with multiple treatments and periods. In addition, it is important to note that, even for single response models, adding more random effects increases the complexity of the model and thus may be difficult or impossible to fit in some cases.

Keywords: Evaluation, Grizzle model, Jones & Kenward model, Performance measures, Simulation

Procedia PDF Downloads 106
29942 Predicting Mixing Patterns of Overflows from a Square Manhole

Authors: Modupe O. Jimoh

Abstract:

During manhole overflows, its contents pollute the immediate environment. Understanding the pollutant transfer characteristics between manhole’s incoming sewer and the overflow is therefore of great importance. A square manhole with sides 388 mm by 388 mm and height 700 mm with an overflow facility was used in the laboratory to carry out overflow concentration measurements. Two scenarios were investigated using three flow rates. The first scenario corresponded to when the exit of the pipe becomes blocked and the only exit for the flow is the manhole. The second scenario is when there is an overflow in combination with a pipe exit. The temporal concentration measurements showed that the peak concentration of pollutants in the flow was attenuated between the inlet and the overflow. A deconvolution software was used to predict the Residence time distribution (RTD) and consequently the Cumulative Residence time distribution (CRTD). The CRTDs suggest that complete mixing is occurring between the pipe inlet and the overflow, like what is obtained in a low surcharged manhole. The results also suggest that an instantaneous stirred tank reactor model can describe the mixing characteristics.

Keywords: CRTDs, instantaneous stirred tank reactor model, overflow, square manholes, surcharge, temporal concentration profiles

Procedia PDF Downloads 120