Search results for: default probability
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1377

Search results for: default probability

507 Distances over Incomplete Diabetes and Breast Cancer Data Based on Bhattacharyya Distance

Authors: Loai AbdAllah, Mahmoud Kaiyal

Abstract:

Missing values in real-world datasets are a common problem. Many algorithms were developed to deal with this problem, most of them replace the missing values with a fixed value that was computed based on the observed values. In our work, we used a distance function based on Bhattacharyya distance to measure the distance between objects with missing values. Bhattacharyya distance, which measures the similarity of two probability distributions. The proposed distance distinguishes between known and unknown values. Where the distance between two known values is the Mahalanobis distance. When, on the other hand, one of them is missing the distance is computed based on the distribution of the known values, for the coordinate that contains the missing value. This method was integrated with Wikaya, a digital health company developing a platform that helps to improve prevention of chronic diseases such as diabetes and cancer. In order for Wikaya’s recommendation system to work distance between users need to be measured. Since there are missing values in the collected data, there is a need to develop a distance function distances between incomplete users profiles. To evaluate the accuracy of the proposed distance function in reflecting the actual similarity between different objects, when some of them contain missing values, we integrated it within the framework of k nearest neighbors (kNN) classifier, since its computation is based only on the similarity between objects. To validate this, we ran the algorithm over diabetes and breast cancer datasets, standard benchmark datasets from the UCI repository. Our experiments show that kNN classifier using our proposed distance function outperforms the kNN using other existing methods.

Keywords: missing values, incomplete data, distance, incomplete diabetes data

Procedia PDF Downloads 225
506 Spatio-Temporal Pest Risk Analysis with ‘BioClass’

Authors: Vladimir A. Todiras

Abstract:

Spatio-temporal models provide new possibilities for real-time action in pest risk analysis. It should be noted that estimation of the possibility and probability of introduction of a pest and of its economic consequences involves many uncertainties. We present a new mapping technique that assesses pest invasion risk using online BioClass software. BioClass is a GIS tool designed to solve multiple-criteria classification and optimization problems based on fuzzy logic and level set methods. This research describes a method for predicting the potential establishment and spread of a plant pest into new areas using a case study: corn rootworm (Diabrotica spp.), tomato leaf miner (Tuta absoluta) and plum fruit moth (Grapholita funebrana). Our study demonstrated that in BioClass we can combine fuzzy logic and geographic information systems with knowledge of pest biology and environmental data to derive new information for decision making. Pests are sensitive to a warming climate, as temperature greatly affects their survival and reproductive rate and capacity. Changes have been observed in the distribution, frequency and severity of outbreaks of Helicoverpa armigera on tomato. BioClass has demonstrated to be a powerful tool for applying dynamic models and map the potential future distribution of a species, enable resource to make decisions about dangerous and invasive species management and control.

Keywords: classification, model, pest, risk

Procedia PDF Downloads 282
505 An Understanding of Child Sexual Abuse in South Africa: Case Study of Eastern Cape Province

Authors: Mandlenkosi Richard Mphatheni

Abstract:

The Constitution of the Republic of South Africa (Act 108 of 1996 section 28(1) (d)) states, ‘Every child has the right to be protected from maltreatment, neglect, abuse, and degradation’. Qualitative research studied perceptions of the selected sample. Objectives of the research were to determine factors that influence perpetrators of sexual violence to target children, the risk factors of child sexual abuse, the effects of child sexual abuse on the development of the child, and the community prevention measures to minimize the risks of child sexual abuse. The research aimed to understand perspective and experiences of the Ngangelizwe community members on the problem of sexual violence against children and the perpetrator’s perceived motive for sexually abusing children. Convenience non-probability sampling technique was adopted to select 20 participants within the Ngangelizwe Township at Mthatha. Thematic analyses were used to analyse data. It was found that sexual abuse of children affects severely child and parents, while the community reported to be trivially affected by the sexual abuse of a child. The research revealed ignorance of some forms of sexual violence, as the commonly known form of sexual violence was rape. Therefore, ignorance of community members regarding various forms of sexual abuse means that such acts are either ignored, tolerated, or even regarded as acceptable. It thus means that community members cannot reject any actions or behaviour if they themselves are ignorant of what constitutes sexual violence. This study recommends that communities should be educated about different sexual offenses.

Keywords: child sexual abuse, community, childhood attachment, adult attachment

Procedia PDF Downloads 138
504 Challenges in the Material and Action-Resistance Factor Design for Embedded Retaining Wall Limit State Analysis

Authors: Kreso Ivandic, Filip Dodigovic, Damir Stuhec

Abstract:

The paper deals with the proposed 'Material' and 'Action-resistance factor' design methods in designing the embedded retaining walls. The parametric analysis of evaluating the differences of the output values mutually and compared with classic approach computation was performed. There is a challenge with the criteria for choosing the proposed calculation design methods in Eurocode 7 with respect to current technical regulations and regular engineering practice. The basic criterion for applying a particular design method is to ensure minimum an equal degree of reliability in relation to the current practice. The procedure of combining the relevant partial coefficients according to design methods was carried out. The use of mentioned partial coefficients should result in the same level of safety, regardless of load combinations, material characteristics and problem geometry. This proposed approach of the partial coefficients related to the material and/or action-resistance should aimed at building a bridge between calculations used so far and pure probability analysis. The measure to compare the results was to determine an equivalent safety factor for each analysis. The results show a visible wide span of equivalent values of the classic safety factors.

Keywords: action-resistance factor design, classic approach, embedded retaining wall, Eurocode 7, limit states, material factor design

Procedia PDF Downloads 231
503 Probabilistic Slope Stability Analysis of Excavation Induced Landslides Using Hermite Polynomial Chaos

Authors: Schadrack Mwizerwa

Abstract:

The characterization and prediction of landslides are crucial for assessing geological hazards and mitigating risks to infrastructure and communities. This research aims to develop a probabilistic framework for analyzing excavation-induced landslides, which is fundamental for assessing geological hazards and mitigating risks to infrastructure and communities. The study uses Hermite polynomial chaos, a non-stationary random process, to analyze the stability of a slope and characterize the failure probability of a real landslide induced by highway construction excavation. The correlation within the data is captured using the Karhunen-Loève (KL) expansion theory, and the finite element method is used to analyze the slope's stability. The research contributes to the field of landslide characterization by employing advanced random field approaches, providing valuable insights into the complex nature of landslide behavior and the effectiveness of advanced probabilistic models for risk assessment and management. The data collected from the Baiyuzui landslide, induced by highway construction, is used as an illustrative example. The findings highlight the importance of considering the probabilistic nature of landslides and provide valuable insights into the complex behavior of such hazards.

Keywords: Hermite polynomial chaos, Karhunen-Loeve, slope stability, probabilistic analysis

Procedia PDF Downloads 78
502 Solutions for Food-Safe 3D Printing

Authors: Geremew Geidare Kailo, Igor Gáspár, András Koris, Ivana Pajčin, Flóra Vitális, Vanja Vlajkov

Abstract:

Three-dimension (3D) printing, a very popular additive manufacturing technology, has recently undergone rapid growth and replaced the use of conventional technology from prototyping to producing end-user parts and products. The 3D Printing technology involves a digital manufacturing machine that produces three-dimensional objects according to designs created by the user via 3D modeling or computer-aided design/manufacturing (CAD/CAM) software. The most popular 3D printing system is Fused Deposition Modeling (FDM) or also called Fused Filament Fabrication (FFF). A 3D-printed object is considered food safe if it can have direct contact with the food without any toxic effects, even after cleaning, storing, and reusing the object. This work analyzes the processing timeline of the filament (material for 3D printing) from unboxing to the extrusion through the nozzle. It is an important task to analyze the growth of bacteria on the 3D printed surface and in gaps between the layers. By default, the 3D-printed object is not food safe after longer usage and direct contact with food (even though they use food-safe filaments), but there are solutions for this problem. The aim of this work was to evaluate the 3D-printed object from different perspectives of food safety. Firstly, testing antimicrobial 3D printing filaments from a food safety aspect since the 3D Printed object in the food industry may have direct contact with the food. Therefore, the main purpose of the work is to reduce the microbial load on the surface of a 3D-printed part. Coating with epoxy resin was investigated, too, to see its effect on mechanical strength, thermal resistance, surface smoothness and food safety (cleanability). Another aim of this study was to test new temperature-resistant filaments and the effect of high temperature on 3D printed materials to see if they can be cleaned with boiling or similar hi-temp treatment. This work proved that all three mentioned methods could improve the food safety of the 3D printed object, but the size of this effect variates. The best result we got was with coating with epoxy resin, and the object was cleanable like any other injection molded plastic object with a smooth surface. Very good results we got by boiling the objects, and it is good to see that nowadays, more and more special filaments have a food-safe certificate and can withstand boiling temperatures too. Using antibacterial filaments reduced bacterial colonies to 1/5, but the biggest advantage of this method is that it doesn’t require any post-processing. The object is ready out of the 3D printer. Acknowledgements: The research was supported by the Hungarian and Serbian bilateral scientific and technological cooperation project funded by the Hungarian National Office for Research, Development and Innovation (NKFI, 2019-2.1.11-TÉT-2020-00249) and the Ministry of Education, Science and Technological Development of the Republic of Serbia. The authors acknowledge the Hungarian University of Agriculture and Life Sciences’s Doctoral School of Food Science for the support in this study

Keywords: food safety, 3D printing, filaments, microbial, temperature

Procedia PDF Downloads 143
501 Flame Volume Prediction and Validation for Lean Blowout of Gas Turbine Combustor

Authors: Ejaz Ahmed, Huang Yong

Abstract:

The operation of aero engines has a critical importance in the vicinity of lean blowout (LBO) limits. Lefebvre’s model of LBO based on empirical correlation has been extended to flame volume concept by the authors. The flame volume takes into account the effects of geometric configuration, the complex spatial interaction of mixing, turbulence, heat transfer and combustion processes inside the gas turbine combustion chamber. For these reasons, flame volume based LBO predictions are more accurate. Although LBO prediction accuracy has improved, it poses a challenge associated with Vf estimation in real gas turbine combustors. This work extends the approach of flame volume prediction previously based on fuel iterative approximation with cold flow simulations to reactive flow simulations. Flame volume for 11 combustor configurations has been simulated and validated against experimental data. To make prediction methodology robust as required in the preliminary design stage, reactive flow simulations were carried out with the combination of probability density function (PDF) and discrete phase model (DPM) in FLUENT 15.0. The criterion for flame identification was defined. Two important parameters i.e. critical injection diameter (Dp,crit) and critical temperature (Tcrit) were identified, and their influence on reactive flow simulation was studied for Vf estimation. Obtained results exhibit ±15% error in Vf estimation with experimental data.

Keywords: CFD, combustion, gas turbine combustor, lean blowout

Procedia PDF Downloads 268
500 An Efficient Propensity Score Method for Causal Analysis With Application to Case-Control Study in Breast Cancer Research

Authors: Ms Azam Najafkouchak, David Todem, Dorothy Pathak, Pramod Pathak, Joseph Gardiner

Abstract:

Propensity score (PS) methods have recently become the standard analysis as a tool for the causal inference in the observational studies where exposure is not randomly assigned, thus, confounding can impact the estimation of treatment effect on the outcome. For the binary outcome, the effect of treatment on the outcome can be estimated by odds ratios, relative risks, and risk differences. However, using the different PS methods may give you a different estimation of the treatment effect on the outcome. Several methods of PS analyses have been used mainly, include matching, inverse probability of weighting, stratification, and covariate adjusted on PS. Due to the dangers of discretizing continuous variables (exposure, covariates), the focus of this paper will be on how the variation in cut-points or boundaries will affect the average treatment effect (ATE) utilizing the stratification of PS method. Therefore, we are trying to avoid choosing arbitrary cut-points, instead, we continuously discretize the PS and accumulate information across all cut-points for inferences. We will use Monte Carlo simulation to evaluate ATE, focusing on two PS methods, stratification and covariate adjusted on PS. We will then show how this can be observed based on the analyses of the data from a case-control study of breast cancer, the Polish Women’s Health Study.

Keywords: average treatment effect, propensity score, stratification, covariate adjusted, monte Calro estimation, breast cancer, case_control study

Procedia PDF Downloads 107
499 A Bi-Objective Model to Optimize the Total Time and Idle Probability for Facility Location Problem Behaving as M/M/1/K Queues

Authors: Amirhossein Chambari

Abstract:

This article proposes a bi-objective model for the facility location problem subject to congestion (overcrowding). Motivated by implementations to locate servers in internet mirror sites, communication networks, one-server-systems, so on. This model consider for situations in which immobile (or fixed) service facilities are congested (or queued) by stochastic demand to behave as M/M/1/K queues. We consider for this problem two simultaneous perspectives; (1) Customers (desire to limit times of accessing and waiting for service) and (2) Service provider (desire to limit average facility idle-time). A bi-objective model is setup for facility location problem with two objective functions; (1) Minimizing sum of expected total traveling and waiting time (customers) and (2) Minimizing the average facility idle-time percentage (service provider). The proposed model belongs to the class of mixed-integer nonlinear programming models and the class of NP-hard problems. In addition, to solve the model, controlled elitist non-dominated sorting genetic algorithms (Controlled NSGA-II) and controlled elitist non-dominated ranking genetic algorithms (NRGA-I) are proposed. Furthermore, the two proposed metaheuristics algorithms are evaluated by establishing standard multiobjective metrics. Finally, the results are analyzed and some conclusions are given.

Keywords: bi-objective, facility location, queueing, controlled NSGA-II, NRGA-I

Procedia PDF Downloads 584
498 Misophonia and Its Relationship with Obsessive-compulsive Disorder, Anxiety, and Depression

Authors: Shahrzad Yektatalab, Agrin Mohammadi, Ladan Zarshenas

Abstract:

Background: Misophonia is a severe emotional response to repetitive sounds. This disorder may limit a person's communication, reduce their ability, or disrupt their social and personal life. This study aims to determine the prevalence of Misophonia and its relationship with obsessive-compulsive disorder, anxiety, and depression in undergraduate students of XXX University of Medical Sciences. Methods: The present study is an analytical descriptive study conducted in October 2020. The study samples are 390 undergraduate students of XXX University of medical sciences. A relative and systematic sampling method was used. In this study, a demographic questionnaire, misophonia questionnaire (A score of 7 or higher is considered as misophonia), Beck anxiety questionnaire, Beck depression questionnaire, and Maudsley questionnaire were used, and the data were analyzed using SPSS 24 software. Due to the non-normality of the data, the Spearman correlation coefficient was used for analyzing data. The significance level was considered less than 0.05. Results: 23.8% of students have misophonia. Of the 390 participants in the study, 93 had experienced misophonia. Among these 93 students, 37 (39.8%) had obsessive-compulsive disorder, 8 (8.6%) had anxiety and 9 (9.7%) had depression. There was a significant and direct relationship between all variables (p <0.001), which means that with the increase of each disorder, the probability of an increase in other disorders increases. Conclusion: Due to the high prevalence of misophonia in students and its direct relationship with obsessive-compulsive disorder, anxiety, and depression, It is recommended that future studies be conducted to identify ways to prevent and reduce the incidence of misophonia.

Keywords: misophonia, obsessive-compulsive disorder, anxiety, depression

Procedia PDF Downloads 63
497 Prevalence Rate and Types of the Domestic Violence Against Deaf in Iran

Authors: Hadi Farahani, Mahsa Tahzibi, Laleh Golamrej Eliasi, Mohammad Torkashvand

Abstract:

Iranian deafs are an under-researched population. The lack of research comes from the fact that if none, there are very few researchers capable of speaking sign language. The exclusion of this minority group from mainstream society often distorts the general understanding of prevalent issues of the deaf in Iran. The topic of this research was co-created through preliminary discussions with the Iranian deaf. Domestic violence then was picked up as an infrastructural issue impacting other dimensions of deaf lives such as work, education, and outside family relationships. For this purpose, we systematically searched the literature seeking a comprehensive questionnaire. We came across a 46-item standardized questionnaire measuring domestic violence in Iran. To adapt this questionnaire, we followed standard procedures reflected in another article. The inclusion criteria of the current research were married (had experienced living with a partner before) and +18-year-old deaf. Sampling was random and recruitment of the participants was through governmental or voluntary organizations for the deaf. 390 questionnaires then were analyzed through SPSS version 27. Analysis showed that the prevalence rate of domestic violence was 26% in general that emotional violence with 29% was the most prevalent type. Findings suggested that the more educated, and economically independent were the participants, the lower the probability of encountering domestic violence. Domestic violence within families where all members were deaf proved to be less usual than in families in which only the participant was deaf. Further interventional research is needed to assess how to empower the Iranian deaf regarding domestic violence.

Keywords: deaf, domestic violence, economic violence, emotional violence, physical violence, sexual violence

Procedia PDF Downloads 133
496 The Impact of Board Director Characteristics on the Quality of Information Disclosure

Authors: Guo Jinhong

Abstract:

The purpose of this study is to explore the association between board member functions and information disclosure levels. Based on the literature variables, such as the characteristics of the board of directors in the past, a single comprehensive indicator is established as a substitute variable for board functions, and the information disclosure evaluation results published by the Securities and Foundation are used to measure the information disclosure level of the company. This study focuses on companies listed on the Taiwan Stock Exchange from 2006 to 2010 and uses descriptive statistical analysis, univariate analysis, correlation analysis and ordered normal probability (Ordered Probit) regression for empirical analysis. The empirical results show that there is a significant positive correlation between the function of board members and the level of information disclosure. This study also conducts a sensitivity test and draws similar conclusions, showing that boards with better board member functions have higher levels of information disclosure. In addition, this study also found that higher board independence, lower director shareholding pledge ratio, higher director shareholding ratio, and directors with rich professional knowledge and practical experience can help improve the level of information disclosure. The empirical results of this study provide strong support for the "relative regulations to improve the level of information disclosure" formulated by the competent authorities in recent years.

Keywords: function of board members, information disclosure, securities, foundation

Procedia PDF Downloads 97
495 Comparison of Leeway Space Predictions Using Moyers and Tanaka-Johnston Upper Jaw and Lower Jaw on Batak Tribe Between Male and Female in Elementary School Students in Medan City, Sumatera Utara, Indonesia: A Cross-sectional Study

Authors: Hilda Fitria Lubis, Erna Sulistyawati

Abstract:

Objective: The study aims to compare Leeway space averages between Moyers and Tanaka-Johnston's analysis of elementary school students from the Batak tribe in Medan City. Material and Methods: The study involved 106 students from the Batak tribe elementary school in Medan, comprising 53 male and 53 female students. The samples obtained were then printed on both jaws to obtain a working model, and the mesiodistal width of the four permanent biting teeth of the lower jaw and the amount of space available on the canine-premolar region, as well as the predicted mesiodistal number of the canine-premolar on the Moyers probability table with a 75% degree of confidence and the Tanaka-Johnston formula. Results: Using Moyers analysis, students at Batak Elementary School in Medan City have an average Leeway space value of 2 mm on the upper jaw and 2.78 mm on the lower jaw. The average Leeway spatial value using Tanaka-Johnston analysis in the Batak tribe in elementary school in Medan City is 1.33 mm on the top jaw and 2.39 mm on the bottom jaw. Conclusion: According to Moyers and Tanaka-Johnsnton's analysis of both the upper and lower jaws in elementary school students of the Batak tribe in Medan City, there is a significant difference between Leeway's average space.

Keywords: leeways space, batak tribe, genders, diagnosis

Procedia PDF Downloads 35
494 Relationship between Electricity Consumption and Economic Growth: Evidence from Nigeria (1971-2012)

Authors: N. E Okoligwe, Okezie A. Ihugba

Abstract:

Few scholars disagrees that electricity consumption is an important supporting factor for economy growth. However, the relationship between electricity consumption and economy growth has different manifestation in different countries according to previous studies. This paper examines the causal relationship between electricity consumption and economic growth for Nigeria. In an attempt to do this, the paper tests the validity of the modernization or depending hypothesis by employing various econometric tools such as Augmented Dickey Fuller (ADF) and Johansen Co-integration test, the Error Correction Mechanism (ECM) and Granger Causality test on time series data from 1971-2012. The Granger causality is found not to run from electricity consumption to real GDP and from GDP to electricity consumption during the year of study. The null hypothesis is accepted at the 5 per cent level of significance where the probability value (0.2251 and 0.8251) is greater than five per cent level of significance because both of them are probably determined by some other factors like; increase in urban population, unemployment rate and the number of Nigerians that benefit from the increase in GDP and increase in electricity demand is not determined by the increase in GDP (income) over the period of study because electricity demand has always been greater than consumption. Consequently; the policy makers in Nigeria should place priority in early stages of reconstruction on building capacity additions and infrastructure development of the electric power sector as this would force the sustainable economic growth in Nigeria.

Keywords: economic growth, electricity consumption, error correction mechanism, granger causality test

Procedia PDF Downloads 311
493 Macular Ganglion Cell Inner Plexiform Layer Thinning

Authors: Hye-Young Shin, Chan Kee Park

Abstract:

Background: To compare the thinning patterns of the ganglion cell-inner plexiform layer (GCIPL) and peripapillary retinal nerve fiber layer (pRNFL) as measured using Cirrus high-definition optical coherence tomography (HD-OCT) in patients with visual field (VF) defects that respect the vertical meridian. Methods: Twenty eyes of eleven patients with VF defects that respect the vertical meridian were enrolled retrospectively. The thicknesses of the macular GCIPL and pRNFL were measured using Cirrus HD-OCT. The 5% and 1% thinning area index (TAI) was calculated as the proportion of abnormally thin sectors at the 5% and 1% probability level within the area corresponding to the affected VF. The 5% and 1% TAI were compared between the GCIPL and pRNFL measurements. Results: The color-coded GCIPL deviation map showed a characteristic vertical thinning pattern of the GCIPL, which is also seen in the VF of patients with brain lesions. The 5% and 1% TAI were significantly higher in the GCIPL measurements than in the pRNFL measurements (all P < 0.01). Conclusions: Macular GCIPL analysis clearly visualized a characteristic topographic pattern of retinal ganglion cell (RGC) loss in patients with VF defects that respect the vertical meridian, unlike pRNFL measurements. Macular GCIPL measurements provide more valuable information than pRNFL measurements for detecting the loss of RGCs in patients with retrograde degeneration of the optic nerve fibers.

Keywords: brain lesion, macular ganglion cell, inner plexiform layer, spectral-domain optical coherence tomography

Procedia PDF Downloads 337
492 Educational Leadership and Artificial Intelligence

Authors: Sultan Ghaleb Aldaihani

Abstract:

- The environment in which educational leadership takes place is becoming increasingly complex due to factors like globalization and rapid technological change. - This is creating a "leadership gap" where the complexity of the environment outpaces the ability of leaders to effectively respond. - Educational leadership involves guiding teachers and the broader school system towards improved student learning and achievement. 2. Implications of Artificial Intelligence (AI) in Educational Leadership: - AI has great potential to enhance education, such as through intelligent tutoring systems and automating routine tasks to free up teachers. - AI can also have significant implications for educational leadership by providing better information and data-driven decision-making capabilities. - Computer-adaptive testing can provide detailed, individualized data on student learning that leaders can use for instructional decisions and accountability. 3. Enhancing Decision-Making Processes: - Statistical models and data mining techniques can help identify at-risk students earlier, allowing for targeted interventions. - Probability-based models can diagnose students likely to drop out, enabling proactive support. - These data-driven approaches can make resource allocation and decision-making more effective. 4. Improving Efficiency and Productivity: - AI systems can automate tasks and change processes to improve the efficiency of educational leadership and administration. - Integrating AI can free up leaders to focus more on their role's human, interactive elements.

Keywords: Education, Leadership, Technology, Artificial Intelligence

Procedia PDF Downloads 45
491 Vehicle Timing Motion Detection Based on Multi-Dimensional Dynamic Detection Network

Authors: Jia Li, Xing Wei, Yuchen Hong, Yang Lu

Abstract:

Detecting vehicle behavior has always been the focus of intelligent transportation, but with the explosive growth of the number of vehicles and the complexity of the road environment, the vehicle behavior videos captured by traditional surveillance have been unable to satisfy the study of vehicle behavior. The traditional method of manually labeling vehicle behavior is too time-consuming and labor-intensive, but the existing object detection and tracking algorithms have poor practicability and low behavioral location detection rate. This paper proposes a vehicle behavior detection algorithm based on the dual-stream convolution network and the multi-dimensional video dynamic detection network. In the videos, the straight-line behavior of the vehicle will default to the background behavior. The Changing lanes, turning and turning around are set as target behaviors. The purpose of this model is to automatically mark the target behavior of the vehicle from the untrimmed videos. First, the target behavior proposals in the long video are extracted through the dual-stream convolution network. The model uses a dual-stream convolutional network to generate a one-dimensional action score waveform, and then extract segments with scores above a given threshold M into preliminary vehicle behavior proposals. Second, the preliminary proposals are pruned and identified using the multi-dimensional video dynamic detection network. Referring to the hierarchical reinforcement learning, the multi-dimensional network includes a Timer module and a Spacer module, where the Timer module mines time information in the video stream and the Spacer module extracts spatial information in the video frame. The Timer and Spacer module are implemented by Long Short-Term Memory (LSTM) and start from an all-zero hidden state. The Timer module uses the Transformer mechanism to extract timing information from the video stream and extract features by linear mapping and other methods. Finally, the model fuses time information and spatial information and obtains the location and category of the behavior through the softmax layer. This paper uses recall and precision to measure the performance of the model. Extensive experiments show that based on the dataset of this paper, the proposed model has obvious advantages compared with the existing state-of-the-art behavior detection algorithms. When the Time Intersection over Union (TIoU) threshold is 0.5, the Average-Precision (MP) reaches 36.3% (the MP of baselines is 21.5%). In summary, this paper proposes a vehicle behavior detection model based on multi-dimensional dynamic detection network. This paper introduces spatial information and temporal information to extract vehicle behaviors in long videos. Experiments show that the proposed algorithm is advanced and accurate in-vehicle timing behavior detection. In the future, the focus will be on simultaneously detecting the timing behavior of multiple vehicles in complex traffic scenes (such as a busy street) while ensuring accuracy.

Keywords: vehicle behavior detection, convolutional neural network, long short-term memory, deep learning

Procedia PDF Downloads 132
490 Reconstructability Analysis for Landslide Prediction

Authors: David Percy

Abstract:

Landslides are a geologic phenomenon that affects a large number of inhabited places and are constantly being monitored and studied for the prediction of future occurrences. Reconstructability analysis (RA) is a methodology for extracting informative models from large volumes of data that work exclusively with discrete data. While RA has been used in medical applications and social science extensively, we are introducing it to the spatial sciences through applications like landslide prediction. Since RA works exclusively with discrete data, such as soil classification or bedrock type, working with continuous data, such as porosity, requires that these data are binned for inclusion in the model. RA constructs models of the data which pick out the most informative elements, independent variables (IVs), from each layer that predict the dependent variable (DV), landslide occurrence. Each layer included in the model retains its classification data as a primary encoding of the data. Unlike other machine learning algorithms that force the data into one-hot encoding type of schemes, RA works directly with the data as it is encoded, with the exception of continuous data, which must be binned. The usual physical and derived layers are included in the model, and testing our results against other published methodologies, such as neural networks, yields accuracy that is similar but with the advantage of a completely transparent model. The results of an RA session with a data set are a report on every combination of variables and their probability of landslide events occurring. In this way, every combination of informative state combinations can be examined.

Keywords: reconstructability analysis, machine learning, landslides, raster analysis

Procedia PDF Downloads 68
489 Effect of Fortification of Expressed Human Breast Milk with Olive Oil and Skimmed Milk in Improving Weight Gain in Very Low Birth Weight Neonates and Shortening Their Length of Hospital Stay

Authors: Sumrina Kousar

Abstract:

Objective: The aim of this study was to observe the effect of fortification of expressed human breast milk with olive oil and skimmed milk in improving weight gain in very low birth weight neonates and shortening their length of hospital stay. Study Design and place: A randomized controlled trial was carried out at the Combined Military Hospital Lahore from March 2018 to March 2019. Methods: Neonates admitted with very low birth weight and gestational age of < 34 weeks were included in the study. Sixty babies were enrolled using non-probability consecutive sampling; a random number table was used to allocate them into a fortification group and a control group. The control group received expressed milk alone, while olive oil 1 ml twice daily and skimmed milk 1 gram in every third feed were added to expressed milk in the fortification group. Data was analyzed on SPSS 20. Proportions were compared by applying the chi-square test. An independent sample t-test was applied for comparing means. A p-value of ≤ 0.05 was considered significant. Results: The study comprised of 60 neonates, with 30 in each of the groups. Weight gain was 24.83±5.63 in the fortification group and 11.72±3.95 in the control group (p =< 0.001). Mean hospital stay was 20.5716.511 in the fortification group and 27.678.89 in the control group (p =< 0.043). Conclusion: Olive oil and skimmed milk fortification of breast milk was effective for weight gain and reducing the length of hospital stay in very low birth weight neonates.

Keywords: fortification, olive oil, skimmed milk, weight gain

Procedia PDF Downloads 169
488 Impact of Drought on Agriculture in the Upper Middle Gangetic Plain in India

Authors: Reshmita Nath

Abstract:

In this study, we investigate the spatiotemporal characteristics of drought in India and its impact on agriculture during the summer season (April to September). For our analysis, we have used Standardized Precipitation Evapotranspiration Index (SPEI) datasets between 1982 and 2012 at six-month timescale. Based on the criteria SPEI<-1 we obtain the vulnerability map and have found that the Humid subtropical Upper Middle Gangetic Plain (UMGP) region is highly drought prone with an occurrence frequency of 40-45%. This UMGP region contributes at least 18-20% of India’s annual cereal production. Not only the probability, but the region becomes more and more drought-prone in the recent decades. Moreover, the cereal production in the UMGP has experienced a gradual declining trend from 2000 onwards and this feature is consistent with the increase in drought affected areas from 20-25% to 50-60%, before and after 2000, respectively. The higher correlation coefficient (-0.69) between the changes in cereal production and drought affected areas confirms that at least 50% of the agricultural (cereal) losses is associated with drought. While analyzing the individual impact of precipitation and surface temperature anomalies on SPEI (6), we have found that in the UMGP region surface temperature plays the primary role in lowering of SPEI. The linkage is further confirmed by the correlation analysis between the SPEI (6) and surface temperature rise, which exhibits strong negative values in the UMGP region. Higher temperature might have caused more evaporation and drying, which therefore increases the area affected by drought in the recent decade.

Keywords: drought, agriculture, SPEI, Indo-Gangetic plain

Procedia PDF Downloads 258
487 Integrated Clean Development Mechanism and Risk Management Approach for Infrastructure Transportation Project

Authors: Debasis Sarkar

Abstract:

Clean development mechanism (CDM) can act as an effective instrument for mitigating climate change. This mechanism can effectively reduce the emission of CO2 and other green house gases (GHG). Construction of a mega infrastructure project like underground corridor construction for metro rail operation involves in consumption of substantial quantity of concrete which consumes huge quantity of energy consuming materials like cement and steel. This paper is an attempt to develop an integrated clean development mechanism and risk management approach for sustainable development for an underground corridor metro rail project in India during its construction phase. It was observed that about 35% reduction in CO2 emission can be obtained by adding fly ash as a part replacement of cement. The reduced emission quantity of CO2 which is of the quantum of about 21,646.36 MT would result in cost savings of approximately INR 8.5 million (USD 1,29,878).But construction and operation of such infrastructure projects of the present era are subject to huge risks and uncertainties throughout all the phases of the project, thus reducing the probability of successful completion of the project within stipulated time and cost frame. Thus, an integrated approach of combining CDM with risk management would enable the metro rail authorities to develop a sustainable risk mitigation measure framework to ensure more cost and energy savings and lesser time and cost over-run.

Keywords: clean development mechanism (CDM), infrastructure transportation, project risk management, underground metro rail

Procedia PDF Downloads 476
486 Patients' Interpretation of Prescribed Medication Instructions: A Pilot Study among Diabetes Mellitus Patients at Makanye Clinic in Limpopo Province, South Africa

Authors: Charity Ngoatle, Tebogo M. Mothiba, Mahlapahlapana J. Themane

Abstract:

Misapprehension of medications instructions due to poor health literacy is common in diabetic patients, predominantly leading to suboptimal medication therapy caused by taking less than expected, or getting inadequate medication concentration. Globally, 50% of adults have been reported to have misunderstood medication instructions which could be the cause of not using medication as prescribed. Reading material has been found not to improve people’s knowledge to the extent where they would be informed and knowledgeable about their health. This, therefore, depicts that instructive materials alone cannot improve health literacy but further patient education is still needed to explain what the information really mean. The aim of this study was to investigate patients’ interpretation of prescribed medication instructions at Makanye Clinic in Limpopo Province, South Africa. The study used a mixed method approach. A non-probability purposive and simple random sampling strategies will be used to select ten (10) participants for the pilot study. Semi-structured interviews with a guide and self- administered structured questionnaires will be used to collect data. Tesch’s eight steps for qualitative data analysis and SPSS version 24 with descriptive statistics will be adopted. The preliminary findings from other studies show that: (a) poor health literacy negatively affect medication adherence, (b) general literacy influence health literacy, and (c) there are poor health outcomes and medication adverse effects due to poor medication comprehension.

Keywords: instructions, diabetes mellitus, patients, prescribed medication

Procedia PDF Downloads 138
485 Landfill Failure Mobility Analysis: A Probabilistic Approach

Authors: Ali Jahanfar, Brajesh Dubey, Bahram Gharabaghi, Saber Bayat Movahed

Abstract:

Ever increasing population growth of major urban centers and environmental challenges in siting new landfills have resulted in a growing trend in design of mega-landfills some with extraordinary heights and dangerously steep slopes. Landfill failure mobility risk analysis is one of the most uncertain types of dynamic rheology models due to very large inherent variabilities in the heterogeneous solid waste material shear strength properties. The waste flow of three historic dumpsite and two landfill failures were back-analyzed using run-out modeling with DAN-W model. The travel distances of the waste flow during landfill failures were calculated approach by taking into account variability in material shear strength properties. The probability distribution function for shear strength properties of the waste material were grouped into four major classed based on waste material compaction (landfills versus dumpsites) and composition (high versus low quantity) of high shear strength waste materials such as wood, metal, plastic, paper and cardboard in the waste. This paper presents a probabilistic method for estimation of the spatial extent of waste avalanches, after a potential landfill failure, to create maps of vulnerability scores to inform property owners and residents of the level of the risk.

Keywords: landfill failure, waste flow, Voellmy rheology, friction coefficient, waste compaction and type

Procedia PDF Downloads 292
484 Kannada HandWritten Character Recognition by Edge Hinge and Edge Distribution Techniques Using Manhatan and Minimum Distance Classifiers

Authors: C. V. Aravinda, H. N. Prakash

Abstract:

In this paper, we tried to convey fusion and state of art pertaining to SIL character recognition systems. In the first step, the text is preprocessed and normalized to perform the text identification correctly. The second step involves extracting relevant and informative features. The third step implements the classification decision. The three stages which involved are Data acquisition and preprocessing, Feature extraction, and Classification. Here we concentrated on two techniques to obtain features, Feature Extraction & Feature Selection. Edge-hinge distribution is a feature that characterizes the changes in direction of a script stroke in handwritten text. The edge-hinge distribution is extracted by means of a windowpane that is slid over an edge-detected binary handwriting image. Whenever the mid pixel of the window is on, the two edge fragments (i.e. connected sequences of pixels) emerging from this mid pixel are measured. Their directions are measured and stored as pairs. A joint probability distribution is obtained from a large sample of such pairs. Despite continuous effort, handwriting identification remains a challenging issue, due to different approaches use different varieties of features, having different. Therefore, our study will focus on handwriting recognition based on feature selection to simplify features extracting task, optimize classification system complexity, reduce running time and improve the classification accuracy.

Keywords: word segmentation and recognition, character recognition, optical character recognition, hand written character recognition, South Indian languages

Procedia PDF Downloads 497
483 Risk Management in Industrial Supervision Projects

Authors: Érick Aragão Ribeiro, George André Pereira Thé, José Marques Soares

Abstract:

Several problems in industrial supervision software development projects may lead to the delay or cancellation of projects. These problems can be avoided or contained by using identification methods, analysis and control of risks. These procedures can give an overview of the possible problems that can happen in the projects and what are the immediate solutions. Therefore, we propose a risk management method applied to the teaching and development of industrial supervision software. The method is developed through a literature review and previous projects can be divided into phases of management and have basic features that are validated with experimental research carried out by mechatronics engineering students and professionals. The management is conducted through the stages of identification, analysis, planning, monitoring, control and communication of risks. Programmers use a method of prioritizing risks considering the gravity and the possibility of occurrence of the risk. The outputs of the method indicate which risks occurred or are about to happen. The first results indicate which risks occur at different stages of the project and what risks have a high probability of occurring. The results show the efficiency of the proposed method compared to other methods, showing the improvement of software quality and leading developers in their decisions. This new way of developing supervision software helps students identify design problems, evaluate software developed and propose effective solutions. We conclude that the risk management optimizes the development of the industrial process control software and provides higher quality to the product.

Keywords: supervision software, risk management, industrial supervision, project management

Procedia PDF Downloads 359
482 Reliability Modeling on Drivers’ Decision during Yellow Phase

Authors: Sabyasachi Biswas, Indrajit Ghosh

Abstract:

The random and heterogeneous behavior of vehicles in India puts up a greater challenge for researchers. Stop-and-go modeling at signalized intersections under heterogeneous traffic conditions has remained one of the most sought-after fields. Vehicles are often caught up in the dilemma zone and are unable to take quick decisions whether to stop or cross the intersection. This hampers the traffic movement and may lead to accidents. The purpose of this work is to develop a stop and go prediction model that depicts the drivers’ decision during the yellow time at signalised intersections. To accomplish this, certain traffic parameters were taken into account to develop surrogate model. This research investigated the Stop and Go behavior of the drivers by collecting data from 4-signalized intersections located in two major Indian cities. Model was developed to predict the drivers’ decision making during the yellow phase of the traffic signal. The parameters used for modeling included distance to stop line, time to stop line, speed, and length of the vehicle. A Kriging base surrogate model has been developed to investigate the drivers’ decision-making behavior in amber phase. It is observed that the proposed approach yields a highly accurate result (97.4 percent) by Gaussian function. It was observed that the accuracy for the crossing probability was 95.45, 90.9 and 86.36.11 percent respectively as predicted by the Kriging models with Gaussian, Exponential and Linear functions.

Keywords: decision-making decision, dilemma zone, surrogate model, Kriging

Procedia PDF Downloads 309
481 Sub-Pixel Mapping Based on New Mixed Interpolation

Authors: Zeyu Zhou, Xiaojun Bi

Abstract:

Due to the limited environmental parameters and the limited resolution of the sensor, the universal existence of the mixed pixels in the process of remote sensing images restricts the spatial resolution of the remote sensing images. Sub-pixel mapping technology can effectively improve the spatial resolution. As the bilinear interpolation algorithm inevitably produces the edge blur effect, which leads to the inaccurate sub-pixel mapping results. In order to avoid the edge blur effect that affects the sub-pixel mapping results in the interpolation process, this paper presents a new edge-directed interpolation algorithm which uses the covariance adaptive interpolation algorithm on the edge of the low-resolution image and uses bilinear interpolation algorithm in the low-resolution image smooth area. By using the edge-directed interpolation algorithm, the super-resolution of the image with low resolution is obtained, and we get the percentage of each sub-pixel under a certain type of high-resolution image. Then we rely on the probability value as a soft attribute estimate and carry out sub-pixel scale under the ‘hard classification’. Finally, we get the result of sub-pixel mapping. Through the experiment, we compare the algorithm and the bilinear algorithm given in this paper to the results of the sub-pixel mapping method. It is found that the sub-pixel mapping method based on the edge-directed interpolation algorithm has better edge effect and higher mapping accuracy. The results of the paper meet our original intention of the question. At the same time, the method does not require iterative computation and training of samples, making it easier to implement.

Keywords: remote sensing images, sub-pixel mapping, bilinear interpolation, edge-directed interpolation

Procedia PDF Downloads 230
480 Peak Frequencies in the Collective Membrane Potential of a Hindmarsh-Rose Small-World Neural Network

Authors: Sun Zhe, Ruggero Micheletto

Abstract:

As discussed extensively in many studies, noise in neural networks have an important role in the functioning and time evolution of the system. The mechanism by which noise induce stochastic resonance enhancing and influencing certain operations is not clarified nor is the mechanism of information storage and coding. With the present research we want to study the role of noise, especially focusing on the frequency peaks in a three variable Hindmarsh−Rose Small−World network. We investigated the behaviour of the network to external noises. We demonstrate that a variation of signal to noise ratio of about 10 dB induces an increase in membrane potential signal of about 15%, averaged over the whole network. We also considered the integral of the whole membrane potential as a paradigm of internal noise, the one generated by the brain network. We showed that this internal noise is attenuated with the size of the network or with the number of random connections. By means of Fourier analysis we found that it has distinct peaks of frequencies, moreover, we showed that increasing the size of the network introducing more neurons, reduced the maximum frequencies generated by the network, whereas the increase in the number of random connections (determined by the small-world probability p) led to a trend toward higher frequencies. This study may give clues on how networks utilize noise to alter the collective behaviour of the system in their operations.

Keywords: neural networks, stochastic processes, small-world networks, discrete Fourier analysis

Procedia PDF Downloads 292
479 Project HDMI: A Hybrid-Differentiated Mathematics Instruction for Grade 11 Senior High School Students at Las Piñas City Technical Vocational High School

Authors: Mary Ann Cristine R. Olgado

Abstract:

Diversity in the classroom might make it difficult to promote individualized learning, but differentiated instruction that caters to students' various learning preferences may prove to be beneficial. Hence, this study examined the effectiveness of Hybrid-Differentiated Mathematics Instruction (HDMI) in improving the students’ academic performance in Mathematics. It employed the quasi-experimental research design by using a comparative analysis of the two variables: the experimental and control groups. The learning styles of the students were identified using the Grasha-Riechmann Student Learning Style Scale (GRSLSS), which served as the basis for designing differentiated action plans in Mathematics. In addition, adapted survey questionnaires, pre-tests, and post-tests were used to gather information and were analyzed using descriptive and correlational statistics to find the relationship between variables. The experimental group received differentiated instruction for a month, while the control group received traditional teaching instruction. The study found that Hybrid-Differentiated Mathematics Instruction (HDMI) improved the academic performance of Grade 11-TVL students, with the experimental group performing better than the control group. This program has effectively tailored the teaching methods to meet the diverse learning needs of the students, fostering and enhancing a deeper understanding of mathematical concepts in Statistics & Probability, both within and beyond the classroom.

Keywords: differentiated instruction, hybrid, learning styles, academic performance

Procedia PDF Downloads 62
478 Efficient Estimation for the Cox Proportional Hazards Cure Model

Authors: Khandoker Akib Mohammad

Abstract:

While analyzing time-to-event data, it is possible that a certain fraction of subjects will never experience the event of interest, and they are said to be cured. When this feature of survival models is taken into account, the models are commonly referred to as cure models. In the presence of covariates, the conditional survival function of the population can be modelled by using the cure model, which depends on the probability of being uncured (incidence) and the conditional survival function of the uncured subjects (latency), and a combination of logistic regression and Cox proportional hazards (PH) regression is used to model the incidence and latency respectively. In this paper, we have shown the asymptotic normality of the profile likelihood estimator via asymptotic expansion of the profile likelihood and obtain the explicit form of the variance estimator with an implicit function in the profile likelihood. We have also shown the efficient score function based on projection theory and the profile likelihood score function are equal. Our contribution in this paper is that we have expressed the efficient information matrix as the variance of the profile likelihood score function. A simulation study suggests that the estimated standard errors from bootstrap samples (SMCURE package) and the profile likelihood score function (our approach) are providing similar and comparable results. The numerical result of our proposed method is also shown by using the melanoma data from SMCURE R-package, and we compare the results with the output obtained from the SMCURE package.

Keywords: Cox PH model, cure model, efficient score function, EM algorithm, implicit function, profile likelihood

Procedia PDF Downloads 146