Search results for: defect prediction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2598

Search results for: defect prediction

1638 Interpretable Deep Learning Models for Medical Condition Identification

Authors: Dongping Fang, Lian Duan, Xiaojing Yuan, Mike Xu, Allyn Klunder, Kevin Tan, Suiting Cao, Yeqing Ji

Abstract:

Accurate prediction of a medical condition with straight clinical evidence is a long-sought topic in the medical management and health insurance field. Although great progress has been made with machine learning algorithms, the medical community is still, to a certain degree, suspicious about the model's accuracy and interpretability. This paper presents an innovative hierarchical attention deep learning model to achieve good prediction and clear interpretability that can be easily understood by medical professionals. This deep learning model uses a hierarchical attention structure that matches naturally with the medical history data structure and reflects the member’s encounter (date of service) sequence. The model attention structure consists of 3 levels: (1) attention on the medical code types (diagnosis codes, procedure codes, lab test results, and prescription drugs), (2) attention on the sequential medical encounters within a type, (3) attention on the medical codes within an encounter and type. This model is applied to predict the occurrence of stage 3 chronic kidney disease (CKD3), using three years’ medical history of Medicare Advantage (MA) members from a top health insurance company. The model takes members’ medical events, both claims and electronic medical record (EMR) data, as input, makes a prediction of CKD3 and calculates the contribution from individual events to the predicted outcome. The model outcome can be easily explained with the clinical evidence identified by the model algorithm. Here are examples: Member A had 36 medical encounters in the past three years: multiple office visits, lab tests and medications. The model predicts member A has a high risk of CKD3 with the following well-contributed clinical events - multiple high ‘Creatinine in Serum or Plasma’ tests and multiple low kidneys functioning ‘Glomerular filtration rate’ tests. Among the abnormal lab tests, more recent results contributed more to the prediction. The model also indicates regular office visits, no abnormal findings of medical examinations, and taking proper medications decreased the CKD3 risk. Member B had 104 medical encounters in the past 3 years and was predicted to have a low risk of CKD3, because the model didn’t identify diagnoses, procedures, or medications related to kidney disease, and many lab test results, including ‘Glomerular filtration rate’ were within the normal range. The model accurately predicts members A and B and provides interpretable clinical evidence that is validated by clinicians. Without extra effort, the interpretation is generated directly from the model and presented together with the occurrence date. Our model uses the medical data in its most raw format without any further data aggregation, transformation, or mapping. This greatly simplifies the data preparation process, mitigates the chance for error and eliminates post-modeling work needed for traditional model explanation. To our knowledge, this is the first paper on an interpretable deep-learning model using a 3-level attention structure, sourcing both EMR and claim data, including all 4 types of medical data, on the entire Medicare population of a big insurance company, and more importantly, directly generating model interpretation to support user decision. In the future, we plan to enrich the model input by adding patients’ demographics and information from free-texted physician notes.

Keywords: deep learning, interpretability, attention, big data, medical conditions

Procedia PDF Downloads 88
1637 Machine Learning Approaches Based on Recency, Frequency, Monetary (RFM) and K-Means for Predicting Electrical Failures and Voltage Reliability in Smart Cities

Authors: Panaya Sudta, Wanchalerm Patanacharoenwong, Prachya Bumrungkun

Abstract:

As With the evolution of smart grids, ensuring the reliability and efficiency of electrical systems in smart cities has become crucial. This paper proposes a distinct approach that combines advanced machine learning techniques to accurately predict electrical failures and address voltage reliability issues. This approach aims to improve the accuracy and efficiency of reliability evaluations in smart cities. The aim of this research is to develop a comprehensive predictive model that accurately predicts electrical failures and voltage reliability in smart cities. This model integrates RFM analysis, K-means clustering, and LSTM networks to achieve this objective. The research utilizes RFM analysis, traditionally used in customer value assessment, to categorize and analyze electrical components based on their failure recency, frequency, and monetary impact. K-means clustering is employed to segment electrical components into distinct groups with similar characteristics and failure patterns. LSTM networks are used to capture the temporal dependencies and patterns in customer data. This integration of RFM, K-means, and LSTM results in a robust predictive tool for electrical failures and voltage reliability. The proposed model has been tested and validated on diverse electrical utility datasets. The results show a significant improvement in prediction accuracy and reliability compared to traditional methods, achieving an accuracy of 92.78% and an F1-score of 0.83. This research contributes to the proactive maintenance and optimization of electrical infrastructures in smart cities. It also enhances overall energy management and sustainability. The integration of advanced machine learning techniques in the predictive model demonstrates the potential for transforming the landscape of electrical system management within smart cities. The research utilizes diverse electrical utility datasets to develop and validate the predictive model. RFM analysis, K-means clustering, and LSTM networks are applied to these datasets to analyze and predict electrical failures and voltage reliability. The research addresses the question of how accurately electrical failures and voltage reliability can be predicted in smart cities. It also investigates the effectiveness of integrating RFM analysis, K-means clustering, and LSTM networks in achieving this goal. The proposed approach presents a distinct, efficient, and effective solution for predicting and mitigating electrical failures and voltage issues in smart cities. It significantly improves prediction accuracy and reliability compared to traditional methods. This advancement contributes to the proactive maintenance and optimization of electrical infrastructures, overall energy management, and sustainability in smart cities.

Keywords: electrical state prediction, smart grids, data-driven method, long short-term memory, RFM, k-means, machine learning

Procedia PDF Downloads 49
1636 Heart Rate Variability Analysis for Early Stage Prediction of Sudden Cardiac Death

Authors: Reeta Devi, Hitender Kumar Tyagi, Dinesh Kumar

Abstract:

In present scenario, cardiovascular problems are growing challenge for researchers and physiologists. As heart disease have no geographic, gender or socioeconomic specific reasons; detecting cardiac irregularities at early stage followed by quick and correct treatment is very important. Electrocardiogram is the finest tool for continuous monitoring of heart activity. Heart rate variability (HRV) is used to measure naturally occurring oscillations between consecutive cardiac cycles. Analysis of this variability is carried out using time domain, frequency domain and non-linear parameters. This paper presents HRV analysis of the online dataset for normal sinus rhythm (taken as healthy subject) and sudden cardiac death (SCD subject) using all three methods computing values for parameters like standard deviation of node to node intervals (SDNN), square root of mean of the sequences of difference between adjacent RR intervals (RMSSD), mean of R to R intervals (mean RR) in time domain, very low-frequency (VLF), low-frequency (LF), high frequency (HF) and ratio of low to high frequency (LF/HF ratio) in frequency domain and Poincare plot for non linear analysis. To differentiate HRV of healthy subject from subject died with SCD, k –nearest neighbor (k-NN) classifier has been used because of its high accuracy. Results show highly reduced values for all stated parameters for SCD subjects as compared to healthy ones. As the dataset used for SCD patients is recording of their ECG signal one hour prior to their death, it is therefore, verified with an accuracy of 95% that proposed algorithm can identify mortality risk of a patient one hour before its death. The identification of a patient’s mortality risk at such an early stage may prevent him/her meeting sudden death if in-time and right treatment is given by the doctor.

Keywords: early stage prediction, heart rate variability, linear and non-linear analysis, sudden cardiac death

Procedia PDF Downloads 337
1635 Implementation of Deep Neural Networks for Pavement Condition Index Prediction

Authors: M. Sirhan, S. Bekhor, A. Sidess

Abstract:

In-service pavements deteriorate with time due to traffic wheel loads, environment, and climate conditions. Pavement deterioration leads to a reduction in their serviceability and structural behavior. Consequently, proper maintenance and rehabilitation (M&R) are necessary actions to keep the in-service pavement network at the desired level of serviceability. Due to resource and financial constraints, the pavement management system (PMS) prioritizes roads most in need of maintenance and rehabilitation action. It recommends a suitable action for each pavement based on the performance and surface condition of each road in the network. The pavement performance and condition are usually quantified and evaluated by different types of roughness-based and stress-based indices. Examples of such indices are Pavement Serviceability Index (PSI), Pavement Serviceability Ratio (PSR), Mean Panel Rating (MPR), Pavement Condition Rating (PCR), Ride Number (RN), Profile Index (PI), International Roughness Index (IRI), and Pavement Condition Index (PCI). PCI is commonly used in PMS as an indicator of the extent of the distresses on the pavement surface. PCI values range between 0 and 100; where 0 and 100 represent a highly deteriorated pavement and a newly constructed pavement, respectively. The PCI value is a function of distress type, severity, and density (measured as a percentage of the total pavement area). PCI is usually calculated iteratively using the 'Paver' program developed by the US Army Corps. The use of soft computing techniques, especially Artificial Neural Network (ANN), has become increasingly popular in the modeling of engineering problems. ANN techniques have successfully modeled the performance of the in-service pavements, due to its efficiency in predicting and solving non-linear relationships and dealing with an uncertain large amount of data. Typical regression models, which require a pre-defined relationship, can be replaced by ANN, which was found to be an appropriate tool for predicting the different pavement performance indices versus different factors as well. Subsequently, the objective of the presented study is to develop and train an ANN model that predicts the PCI values. The model’s input consists of percentage areas of 11 different damage types; alligator cracking, swelling, rutting, block cracking, longitudinal/transverse cracking, edge cracking, shoving, raveling, potholes, patching, and lane drop off, at three severity levels (low, medium, high) for each. The developed model was trained using 536,000 samples and tested on 134,000 samples. The samples were collected and prepared by The National Transport Infrastructure Company. The predicted results yielded satisfactory compliance with field measurements. The proposed model predicted PCI values with relatively low standard deviations, suggesting that it could be incorporated into the PMS for PCI determination. It is worth mentioning that the most influencing variables for PCI prediction are damages related to alligator cracking, swelling, rutting, and potholes.

Keywords: artificial neural networks, computer programming, pavement condition index, pavement management, performance prediction

Procedia PDF Downloads 131
1634 Validation of Nutritional Assessment Scores in Prediction of Mortality and Duration of Admission in Elderly, Hospitalized Patients: A Cross-Sectional Study

Authors: Christos Lampropoulos, Maria Konsta, Vicky Dradaki, Irini Dri, Konstantina Panouria, Tamta Sirbilatze, Ifigenia Apostolou, Vaggelis Lambas, Christina Kordali, Georgios Mavras

Abstract:

Objectives: Malnutrition in hospitalized patients is related to increased morbidity and mortality. The purpose of our study was to compare various nutritional scores in order to detect the most suitable one for assessing the nutritional status of elderly, hospitalized patients and correlate them with mortality and extension of admission duration, due to patients’ critical condition. Methods: Sample population included 150 patients (78 men, 72 women, mean age 80±8.2). Nutritional status was assessed by Mini Nutritional Assessment (MNA full, short-form), Malnutrition Universal Screening Tool (MUST) and short Nutritional Appetite Questionnaire (sNAQ). Sensitivity, specificity, positive and negative predictive values and ROC curves were assessed after adjustment for the cause of current admission, a known prognostic factor according to previously applied multivariate models. Primary endpoints were mortality (from admission until 6 months afterwards) and duration of hospitalization, compared to national guidelines for closed consolidated medical expenses. Results: Concerning mortality, MNA (short-form and full) and SNAQ had similar, low sensitivity (25.8%, 25.8% and 35.5% respectively) while MUST had higher sensitivity (48.4%). In contrast, all the questionnaires had high specificity (94%-97.5%). Short-form MNA and sNAQ had the best positive predictive value (72.7% and 78.6% respectively) whereas all the questionnaires had similar negative predictive value (83.2%-87.5%). MUST had the highest ROC curve (0.83) in contrast to the rest questionnaires (0.73-0.77). With regard to extension of admission duration, all four scores had relatively low sensitivity (48.7%-56.7%), specificity (68.4%-77.6%), positive predictive value (63.1%-69.6%), negative predictive value (61%-63%) and ROC curve (0.67-0.69). Conclusion: MUST questionnaire is more advantageous in predicting mortality due to its higher sensitivity and ROC curve. None of the nutritional scores is suitable for prediction of extended hospitalization.

Keywords: duration of admission, malnutrition, nutritional assessment scores, prognostic factors for mortality

Procedia PDF Downloads 339
1633 Advancements in Predicting Diabetes Biomarkers: A Machine Learning Epigenetic Approach

Authors: James Ladzekpo

Abstract:

Background: The urgent need to identify new pharmacological targets for diabetes treatment and prevention has been amplified by the disease's extensive impact on individuals and healthcare systems. A deeper insight into the biological underpinnings of diabetes is crucial for the creation of therapeutic strategies aimed at these biological processes. Current predictive models based on genetic variations fall short of accurately forecasting diabetes. Objectives: Our study aims to pinpoint key epigenetic factors that predispose individuals to diabetes. These factors will inform the development of an advanced predictive model that estimates diabetes risk from genetic profiles, utilizing state-of-the-art statistical and data mining methods. Methodology: We have implemented a recursive feature elimination with cross-validation using the support vector machine (SVM) approach for refined feature selection. Building on this, we developed six machine learning models, including logistic regression, k-Nearest Neighbors (k-NN), Naive Bayes, Random Forest, Gradient Boosting, and Multilayer Perceptron Neural Network, to evaluate their performance. Findings: The Gradient Boosting Classifier excelled, achieving a median recall of 92.17% and outstanding metrics such as area under the receiver operating characteristics curve (AUC) with a median of 68%, alongside median accuracy and precision scores of 76%. Through our machine learning analysis, we identified 31 genes significantly associated with diabetes traits, highlighting their potential as biomarkers and targets for diabetes management strategies. Conclusion: Particularly noteworthy were the Gradient Boosting Classifier and Multilayer Perceptron Neural Network, which demonstrated potential in diabetes outcome prediction. We recommend future investigations to incorporate larger cohorts and a wider array of predictive variables to enhance the models' predictive capabilities.

Keywords: diabetes, machine learning, prediction, biomarkers

Procedia PDF Downloads 49
1632 Excitation Density and Energy Dependent Relaxation Dynamics of Charge Carriers in Large Area 2D TMDCs

Authors: Ashish Soni, Suman Kalyan Pal

Abstract:

Transition metal dichalcogenides (TMDCs) are an emerging paradigm for the generation of advanced materials which are capable of utilizing in future device applications. In recent years TMDCs have attracted researchers for their unique band structure in monolayers. Large-area monolayers could become the most appropriate candidate for flexible and thin optoelectronic devices. For this purpose, it is crucial to understand the generation and transport of charge carriers in low dimensions. A deep understanding of photo-generated hot charges and trapped charges is essential to improve the performance of optoelectronic devices. Carrier trapping by the defect states that are introduced during the growth process of the monolayer could influence the dynamical behaviour of charge carriers. Herein, we investigated some aspects of the ultrafast evolution of the initially generated hot carriers and trapped charges in large-area monolayer WS₂ by measuring transient absorption at energies above and below the band gap energy. Our excitation density and energy-dependent measurements reveal the trapping of the initially generated charge carrier. Our results could be beneficial for the development of TMDC-based optoelectronic devices.

Keywords: transient absorption, optoelectronics, 2D materials, TMDCs, exciton

Procedia PDF Downloads 63
1631 The Prediction of Evolutionary Process of Coloured Vision in Mammals: A System Biology Approach

Authors: Shivani Sharma, Prashant Saxena, Inamul Hasan Madar

Abstract:

Since the time of Darwin, it has been considered that genetic change is the direct indicator of variation in phenotype. But a few studies in system biology in the past years have proposed that epigenetic developmental processes also affect the phenotype thus shifting the focus from a linear genotype-phenotype map to a non-linear G-P map. In this paper, we attempt at explaining the evolution of colour vision in mammals by taking LWS/ Long-wave sensitive gene under consideration.

Keywords: evolution, phenotypes, epigenetics, LWS gene, G-P map

Procedia PDF Downloads 516
1630 Studying the Possibility to Weld AA1100 Aluminum Alloy by Friction Stir Spot Welding

Authors: Ahmad K. Jassim, Raheem Kh. Al-Subar

Abstract:

Friction stir welding is a modern and an environmentally friendly solid state joining process used to joint relatively lighter family of materials. Recently, friction stir spot welding has been used instead of resistance spot welding which has received considerable attention from the automotive industry. It is environmentally friendly process that eliminated heat and pollution. In this research, friction stir spot welding has been used to study the possibility to weld AA1100 aluminum alloy sheet with 3 mm thickness by overlapping the edges of sheet as lap joint. The process was done using a drilling machine instead of milling machine. Different tool rotational speeds of 760, 1065, 1445, and 2000 RPM have been applied with manual and automatic compression to study their effect on the quality of welded joints. Heat generation, pressure applied, and depth of tool penetration have been measured during the welding process. The result shows that there is a possibility to weld AA1100 sheets; however, there is some surface defect that happened due to insufficient condition of welding. Moreover, the relationship between rotational speed, pressure, heat generation and tool depth penetration was created.

Keywords: friction, spot, stir, environmental, sustainable, AA1100 aluminum alloy

Procedia PDF Downloads 192
1629 Optimization of Pressure in Deep Drawing Process

Authors: Ajay Kumar Choubey, Geeta Agnihotri, C. Sasikumar, Rashmi Dwivedi

Abstract:

Deep-drawing operations are performed widely in industrial applications. It is very important for efficiency to achieve parts with no or minimum defects. Deep drawn parts are used in high performance, high strength and high reliability applications where tension, stress, load and human safety are critical considerations. Wrinkling is a kind of defect caused by stresses in the flange part of the blank during metal forming operations. To avoid wrinkling appropriate blank-holder pressure/force or drawbead can be applied. Now-a-day computer simulation plays a vital role in the field of manufacturing process. So computer simulation of manufacturing has much advantage over previous conventional process i.e. mass production, good quality of product, fast working etc. In this study, a two dimensional elasto-plastic Finite Element (F.E.) model for Mild Steel material blank has been developed to study the behavior of the flange wrinkling and deep drawing parameters under different Blank-Holder Pressure (B.H.P.). For this, commercially available Finite Element software ANSYS 14 has been used in this study. Simulation results are critically studied and salient conclusions have been drawn.

Keywords: ANSYS, deep drawing, BHP, finite element simulation, wrinkling

Procedia PDF Downloads 446
1628 Analysis of Replication Protein A (RPA): The Role of Homolog Interaction and Recombination during Meiosis

Authors: Jeong Hwan Joo, Keun Pil Kim

Abstract:

During meiosis, meiotic recombination is initiated by Spo11-mediated DSB formation and exonuclease-mediated DSB resection occurs to expose single stranded DNA formation. RPA is further required to inhibit secondary structure formation of ssDNA that can be formed Watson-Crick pairing. Rad51-Dmc1, RecA homologs in eukaryote and their accessory factors involve in searching homolog templates to mediate strand exchange. In this study, we investigate the recombinational roles of replication protein A (RPA), which is heterotrimeric protein that is composed of RPA1, RPA2, and RPA3. Here, we investigated meiotic recombination using DNA physical analysis at the HIS4LEU2 hot spot. In rfa1-119 (K45E, N316S) cells, crossover (CO) and non-crossover (NCO) products reduced than WT. rfa1-119 delayed in single end invasion-to-double holiday junction (SEI-to-dHJ) transition and exhibits a defect in second-end capture that is also modulated by Rad52. In the further experiment, we observed that in rfa1-119 mutant, RPA could not be released in timely manner. Furthermore, rfa1-119 exhibits failure in the second end capture, implying reduction of COs and NCOs. In this talk, we will discuss more detail how RPA involves in chromatin axis association via formation of axis-bridge and why RPA is required for Rad52-mediated second-end capture progression.

Keywords: homolog interaction, meiotic recombination, replication protein A, RPA1

Procedia PDF Downloads 198
1627 Applying Semi-Automatic Digital Aerial Survey Technology and Canopy Characters Classification for Surface Vegetation Interpretation of Archaeological Sites

Authors: Yung-Chung Chuang

Abstract:

The cultural layers of archaeological sites are mainly affected by surface land use, land cover, and root system of surface vegetation. For this reason, continuous monitoring of land use and land cover change is important for archaeological sites protection and management. However, in actual operation, on-site investigation and orthogonal photograph interpretation require a lot of time and manpower. For this reason, it is necessary to perform a good alternative for surface vegetation survey in an automated or semi-automated manner. In this study, we applied semi-automatic digital aerial survey technology and canopy characters classification with very high-resolution aerial photographs for surface vegetation interpretation of archaeological sites. The main idea is based on different landscape or forest type can easily be distinguished with canopy characters (e.g., specific texture distribution, shadow effects and gap characters) extracted by semi-automatic image classification. A novel methodology to classify the shape of canopy characters using landscape indices and multivariate statistics was also proposed. Non-hierarchical cluster analysis was used to assess the optimal number of canopy character clusters and canonical discriminant analysis was used to generate the discriminant functions for canopy character classification (seven categories). Therefore, people could easily predict the forest type and vegetation land cover by corresponding to the specific canopy character category. The results showed that the semi-automatic classification could effectively extract the canopy characters of forest and vegetation land cover. As for forest type and vegetation type prediction, the average prediction accuracy reached 80.3%~91.7% with different sizes of test frame. It represented this technology is useful for archaeological site survey, and can improve the classification efficiency and data update rate.

Keywords: digital aerial survey, canopy characters classification, archaeological sites, multivariate statistics

Procedia PDF Downloads 137
1626 Chaotic Analysis of Acid Rains with Times Series of pH Degree, Nitrate and Sulphate Concentration on Wet Samples

Authors: Aysegul Sener, Gonca Tuncel Memis, Mirac Kamislioglu

Abstract:

Chaos theory is one of the new paradigms of science since the last century. After determining chaos in the weather systems by Edward Lorenz the popularity of the theory was increased. Chaos is observed in many natural systems and studies continue to defect chaos to other natural systems. Acid rain is one of the environmental problems that have negative effects on environment and acid rains values are monitored continuously. In this study, we aim that analyze the chaotic behavior of acid rains in Turkey with the chaotic defecting approaches. The data of pH degree of rain waters, concentration of sulfate and nitrate data of wet rain water samples in the rain collecting stations which are located in different regions of Turkey are provided by Turkish State Meteorology Service. Lyapunov exponents, reconstruction of the phase space, power spectrums are used in this study to determine and predict the chaotic behaviors of acid rains. As a result of the analysis it is found that acid rain time series have positive Lyapunov exponents and wide power spectrums and chaotic behavior is observed in the acid rain time series.

Keywords: acid rains, chaos, chaotic analysis, Lypapunov exponents

Procedia PDF Downloads 141
1625 Developing an Empirical Relationship to Predict Tensile Strength and Micro Hardness of Friction Stir Welded Aluminium Alloy Joints

Authors: Gurmeet Singh Cheema, Gurjinder Singh, Amardeep Singh Kang

Abstract:

Aluminium alloy 6061 is a medium to high strength heat-treatable alloy which has very good corrosion resistance and very good weldability. Friction Stir Welding was developed and this technique has attracted considerable interest from the aerospace and automotive industries since it is able to produce defect free joints particularly for light metals i.e aluminum alloy and magnesium alloy. In the friction stir welding process, welding parameters such as tool rotational speed, welding speed and tool shoulder diameter play a major role in deciding the weld quality. In this research work, an attempt has been made to understand the effect of tool rotational speed, welding speed and tool shoulder diameter on friction stir welded AA6061 aluminium alloy joints. Statistical tool such as central composite design is used to develop the mathematical relationships. The mathematical model was developed to predict mechanical properties of friction stir welded aluminium alloy joints at the 95% confidence level.

Keywords: aluminium alloy, friction stir welding, central composite design, mathematical relationship

Procedia PDF Downloads 496
1624 Effect of Low Level Laser on Healing of Congenital Septal Defects on Dogs

Authors: Hady Atef, Zinab Helmy, Heba Abdeen, Mostafa Fadel

Abstract:

Background and purpose: After the success of the first trials of this experiment which were done on rabbits, a new study were conducted on dogs to ensure the past results; in a step forward to use low-level LASER therapy in the treatment of congenital septal defects in infants. The aim of this study was to investigate the effect of low-level LASER irradiation on congenital septal defects in dogs. Subjects and Methodology: six male dogs who have congenital septal defects in their hearts -with age ranged 6-10 months- enrolled in this study for one and half months. They were assigned into two groups: Group (A): The study group consisted of 3 canine hearts who received routine animal care associated with LASER irradiation. Group (B): The control group consisted of 3 canine hearts who received only routine animal care. Sizes of the septal defects were measured for both groups at the beginning and after the end of the study. Results: There was a significant decrease in the size of the diameter of the congenital septal defect with the study group (percentage of improvement was 42.19%) when compared with control group. Conclusion: It was concluded that low-level LASER therapy can be considered as a promising therapy for congenital heart defects in animals and to be examined on children with similar congenital lesions after then.

Keywords: laser, congenital septal defects, dogs, infants

Procedia PDF Downloads 273
1623 A Generalized Weighted Loss for Support Vextor Classification and Multilayer Perceptron

Authors: Filippo Portera

Abstract:

Usually standard algorithms employ a loss where each error is the mere absolute difference between the true value and the prediction, in case of a regression task. In the present, we present several error weighting schemes that are a generalization of the consolidated routine. We study both a binary classification model for Support Vextor Classification and a regression net for Multylayer Perceptron. Results proves that the error is never worse than the standard procedure and several times it is better.

Keywords: loss, binary-classification, MLP, weights, regression

Procedia PDF Downloads 87
1622 Evaluation of Lactobacillus helveticus as an Adjunct Culture for Removal of Bitterness in Iranian White-Brined Cheese

Authors: F. Nejati, Sh. Dokhani

Abstract:

Bitterness is a flavor defect encountered in some cheeses, such as Iranian white brined cheese and is responsible for reducing acceptability of the cheeses. The objective of this study was to investigate the effect of an adjunct culture on removal of bitterness fro, Iranian white-brined cheese. The chemical and proteolysis characteristics of the cheese were also monitored. Bitter cheeses were made using overdose of clotting enzyme with and without L. helveticus CH-1 as an adjunct culture. Cheese made with normal doses of clotting enzyme was used as the control. Adjunct culture was applied in two different forms: attenuated and non-attenuated. Proteolysis was assessed by measuring the amount of water soluble nitrogen, 12% trichloroacetic acid soluble nitrogen and total free amino acids during ripening. A taste panel group also evaluated the cheeses at the end of ripening period. Results of the statistical analysis showed that the adjunct caused considerable proteolysis and the level of water soluble nitrogen and 12% soluble nitrogen fractions were found to be significantly higher in the treatment involving L. helveticus (respectively P < 0.05 and P < 0.01). Regarding to organoleptic evaluations, the non-shocked adjunct culture caused reduction in bitterness and enhancement of flavor in cheese.

Keywords: bitterness, Iranian white brined cheese, Lactobacillus helveticus, ripening

Procedia PDF Downloads 365
1621 Lactobacillus Helveticus as an Adjunct Culture for Removal of Bitterness in White-Brined Cheese

Authors: Fatemeh Nejati, Shahram Dokhani

Abstract:

Bitterness is a flavor defect encountered in some cheeses, such as Iranian white brined cheese and is responsible for reducing acceptability of the cheeses. The objective of this study was to investigate the effect of an adjunct culture on removal of bitterness fro, Iranian white-brined cheese. The chemical and proteolysis characteristics of the cheese were also monitored. Bitter cheeses were made using overdose of clotting enzyme with and without L. helveticus CH-1 as an adjunct culture. Cheese made with normal doses of clotting enzyme was used as the control. Adjunct culture was applied in two different forms: attenuated and non-attenuated. Proteolysis was assessed by measuring the amount of water soluble nitrogen, 12% trichloroacetic acid soluble nitrogen and total free amino acids during ripening. A taste panel group also evaluated the cheeses at the end of ripening period. Results of the statistical analysis showed that the adjunct caused considerable proteolysis and the level of water soluble nitrogen and 12% soluble nitrogen fractions were found to be significantly higher in the treatment involving L. helveticus (respectively P < 0.05 and P < 0.01). Regarding to organoleptic evaluations, the non-shocked adjunct culture caused reduction in bitterness and enhancement of flavor in cheese.

Keywords: Bitterness, Iranian white brined Cheese, Lactobacillus helveticus, Ripening

Procedia PDF Downloads 454
1620 Analysis of Biomarkers Intractable Epileptogenic Brain Networks with Independent Component Analysis and Deep Learning Algorithms: A Comprehensive Framework for Scalable Seizure Prediction with Unimodal Neuroimaging Data in Pediatric Patients

Authors: Bliss Singhal

Abstract:

Epilepsy is a prevalent neurological disorder affecting approximately 50 million individuals worldwide and 1.2 million Americans. There exist millions of pediatric patients with intractable epilepsy, a condition in which seizures fail to come under control. The occurrence of seizures can result in physical injury, disorientation, unconsciousness, and additional symptoms that could impede children's ability to participate in everyday tasks. Predicting seizures can help parents and healthcare providers take precautions, prevent risky situations, and mentally prepare children to minimize anxiety and nervousness associated with the uncertainty of a seizure. This research proposes a comprehensive framework to predict seizures in pediatric patients by evaluating machine learning algorithms on unimodal neuroimaging data consisting of electroencephalogram signals. The bandpass filtering and independent component analysis proved to be effective in reducing the noise and artifacts from the dataset. Various machine learning algorithms’ performance is evaluated on important metrics such as accuracy, precision, specificity, sensitivity, F1 score and MCC. The results show that the deep learning algorithms are more successful in predicting seizures than logistic Regression, and k nearest neighbors. The recurrent neural network (RNN) gave the highest precision and F1 Score, long short-term memory (LSTM) outperformed RNN in accuracy and convolutional neural network (CNN) resulted in the highest Specificity. This research has significant implications for healthcare providers in proactively managing seizure occurrence in pediatric patients, potentially transforming clinical practices, and improving pediatric care.

Keywords: intractable epilepsy, seizure, deep learning, prediction, electroencephalogram channels

Procedia PDF Downloads 77
1619 Gradient Boosted Trees on Spark Platform for Supervised Learning in Health Care Big Data

Authors: Gayathri Nagarajan, L. D. Dhinesh Babu

Abstract:

Health care is one of the prominent industries that generate voluminous data thereby finding the need of machine learning techniques with big data solutions for efficient processing and prediction. Missing data, incomplete data, real time streaming data, sensitive data, privacy, heterogeneity are few of the common challenges to be addressed for efficient processing and mining of health care data. In comparison with other applications, accuracy and fast processing are of higher importance for health care applications as they are related to the human life directly. Though there are many machine learning techniques and big data solutions used for efficient processing and prediction in health care data, different techniques and different frameworks are proved to be effective for different applications largely depending on the characteristics of the datasets. In this paper, we present a framework that uses ensemble machine learning technique gradient boosted trees for data classification in health care big data. The framework is built on Spark platform which is fast in comparison with other traditional frameworks. Unlike other works that focus on a single technique, our work presents a comparison of six different machine learning techniques along with gradient boosted trees on datasets of different characteristics. Five benchmark health care datasets are considered for experimentation, and the results of different machine learning techniques are discussed in comparison with gradient boosted trees. The metric chosen for comparison is misclassification error rate and the run time of the algorithms. The goal of this paper is to i) Compare the performance of gradient boosted trees with other machine learning techniques in Spark platform specifically for health care big data and ii) Discuss the results from the experiments conducted on datasets of different characteristics thereby drawing inference and conclusion. The experimental results show that the accuracy is largely dependent on the characteristics of the datasets for other machine learning techniques whereas gradient boosting trees yields reasonably stable results in terms of accuracy without largely depending on the dataset characteristics.

Keywords: big data analytics, ensemble machine learning, gradient boosted trees, Spark platform

Procedia PDF Downloads 233
1618 Validation of Asymptotic Techniques to Predict Bistatic Radar Cross Section

Authors: M. Pienaar, J. W. Odendaal, J. C. Smit, J. Joubert

Abstract:

Simulations are commonly used to predict the bistatic radar cross section (RCS) of military targets since characterization measurements can be expensive and time consuming. It is thus important to accurately predict the bistatic RCS of targets. Computational electromagnetic (CEM) methods can be used for bistatic RCS prediction. CEM methods are divided into full-wave and asymptotic methods. Full-wave methods are numerical approximations to the exact solution of Maxwell’s equations. These methods are very accurate but are computationally very intensive and time consuming. Asymptotic techniques make simplifying assumptions in solving Maxwell's equations and are thus less accurate but require less computational resources and time. Asymptotic techniques can thus be very valuable for the prediction of bistatic RCS of electrically large targets, due to the decreased computational requirements. This study extends previous work by validating the accuracy of asymptotic techniques to predict bistatic RCS through comparison with full-wave simulations as well as measurements. Validation is done with canonical structures as well as complex realistic aircraft models instead of only looking at a complex slicy structure. The slicy structure is a combination of canonical structures, including cylinders, corner reflectors and cubes. Validation is done over large bistatic angles and at different polarizations. Bistatic RCS measurements were conducted in a compact range, at the University of Pretoria, South Africa. The measurements were performed at different polarizations from 2 GHz to 6 GHz. Fixed bistatic angles of β = 30.8°, 45° and 90° were used. The measurements were calibrated with an active calibration target. The EM simulation tool FEKO was used to generate simulated results. The full-wave multi-level fast multipole method (MLFMM) simulated results together with the measured data were used as reference for validation. The accuracy of physical optics (PO) and geometrical optics (GO) was investigated. Differences relating to amplitude, lobing structure and null positions were observed between the asymptotic, full-wave and measured data. PO and GO were more accurate at angles close to the specular scattering directions and the accuracy seemed to decrease as the bistatic angle increased. At large bistatic angles PO did not perform well due to the shadow regions not being treated appropriately. PO also did not perform well for canonical structures where multi-bounce was the main scattering mechanism. PO and GO do not account for diffraction but these inaccuracies tended to decrease as the electrical size of objects increased. It was evident that both asymptotic techniques do not properly account for bistatic structural shadowing. Specular scattering was calculated accurately even if targets did not meet the electrically large criteria. It was evident that the bistatic RCS prediction performance of PO and GO depends on incident angle, frequency, target shape and observation angle. The improved computational efficiency of the asymptotic solvers yields a major advantage over full-wave solvers and measurements; however, there is still much room for improvement of the accuracy of these asymptotic techniques.

Keywords: asymptotic techniques, bistatic RCS, geometrical optics, physical optics

Procedia PDF Downloads 251
1617 The Osteocutaneous Distal Tibia Turn-over Fillet Flap: A Novel Spare-parts Orthoplastic Surgery Option for Functional Below-knee Amputation

Authors: Harry Burton, Alexios Dimitrios Iliadis, Neil Jones, Aaron Saini, Nicola Bystrzonowski, Alexandros Vris, Georgios Pafitanis

Abstract:

This article portrays the authors’ experience with a complex lower limb bone and soft tissue defect, following chronic osteomyelitis and pathological fracture, which was managed by the multidisciplinary orthoplastic team. The decision for functional amputation versus limb salvage was deemed necessary, enhanced by the principles of “spares parts” in reconstructive microsurgery. This case describes a successful use of the osteocutaneous distal tibia turn-over fillet flap that allowed ‘lowering the level of the amputation’ from a through knee to the conventional level of a below-knee amputation to preserve the knee joint function. This case demonstrates the value of ‘spare-parts’ surgery principles and how these concepts refine complex orthoplastic approaches when limb salvage is not possible to enhance function. The osteocutaneous distal tibia turn-over fillet flap is a robust technique for modified BKA reconstructions that provides sufficient bone length to achieve a tough, sensate stump and functional knee joint.

Keywords: osteocutaneous flap, fillet flap, spare-parts surgery, Below knee amputation

Procedia PDF Downloads 158
1616 Field Prognostic Factors on Discharge Prediction of Traumatic Brain Injuries

Authors: Mohammad Javad Behzadnia, Amir Bahador Boroumand

Abstract:

Introduction: Limited facility situations require allocating the most available resources for most casualties. Accordingly, Traumatic Brain Injury (TBI) is the one that may need to transport the patient as soon as possible. In a mass casualty event, deciding when the facilities are restricted is hard. The Extended Glasgow Outcome Score (GOSE) has been introduced to assess the global outcome after brain injuries. Therefore, we aimed to evaluate the prognostic factors associated with GOSE. Materials and Methods: In a multicenter cross-sectional study conducted on 144 patients with TBI admitted to trauma emergency centers. All the patients with isolated TBI who were mentally and physically healthy before the trauma entered the study. The patient’s information was evaluated, including demographic characteristics, duration of hospital stays, mechanical ventilation on admission laboratory measurements, and on-admission vital signs. We recorded the patients’ TBI-related symptoms and brain computed tomography (CT) scan findings. Results: GOSE assessments showed an increasing trend by the comparison of on-discharge (7.47 ± 1.30), within a month (7.51 ± 1.30), and within three months (7.58 ± 1.21) evaluations (P < 0.001). On discharge, GOSE was positively correlated with Glasgow Coma Scale (GCS) (r = 0.729, P < 0.001) and motor GCS (r = 0.812, P < 0.001), and inversely with age (r = −0.261, P = 0.002), hospitalization period (r = −0.678, P < 0.001), pulse rate (r = −0.256, P = 0.002) and white blood cell (WBC). Among imaging signs and trauma-related symptoms in univariate analysis, intracranial hemorrhage (ICH), interventricular hemorrhage (IVH) (P = 0.006), subarachnoid hemorrhage (SAH) (P = 0.06; marginally at P < 0.1), subdural hemorrhage (SDH) (P = 0.032), and epidural hemorrhage (EDH) (P = 0.037) were significantly associated with GOSE at discharge in multivariable analysis. Conclusion: Our study showed some predictive factors that could help to decide which casualty should transport earlier to a trauma center. According to the current study findings, GCS, pulse rate, WBC, and among imaging signs and trauma-related symptoms, ICH, IVH, SAH, SDH, and EDH are significant independent predictors of GOSE at discharge in TBI patients.

Keywords: field, Glasgow outcome score, prediction, traumatic brain injury.

Procedia PDF Downloads 69
1615 Multifunctional 1D α-Fe2O3/ZnO Core/Shell Semiconductor Nano-Heterostructures: Heterojunction Engineering

Authors: Gobinda Gopal Khan, Ashutosh K. Singh, Debasish Sarkar

Abstract:

This study reports the facile fabrication of 1D ZnO/α-Fe2O3 semiconductor nano-heterostructures (SNHs), and we investigate the strong interfacial interactions at the heterojunction, resulting in novel multifunctionality in the hybrid structure. ZnO-coated α-Fe2O3 nanowires (NWs) have been prepared by combining electrodeposition and wet chemical methods. Significant improvement in electrical conductivity, photoluminescence, and room temperature magnetic properties have been observed for the ZnO/α-Fe2O3 SNHs over the pristine α-Fe2O3 NWs because of the contribution of the ZnO nanolayer. The increase in electrical conductivity in ZnO/α-Fe2O3 SNHs is because of the increase in free electrons in the conduction band of the SNHs due to the formation of type-II n-n band configuration at the heterojunction. The SNHs are found to exhibit enhanced visible green photoluminescence along with the UV emission at room temperature. The band-gap emission of the α-Fe2O3 NWs coupled to the defect emissions of the ZnO in SNHs can be attributed to the profound enhancement of the visible green luminescence. Ferromagnetism of the SNHs is found to be increased nearly five times in magnitude over the primeval α-Fe2O3 NWs, which can be ascribed to the exchange coupling of the interfacial spin at ZnO/α-Fe2O3 interface, the surface spin of ZnO nanolayer, along with the structural defects like the cation vacancies (VZn) and the singly ionized oxygen vacancies (Vo•) present in SNHs.

Keywords: nano-heterostructures, photoluminescence, electrical property, magnetism

Procedia PDF Downloads 252
1614 Estimation of Fragility Curves Using Proposed Ground Motion Selection and Scaling Procedure

Authors: Esra Zengin, Sinan Akkar

Abstract:

Reliable and accurate prediction of nonlinear structural response requires specification of appropriate earthquake ground motions to be used in nonlinear time history analysis. The current research has mainly focused on selection and manipulation of real earthquake records that can be seen as the most critical step in the performance based seismic design and assessment of the structures. Utilizing amplitude scaled ground motions that matches with the target spectra is commonly used technique for the estimation of nonlinear structural response. Representative ground motion ensembles are selected to match target spectrum such as scenario-based spectrum derived from ground motion prediction equations, Uniform Hazard Spectrum (UHS), Conditional Mean Spectrum (CMS) or Conditional Spectrum (CS). Different sets of criteria exist among those developed methodologies to select and scale ground motions with the objective of obtaining robust estimation of the structural performance. This study presents ground motion selection and scaling procedure that considers the spectral variability at target demand with the level of ground motion dispersion. The proposed methodology provides a set of ground motions whose response spectra match target median and corresponding variance within a specified period interval. The efficient and simple algorithm is used to assemble the ground motion sets. The scaling stage is based on the minimization of the error between scaled median and the target spectra where the dispersion of the earthquake shaking is preserved along the period interval. The impact of the spectral variability on nonlinear response distribution is investigated at the level of inelastic single degree of freedom systems. In order to see the effect of different selection and scaling methodologies on fragility curve estimations, results are compared with those obtained by CMS-based scaling methodology. The variability in fragility curves due to the consideration of dispersion in ground motion selection process is also examined.

Keywords: ground motion selection, scaling, uncertainty, fragility curve

Procedia PDF Downloads 581
1613 Development and Characterization of Hydroxyapatite Based Nanocomposites for Local Drug Delivery to Periodontal Pockets

Authors: Indu Lata Kanwar, Preeti K. Suresh

Abstract:

The aim of this study is to fabricate hydroxyapatite based nanocomposites for local drug delivery in periodontal pockets. Hydroxyapatite is chemically similar to the mineral component of bones and hard tissues in mammals. Synthetic biocompatibility and bioactivity with human teeth and bone, making it very attractive for biomedical applications. Nanocomposite is a multiphase solid material where one of the phases has one, two or three dimensions of less than 100 nanometres (nm), or structures having nano­scale repeat distances between the different phases that make up the material. Nanostructured calcium phosphate materials play an important role in the formation of hard tissues in nature. It is reported that calcium phosphates materials in nano-size can mimic the dimensions of constituent components of calcified tissues. Nano-sized materials offer improved performances compared with conventional materials due to their large surface-to-volume ratios. The specific biological properties of the nanocomposites, as well as their interaction with cells, include the use of bioactive molecules. The approach of periodontal tissue engineering is considered promising to restore bone defect through the use of engineered materials with the aim that they will prohibit the invasion of fibrous connective tissue and help repair the function during bone regeneration.

Keywords: bioactive, hydroxyapatite, nanocomposities, periondontal

Procedia PDF Downloads 322
1612 A Long Short-Term Memory Based Deep Learning Model for Corporate Bond Price Predictions

Authors: Vikrant Gupta, Amrit Goswami

Abstract:

The fixed income market forms the basis of the modern financial market. All other assets in financial markets derive their value from the bond market. Owing to its over-the-counter nature, corporate bonds have relatively less data publicly available and thus is researched upon far less compared to Equities. Bond price prediction is a complex financial time series forecasting problem and is considered very crucial in the domain of finance. The bond prices are highly volatile and full of noise which makes it very difficult for traditional statistical time-series models to capture the complexity in series patterns which leads to inefficient forecasts. To overcome the inefficiencies of statistical models, various machine learning techniques were initially used in the literature for more accurate forecasting of time-series. However, simple machine learning methods such as linear regression, support vectors, random forests fail to provide efficient results when tested on highly complex sequences such as stock prices and bond prices. hence to capture these intricate sequence patterns, various deep learning-based methodologies have been discussed in the literature. In this study, a recurrent neural network-based deep learning model using long short term networks for prediction of corporate bond prices has been discussed. Long Short Term networks (LSTM) have been widely used in the literature for various sequence learning tasks in various domains such as machine translation, speech recognition, etc. In recent years, various studies have discussed the effectiveness of LSTMs in forecasting complex time-series sequences and have shown promising results when compared to other methodologies. LSTMs are a special kind of recurrent neural networks which are capable of learning long term dependencies due to its memory function which traditional neural networks fail to capture. In this study, a simple LSTM, Stacked LSTM and a Masked LSTM based model has been discussed with respect to varying input sequences (three days, seven days and 14 days). In order to facilitate faster learning and to gradually decompose the complexity of bond price sequence, an Empirical Mode Decomposition (EMD) has been used, which has resulted in accuracy improvement of the standalone LSTM model. With a variety of Technical Indicators and EMD decomposed time series, Masked LSTM outperformed the other two counterparts in terms of prediction accuracy. To benchmark the proposed model, the results have been compared with traditional time series models (ARIMA), shallow neural networks and above discussed three different LSTM models. In summary, our results show that the use of LSTM models provide more accurate results and should be explored more within the asset management industry.

Keywords: bond prices, long short-term memory, time series forecasting, empirical mode decomposition

Procedia PDF Downloads 133
1611 Measuring Enterprise Growth: Pitfalls and Implications

Authors: N. Šarlija, S. Pfeifer, M. Jeger, A. Bilandžić

Abstract:

Enterprise growth is generally considered as a key driver of competitiveness, employment, economic development and social inclusion. As such, it is perceived to be a highly desirable outcome of entrepreneurship for scholars and decision makers. The huge academic debate resulted in the multitude of theoretical frameworks focused on explaining growth stages, determinants and future prospects. It has been widely accepted that enterprise growth is most likely nonlinear, temporal and related to the variety of factors which reflect the individual, firm, organizational, industry or environmental determinants of growth. However, factors that affect growth are not easily captured, instruments to measure those factors are often arbitrary, causality between variables and growth is elusive, indicating that growth is not easily modeled. Furthermore, in line with heterogeneous nature of the growth phenomenon, there is a vast number of measurement constructs assessing growth which are used interchangeably. Differences among various growth measures, at conceptual as well as at operationalization level, can hinder theory development which emphasizes the need for more empirically robust studies. In line with these highlights, the main purpose of this paper is twofold. Firstly, to compare structure and performance of three growth prediction models based on the main growth measures: Revenues, employment and assets growth. Secondly, to explore the prospects of financial indicators, set as exact, visible, standardized and accessible variables, to serve as determinants of enterprise growth. Finally, to contribute to the understanding of the implications on research results and recommendations for growth caused by different growth measures. The models include a range of financial indicators as lag determinants of the enterprises’ performances during the 2008-2013, extracted from the national register of the financial statements of SMEs in Croatia. The design and testing stage of the modeling used the logistic regression procedures. Findings confirm that growth prediction models based on different measures of growth have different set of predictors. Moreover, the relationship between particular predictors and growth measure is inconsistent, namely the same predictor positively related to one growth measure may exert negative effect on a different growth measure. Overall, financial indicators alone can serve as good proxy of growth and yield adequate predictive power of the models. The paper sheds light on both methodology and conceptual framework of enterprise growth by using a range of variables which serve as a proxy for the multitude of internal and external determinants, but are unlike them, accessible, available, exact and free of perceptual nuances in building up the model. Selection of the growth measure seems to have significant impact on the implications and recommendations related to growth. Furthermore, the paper points out to potential pitfalls of measuring and predicting growth. Overall, the results and the implications of the study are relevant for advancing academic debates on growth-related methodology, and can contribute to evidence-based decisions of policy makers.

Keywords: growth measurement constructs, logistic regression, prediction of growth potential, small and medium-sized enterprises

Procedia PDF Downloads 247
1610 Lineup Optimization Model of Basketball Players Based on the Prediction of Recursive Neural Networks

Authors: Wang Yichen, Haruka Yamashita

Abstract:

In recent years, in the field of sports, decision making such as member in the game and strategy of the game based on then analysis of the accumulated sports data are widely attempted. In fact, in the NBA basketball league where the world's highest level players gather, to win the games, teams analyze the data using various statistical techniques. However, it is difficult to analyze the game data for each play such as the ball tracking or motion of the players in the game, because the situation of the game changes rapidly, and the structure of the data should be complicated. Therefore, it is considered that the analysis method for real time game play data is proposed. In this research, we propose an analytical model for "determining the optimal lineup composition" using the real time play data, which is considered to be difficult for all coaches. In this study, because replacing the entire lineup is too complicated, and the actual question for the replacement of players is "whether or not the lineup should be changed", and “whether or not Small Ball lineup is adopted”. Therefore, we propose an analytical model for the optimal player selection problem based on Small Ball lineups. In basketball, we can accumulate scoring data for each play, which indicates a player's contribution to the game, and the scoring data can be considered as a time series data. In order to compare the importance of players in different situations and lineups, we combine RNN (Recurrent Neural Network) model, which can analyze time series data, and NN (Neural Network) model, which can analyze the situation on the field, to build the prediction model of score. This model is capable to identify the current optimal lineup for different situations. In this research, we collected all the data of accumulated data of NBA from 2019-2020. Then we apply the method to the actual basketball play data to verify the reliability of the proposed model.

Keywords: recurrent neural network, players lineup, basketball data, decision making model

Procedia PDF Downloads 126
1609 Comparing Performance of Neural Network and Decision Tree in Prediction of Myocardial Infarction

Authors: Reza Safdari, Goli Arji, Robab Abdolkhani Maryam zahmatkeshan

Abstract:

Background and purpose: Cardiovascular diseases are among the most common diseases in all societies. The most important step in minimizing myocardial infarction and its complications is to minimize its risk factors. The amount of medical data is increasingly growing. Medical data mining has a great potential for transforming these data into information. Using data mining techniques to generate predictive models for identifying those at risk for reducing the effects of the disease is very helpful. The present study aimed to collect data related to risk factors of heart infarction from patients’ medical record and developed predicting models using data mining algorithm. Methods: The present work was an analytical study conducted on a database containing 350 records. Data were related to patients admitted to Shahid Rajaei specialized cardiovascular hospital, Iran, in 2011. Data were collected using a four-sectioned data collection form. Data analysis was performed using SPSS and Clementine version 12. Seven predictive algorithms and one algorithm-based model for predicting association rules were applied to the data. Accuracy, precision, sensitivity, specificity, as well as positive and negative predictive values were determined and the final model was obtained. Results: five parameters, including hypertension, DLP, tobacco smoking, diabetes, and A+ blood group, were the most critical risk factors of myocardial infarction. Among the models, the neural network model was found to have the highest sensitivity, indicating its ability to successfully diagnose the disease. Conclusion: Risk prediction models have great potentials in facilitating the management of a patient with a specific disease. Therefore, health interventions or change in their life style can be conducted based on these models for improving the health conditions of the individuals at risk.

Keywords: decision trees, neural network, myocardial infarction, Data Mining

Procedia PDF Downloads 426