Search results for: prediction error
1299 Prediction of Changes in Optical Quality by Tissue Redness after Pterygium Surgery
Authors: Mohd Radzi Hilmi, Mohd Zulfaezal Che Azemin, Khairidzan Mohd Kamal, Azrin Esmady Ariffin, Mohd Izzuddin Mohd Tamrin, Norfazrina Abdul Gaffur, Tengku Mohd Tengku Sembok
Abstract:
Purpose: The purpose of this study is to predict optical quality changes after pterygium surgery using tissue redness grading. Methods: Sixty-eight primary pterygium participants were selected from patients who visited an ophthalmology clinic. We developed a semi-automated computer program to measure the pterygium fibrovascular redness from digital pterygium images. The outcome of this software is a continuous scale grading of 1 (minimum redness) to 3 (maximum redness). The region of interest (ROI) was selected manually using the software. Reliability was determined by repeat grading of all 68 images and its association with contrast sensitivity function (CSF) and visual acuity (VA) was examined. Results: The mean and standard deviation of redness of the pterygium fibrovascular images was 1.88 ± 0.55. Intra- and inter-grader reliability estimates were high with intraclass correlation ranging from 0.97 to 0.98. The new grading was positively associated with CSF (p<0.01) and VA (p<0.01). The redness grading was able to predict 25% and 23% of the variance in the CSF and the VA respectively. Conclusions: The new grading of pterygium fibrovascular redness can be reliably measured from digital images and show a good correlation with CSF and VA. The redness grading can be used in addition to the existing pterygium grading.Keywords: contrast sensitivity, pterygium, redness, visual acuity
Procedia PDF Downloads 5121298 Application of the Mobile Phone for Occupational Self-Inspection Program in Small-Scale Industries
Authors: Jia-Sin Li, Ying-Fang Wang, Cheing-Tong Yan
Abstract:
In this study, an integrated approach of Google Spreadsheet and QR code which is free internet resources was used to improve the inspection procedure. The mobile phone Application(App)was also designed to combine with a web page to create an automatic checklist in order to provide a new integrated information of inspection management system. By means of client-server model, the client App is developed for Android mobile OS and the back end is a web server. It can set up App accounts including authorized data and store some checklist documents in the website. The checklist document URL could generate QR code first and then print and paste on the machine. The user can scan the QR code by the app and filled the checklist in the factory. In the meanwhile, the checklist data will send to the server, it not only save the filled data but also executes the related functions and charts. On the other hand, it also enables auditors and supervisors to facilitate the prevention and response to hazards, as well as immediate report data checks. Finally, statistics and professional analysis are performed using inspection records and other relevant data to not only improve the reliability, integrity of inspection operations and equipment loss control, but also increase plant safety and personnel performance. Therefore, it suggested that the traditional paper-based inspection method could be replaced by the APP which promotes the promotion of industrial security and reduces human error.Keywords: checklist, Google spreadsheet, APP, self-inspection
Procedia PDF Downloads 1171297 Fear of Negative Evaluation, Social Support and Wellbeing in People with Vitiligo
Authors: Rafia Rafique, Mutmina Zainab
Abstract:
The present study investigated the relationship between fear of negative evaluation (FNE), social support and well-being in people with Vitiligo. It was hypothesized that low level of FNE and greater social support is likely to predict well-being. It was also hypothesized that social support is likely to moderate the relationship between FNE and well-being. Correlational research design was used for the present study. Non-probability purposive sampling technique was used to collect a sample (N=122) of people with Vitiligo. Hierarchical Moderated Regression analysis was used to test prediction and moderation. Brief Fear of Negative Evaluation Scale, Multidimensional Scale of Perceived Social Support (MSPSS) and Mental Health Continuum-Short form (MHC-SF) were used to evaluate the study variables. Fear of negative evaluation negatively predicted well-being (emotional and psychological). Social support from significant others and friends predicted social well-being. Social Support from family predicted emotional and psychological well-being. It was found that social support from significant others moderated the relationship between FNE and emotional well-being and social support from family moderated the relationship between FNE and social well-being. Dermatologists treating people with Vitiligo need to educate them and their families about the buffering role of social support (family and significant others). Future studies need to focus on other important mediating factors that can possibly explain the relationship between fear of negative evaluation and wellbeing.Keywords: fear of negative evaluation, hierarchical moderated regression, vitiligo, well-being
Procedia PDF Downloads 3011296 Meteorological Risk Assessment for Ships with Fuzzy Logic Designer
Authors: Ismail Karaca, Ridvan Saracoglu, Omer Soner
Abstract:
Fuzzy Logic, an advanced method to support decision-making, is used by various scientists in many disciplines. Fuzzy programming is a product of fuzzy logic, fuzzy rules, and implication. In marine science, fuzzy programming for ships is dramatically increasing together with autonomous ship studies. In this paper, a program to support the decision-making process for ship navigation has been designed. The program is produced in fuzzy logic and rules, by taking the marine accidents and expert opinions into account. After the program was designed, the program was tested by 46 ship accidents reported by the Transportation Safety Investigation Center of Turkey. Wind speed, sea condition, visibility, day/night ratio have been used as input data. They have been converted into a risk factor within the Fuzzy Logic Designer application and fuzzy rules set by marine experts. Finally, the expert's meteorological risk factor for each accident is compared with the program's risk factor, and the error rate was calculated. The main objective of this study is to improve the navigational safety of ships, by using the advance decision support model. According to the study result, fuzzy programming is a robust model that supports safe navigation.Keywords: calculation of risk factor, fuzzy logic, fuzzy programming for ship, safety navigation of ships
Procedia PDF Downloads 1881295 Energy Detection Based Sensing and Primary User Traffic Classification for Cognitive Radio
Authors: Urvee B. Trivedi, U. D. Dalal
Abstract:
As wireless communication services grow quickly; the seriousness of spectrum utilization has been on the rise gradually. An emerging technology, cognitive radio has come out to solve today’s spectrum scarcity problem. To support the spectrum reuse functionality, secondary users are required to sense the radio frequency environment, and once the primary users are found to be active, the secondary users are required to vacate the channel within a certain amount of time. Therefore, spectrum sensing is of significant importance. Once sensing is done, different prediction rules apply to classify the traffic pattern of primary user. Primary user follows two types of traffic patterns: periodic and stochastic ON-OFF patterns. A cognitive radio can learn the patterns in different channels over time. Two types of classification methods are discussed in this paper, by considering edge detection and by using autocorrelation function. Edge detection method has a high accuracy but it cannot tolerate sensing errors. Autocorrelation-based classification is applicable in the real environment as it can tolerate some amount of sensing errors.Keywords: cognitive radio (CR), probability of detection (PD), probability of false alarm (PF), primary user (PU), secondary user (SU), fast Fourier transform (FFT), signal to noise ratio (SNR)
Procedia PDF Downloads 3441294 Probability-Based Damage Detection of Structures Using Kriging Surrogates and Enhanced Ideal Gas Molecular Movement Algorithm
Authors: M. R. Ghasemi, R. Ghiasi, H. Varaee
Abstract:
Surrogate model has received increasing attention for use in detecting damage of structures based on vibration modal parameters. However, uncertainties existing in the measured vibration data may lead to false or unreliable output result from such model. In this study, an efficient approach based on Monte Carlo simulation is proposed to take into account the effect of uncertainties in developing a surrogate model. The probability of damage existence (PDE) is calculated based on the probability density function of the existence of undamaged and damaged states. The kriging technique allows one to genuinely quantify the surrogate error, therefore it is chosen as metamodeling technique. Enhanced version of ideal gas molecular movement (EIGMM) algorithm is used as main algorithm for model updating. The developed approach is applied to detect simulated damage in numerical models of 72-bar space truss and 120-bar dome truss. The simulation results show the proposed method can perform well in probability-based damage detection of structures with less computational effort compared to direct finite element model.Keywords: probability-based damage detection (PBDD), Kriging, surrogate modeling, uncertainty quantification, artificial intelligence, enhanced ideal gas molecular movement (EIGMM)
Procedia PDF Downloads 2381293 A Study on Reliability of Gender and Stature Determination by Odontometric and Craniofacial Anthropometric Parameters
Authors: Churamani Pokhrel, C. B. Jha, S. R. Niraula, P. R. Pokharel
Abstract:
Human identification is one of the most challenging subjects that man has confronted. The determination of adult sex and stature are two of the four key factors (sex, stature, age, and race) in identification of an individual. Craniofacial and odontometric parameters are important tools for forensic anthropologists when it is not possible to apply advanced techniques for identification purposes. The present study provides anthropometric correlation of the parameters with stature and gender and also devises regression formulae for reconstruction of stature. A total of 312 Nepalese students with equal distribution of sex i.e., 156 male and 156 female students of age 18-35 years were taken for the study. Total of 10 parameters were measured (age, sex, stature, head circumference, head length, head breadth, facial height, bi-zygomatic width, mesio-distal canine width and inter-canine distance of both maxilla and mandible). Co-relation and regression analysis was done to find the association between the parameters. All parameters were found to be greater in males than females and each was found to be statistically significant. Out of total 312 samples, the best regressor for the determination of stature was head circumference and mandibular inter-canine width and that for gender was head circumference and right mandibular teeth. The accuracy of prediction was 83%. Regression equations and analysis generated from craniofacial and odontometric parameters can be a supplementary approach for the estimation of stature and gender when extremities are not available.Keywords: craniofacial, gender, odontometric, stature
Procedia PDF Downloads 1901292 Medical Image Watermark and Tamper Detection Using Constant Correlation Spread Spectrum Watermarking
Authors: Peter U. Eze, P. Udaya, Robin J. Evans
Abstract:
Data hiding can be achieved by Steganography or invisible digital watermarking. For digital watermarking, both accurate retrieval of the embedded watermark and the integrity of the cover image are important. Medical image security in Teleradiology is one of the applications where the embedded patient record needs to be extracted with accuracy as well as the medical image integrity verified. In this research paper, the Constant Correlation Spread Spectrum digital watermarking for medical image tamper detection and accurate embedded watermark retrieval is introduced. In the proposed method, a watermark bit from a patient record is spread in a medical image sub-block such that the correlation of all watermarked sub-blocks with a spreading code, W, would have a constant value, p. The constant correlation p, spreading code, W and the size of the sub-blocks constitute the secret key. Tamper detection is achieved by flagging any sub-block whose correlation value deviates by more than a small value, ℇ, from p. The major features of our new scheme include: (1) Improving watermark detection accuracy for high-pixel depth medical images by reducing the Bit Error Rate (BER) to Zero and (2) block-level tamper detection in a single computational process with simultaneous watermark detection, thereby increasing utility with the same computational cost.Keywords: Constant Correlation, Medical Image, Spread Spectrum, Tamper Detection, Watermarking
Procedia PDF Downloads 1921291 Health Outcomes and Economic Growth Nexus: Testing for Long-run Relationships and Causal Links in Nigeria
Authors: Haruna Modibbo Usman, Mustapha Muktar, Nasiru Inuwa
Abstract:
This paper examined the long run relationship between health outcomes and economic growth in Nigeria from 1961 to 2012. Using annual time series data, Augmented Dickey-Fuller (ADF) test is conducted to check the stochastic properties of the variables. Also, the long run relationship among the variables is confirmed based on Johansen Multivariate Cointegration approach whereas the long run and short run dynamics are observed using Vector Error Correction Mechanism (VECM). In addition, VEC Granger causality test is employed to examine the direction of causality among the variables. On the whole, the results obtained revealed the existence of a long run relationship between health outcomes and economic growth in Nigeria and that both life expectancy and crude death rate as measures of health are found to have a long run negative and statistically significant impact on the economic growth over the study period. This is further buttressed by the results of Granger causality test which indicated the existence of unidirectional causality running from life expectancy and crude death rate to economic growth. The study therefore, calls for governments at various levels to create preconditions for health improvements in Nigeria in order to boost the level of health outcomes.Keywords: cointegration, economic growth, Granger causality, health outcomes, VECM
Procedia PDF Downloads 4891290 Implementation of Model Reference Adaptive Control in Tuning of Controller Gains for Following-Vehicle System with Fixed Time Headway
Authors: Fatemeh Behbahani, Rubiyah Yusof
Abstract:
To avoid collision between following vehicles and vehicles in front, it is vital to keep appropriate, safe spacing between both vehicles over all speeds. Therefore, the following vehicle needs to have exact information regarding the speed and spacing between vehicles. This project is conducted to simulate the tuning of controller gain for a vehicle-following system through the selected control strategy, spacing control policy and fixed-time headway policy. In addition, the paper simulates and designs an adaptive gain controller for a road-vehicle-following system which uses information on the spacing, velocity and also acceleration of a preceding vehicle in the proposed one-vehicle look-ahead strategy. The mathematical model is implemented using Kirchhoff and Newton’s Laws, and stability simulated. The trial-error method was used to obtain a suitable value of controller gain. However, the adaptive-based controller system was able to optimize the gain value automatically. Model Reference Adaptive Control (MRAC) is designed and utilized and based on firstly the Gradient and secondly the Lyapunov approach. The Lyapunov approach considers stability. The Gradient approach was found to improve the best value of gain in the controller system with fixed-time headway.Keywords: one-vehicle look-ahead, model reference adaptive, stability, tuning gain controller, MRAC
Procedia PDF Downloads 2361289 Predictive Analytics of Student Performance Determinants
Authors: Mahtab Davari, Charles Edward Okon, Somayeh Aghanavesi
Abstract:
Every institute of learning is usually interested in the performance of enrolled students. The level of these performances determines the approach an institute of study may adopt in rendering academic services. The focus of this paper is to evaluate students' academic performance in given courses of study using machine learning methods. This study evaluated various supervised machine learning classification algorithms such as Logistic Regression (LR), Support Vector Machine, Random Forest, Decision Tree, K-Nearest Neighbors, Linear Discriminant Analysis, and Quadratic Discriminant Analysis, using selected features to predict study performance. The accuracy, precision, recall, and F1 score obtained from a 5-Fold Cross-Validation were used to determine the best classification algorithm to predict students’ performances. SVM (using a linear kernel), LDA, and LR were identified as the best-performing machine learning methods. Also, using the LR model, this study identified students' educational habits such as reading and paying attention in class as strong determinants for a student to have an above-average performance. Other important features include the academic history of the student and work. Demographic factors such as age, gender, high school graduation, etc., had no significant effect on a student's performance.Keywords: student performance, supervised machine learning, classification, cross-validation, prediction
Procedia PDF Downloads 1251288 Prediction of Phonon Thermal Conductivity of F.C.C. Al by Molecular Dynamics Simulation
Authors: Leila Momenzadeh, Alexander V. Evteev, Elena V. Levchenko, Tanvir Ahmed, Irina Belova, Graeme Murch
Abstract:
In this work, the phonon thermal conductivity of f.c.c. Al is investigated in detail in the temperature range 100 – 900 K within the framework of equilibrium molecular dynamics simulations making use of the Green-Kubo formalism and one of the most reliable embedded-atom method potentials. It is found that the heat current auto-correlation function of the f.c.c. Al model demonstrates a two-stage temporal decay similar to the previously observed for f.c.c Cu model. After the first stage of decay, the heat current auto-correlation function of the f.c.c. Al model demonstrates a peak in the temperature range 100-800 K. The intensity of the peak decreases as the temperature increases. At 900 K, it transforms to a shoulder. To describe the observed two-stage decay of the heat current auto-correlation function of the f.c.c. Al model, we employ decomposition model recently developed for phonon-mediated thermal transport in a monoatomic lattice. We found that the electronic contribution to the total thermal conductivity of f.c.c. Al dominates over the whole studied temperature range. However, the phonon contribution to the total thermal conductivity of f.c.c. Al increases as temperature decreases. It is about 1.05% at 900 K and about 12.5% at 100 K.Keywords: aluminum, gGreen-Kubo formalism, molecular dynamics, phonon thermal conductivity
Procedia PDF Downloads 4111287 Parametric Analysis and Optimal Design of Functionally Graded Plates Using Particle Swarm Optimization Algorithm and a Hybrid Meshless Method
Authors: Foad Nazari, Seyed Mahmood Hosseini, Mohammad Hossein Abolbashari, Mohammad Hassan Abolbashari
Abstract:
The present study is concerned with the optimal design of functionally graded plates using particle swarm optimization (PSO) algorithm. In this study, meshless local Petrov-Galerkin (MLPG) method is employed to obtain the functionally graded (FG) plate’s natural frequencies. Effects of two parameters including thickness to height ratio and volume fraction index on the natural frequencies and total mass of plate are studied by using the MLPG results. Then the first natural frequency of the plate, for different conditions where MLPG data are not available, is predicted by an artificial neural network (ANN) approach which is trained by back-error propagation (BEP) technique. The ANN results show that the predicted data are in good agreement with the actual one. To maximize the first natural frequency and minimize the mass of FG plate simultaneously, the weighted sum optimization approach and PSO algorithm are used. However, the proposed optimization process of this study can provide the designers of FG plates with useful data.Keywords: optimal design, natural frequency, FG plate, hybrid meshless method, MLPG method, ANN approach, particle swarm optimization
Procedia PDF Downloads 3661286 High-Resolution ECG Automated Analysis and Diagnosis
Authors: Ayad Dalloo, Sulaf Dalloo
Abstract:
Electrocardiogram (ECG) recording is prone to complications, on analysis by physicians, due to noise and artifacts, thus creating ambiguity leading to possible error of diagnosis. Such drawbacks may be overcome with the advent of high resolution Methods, such as Discrete Wavelet Analysis and Digital Signal Processing (DSP) techniques. This ECG signal analysis is implemented in three stages: ECG preprocessing, features extraction and classification with the aim of realizing high resolution ECG diagnosis and improved detection of abnormal conditions in the heart. The preprocessing stage involves removing spurious artifacts (noise), due to such factors as muscle contraction, motion, respiration, etc. ECG features are extracted by applying DSP and suggested sloping method techniques. These measured features represent peak amplitude values and intervals of P, Q, R, S, R’, and T waves on ECG, and other features such as ST elevation, QRS width, heart rate, electrical axis, QR and QT intervals. The classification is preformed using these extracted features and the criteria for cardiovascular diseases. The ECG diagnostic system is successfully applied to 12-lead ECG recordings for 12 cases. The system is provided with information to enable it diagnoses 15 different diseases. Physician’s and computer’s diagnoses are compared with 90% agreement, with respect to physician diagnosis, and the time taken for diagnosis is 2 seconds. All of these operations are programmed in Matlab environment.Keywords: ECG diagnostic system, QRS detection, ECG baseline removal, cardiovascular diseases
Procedia PDF Downloads 2951285 Effect of Information and Communication Intervention on Stable Economic Growth in Ethiopia
Authors: Medhin Haftom Hailu
Abstract:
The advancement of information technology has significantly impacted Ethiopia's economy, driving innovation, productivity, job creation, and global connectivity. This research examined the impact of contemporary information and communication technologies on Ethiopian economic progress. The study examined eight variables, including mobile, internet, and fixed-line penetration rates, and five macroeconomic control variables. The results showed a positive and strong effect of ICT on economic growth in Ethiopia, with 1% increase in mobile, internet, and fixed line services penetration indexes resulting in an 8.03, 10.05, and 30.06% increase in real GDP. The Granger causality test showed that all ICT variables Granger caused economic growth, but economic growth Granger caused mobile penetration rate only. The study suggests that coordinated ICT infrastructure development, increased telecom service accessibility, and increased competition in the telecom market are crucial for Ethiopia's economic growth. Ethiopia is attempting to establish a digital economy through massive investment in ensuring ICT quality and accessibility. Thus, the research could enhance in understanding of the economic impact of ICT expansion for successful ICT policy interventions for future research.Keywords: economic growth, cointegration and error correction, ICT expansion, granger causality, penetration
Procedia PDF Downloads 771284 Analysis of Cardiovascular Diseases Using Artificial Neural Network
Authors: Jyotismita Talukdar
Abstract:
In this paper, a study has been made on the possibility and accuracy of early prediction of several Heart Disease using Artificial Neural Network. (ANN). The study has been made in both noise free environment and noisy environment. The data collected for this analysis are from five Hospitals. Around 1500 heart patient’s data has been collected and studied. The data is analysed and the results have been compared with the Doctor’s diagnosis. It is found that, in noise free environment, the accuracy varies from 74% to 92%and in noisy environment (2dB), the results of accuracy varies from 62% to 82%. In the present study, four basic attributes considered are Blood Pressure (BP), Fasting Blood Sugar (FBS), Thalach (THAL) and Cholesterol (CHOL.). It has been found that highest accuracy(93%), has been achieved in case of PPI( Post-Permanent-Pacemaker Implementation ), around 79% in case of CAD(Coronary Artery disease), 87% in DCM (Dilated Cardiomyopathy), 89% in case of RHD&MS(Rheumatic heart disease with Mitral Stenosis), 75 % in case of RBBB +LAFB (Right Bundle Branch Block + Left Anterior Fascicular Block), 72% for CHB(Complete Heart Block) etc. The lowest accuracy has been obtained in case of ICMP (Ischemic Cardiomyopathy), about 38% and AF( Atrial Fibrillation), about 60 to 62%.Keywords: coronary heart disease, chronic stable angina, sick sinus syndrome, cardiovascular disease, cholesterol, Thalach
Procedia PDF Downloads 1741283 Methods of Variance Estimation in Two-Phase Sampling
Authors: Raghunath Arnab
Abstract:
The two-phase sampling which is also known as double sampling was introduced in 1938. In two-phase sampling, samples are selected in phases. In the first phase, a relatively large sample of size is selected by some suitable sampling design and only information on the auxiliary variable is collected. During the second phase, a sample of size is selected either from, the sample selected in the first phase or from the entire population by using a suitable sampling design and information regarding the study and auxiliary variable is collected. Evidently, two phase sampling is useful if the auxiliary information is relatively easy and cheaper to collect than the study variable as well as if the strength of the relationship between the variables and is high. If the sample is selected in more than two phases, the resulting sampling design is called a multi-phase sampling. In this article we will consider how one can use data collected at the first phase sampling at the stages of estimation of the parameter, stratification, selection of sample and their combinations in the second phase in a unified setup applicable to any sampling design and wider classes of estimators. The problem of the estimation of variance will also be considered. The variance of estimator is essential for estimating precision of the survey estimates, calculation of confidence intervals, determination of the optimal sample sizes and for testing of hypotheses amongst others. Although, the variance is a non-negative quantity but its estimators may not be non-negative. If the estimator of variance is negative, then it cannot be used for estimation of confidence intervals, testing of hypothesis or measure of sampling error. The non-negativity properties of the variance estimators will also be studied in details.Keywords: auxiliary information, two-phase sampling, varying probability sampling, unbiased estimators
Procedia PDF Downloads 5861282 Analysis of the Lung Microbiome in Cystic Fibrosis Patients Using 16S Sequencing
Authors: Manasvi Pinnaka, Brianna Chrisman
Abstract:
Cystic fibrosis patients often develop lung infections that range anywhere in severity from mild to life-threatening due to the presence of thick and sticky mucus that fills their airways. Since many of these infections are chronic, they not only affect a patient’s ability to breathe but also increase the chances of mortality by respiratory failure. With a publicly available dataset of DNA sequences from bacterial species in the lung microbiome of cystic fibrosis patients, the correlations between different microbial species in the lung and the extent of deterioration of lung function were investigated. 16S sequencing technologies were used to determine the microbiome composition of the samples in the dataset. For the statistical analyses, referencing helped distinguish between taxonomies, and the proportions of certain taxa relative to another were determined. It was found that the Fusobacterium, Actinomyces, and Leptotrichia microbial types all had a positive correlation with the FEV1 score, indicating the potential displacement of these species by pathogens as the disease progresses. However, the dominant pathogens themselves, including Pseudomonas aeruginosa and Staphylococcus aureus, did not have statistically significant negative correlations with the FEV1 score as described by past literature. Examining the lung microbiology of cystic fibrosis patients can help with the prediction of the current condition of lung function, with the potential to guide doctors when designing personalized treatment plans for patients.Keywords: bacterial infections, cystic fibrosis, lung microbiome, 16S sequencing
Procedia PDF Downloads 971281 Attempt Survivor Families’ Views on Criminalizing Attempted Suicide in Ghana
Authors: Joseph Osafo, Winifred Asare-Doku, Charity Akotia
Abstract:
Decriminalizing suicide is one of the major goals of suicide prevention worldwide. In Ghana, suicide is legally prescribed and there is a wide-spread societal condemnation of the act, the survivor and families share the stigma. Evidence and advocacy continue to mount towards pressuring the government, the legal fraternity and lawmakers to consider decriminalizing the act. However, within this discourse, the views of families of attempt survivors are absent. The purpose of this study was to explore from relatives of suicide attempters their reactions towards the criminality of suicide attempt in the country. A total of 10 relatives of suicide attempters were interviewed using a semi-structured interview guide. Thematic analysis was used to analyze the data. We found that there were divergent views from families on decriminalizing suicide. We generated two major themes; Out-group bias versus In-group bias. Half of the participants opined that suicide attempt should not be decriminalized and others advocated for help and mental health care for victims of the suicide attempt. It was generally observed that although all 10 participants were cognizant that suicide attempt is a crime in Ghana, they preferred their relatives were spared from prosecution. The findings indicate incongruity, especially when participants want their relatives to avoid jail term but want the law that criminalizes suicide to remain. Findings are explained using the Fundamental Attribution Error and the concept of Kin selection. Implications for public education on decriminalization and advocacy are addressed.Keywords: decriminalization, families, Ghana suicide, suicide attempt
Procedia PDF Downloads 5191280 Development of a Novel Score for Early Detection of Hepatocellular Carcinoma in Patients with Hepatitis C Virus
Authors: Hatem A. El-Mezayen, Hossam Darwesh
Abstract:
Background/Aim: Hepatocellular carcinoma (HCC) is often diagnosed at advanced stage where effective therapies are lacking. Identification of new scoring system is needed to discriminate HCC patients from those with chronic liver disease. Based on the link between vascular endothelial growth factor (VEGF) and HCC progression, we aimed to develop a novel score based on combination of VEGF and routine laboratory tests for early prediction of HCC. Methods: VEGF was assayed for HCC group (123), liver cirrhosis group (210) and control group (50) by Enzyme Linked Immunosorbent Assay (ELISA). Data from all groups were retrospectively analyzed including α feto protein (AFP), international normalized ratio (INR), albumin and platelet count, transaminases, and age. Areas under ROC curve were used to develop the score. Results: A novel index named hepatocellular carcinoma-vascular endothelial growth factor score (HCC-VEGF score)=1.26 (numerical constant) + 0.05 ×AFP (U L-1)+0.038 × VEGF(ng ml-1)+0.004× INR –1.02 × Albumin (g l-1)–0.002 × Platelet count × 109 l-1 was developed. HCC-VEGF score produce area under ROC curve of 0.98 for discriminating HCC patients from liver cirrhosis with sensitivity of 91% and specificity of 82% at cut-off 4.4 (ie less than 4.4 considered cirrhosis and greater than 4.4 considered HCC). Conclusion: Hepatocellular carcinoma-VEGF score could replace AFP in HCC screening and follow up of cirrhotic patients.Keywords: Hepatocellular carcinoma, cirrhosis, HCV, diagnosis, tumor markers
Procedia PDF Downloads 3201279 Design and Control of a Knee Rehabilitation Device Using an MR-Fluid Brake
Authors: Mina Beheshti, Vida Shams, Mojtaba Esfandiari, Farzaneh Abdollahi, Abdolreza Ohadi
Abstract:
Most of the people who survive a stroke need rehabilitation tools to regain their mobility. The core function of these devices is a brake actuator. The goal of this study is to design and control a magnetorheological brake which can be used as a rehabilitation tool. In fact, the fluid used in this brake is called magnetorheological fluid or MR that properties can change by variation of the magnetic field. The braking properties can be set as control by using this feature of the fluid. In this research, different MR brake designs are first introduced in each design, and the dimensions of the brake have been determined based on the required torque for foot movement. To calculate the brake dimensions, it is assumed that the shear stress distribution in the fluid is uniform and the fluid is in its saturated state. After designing the rehabilitation brake, the mathematical model of the healthy movement of a healthy person is extracted. Due to the nonlinear nature of the system and its variability, various adaptive controllers, neural networks, and robust have been implemented to estimate the parameters and control the system. After calculating torque and control current, the best type of controller in terms of error and control current has been selected. Finally, this controller is implemented on the experimental data of the patient's movements, and the control current is calculated to achieve the desired torque and motion.Keywords: rehabilitation, magnetorheological fluid, knee, brake, adaptive control, robust control, neural network control, torque control
Procedia PDF Downloads 1511278 Development of Time Series Forecasting Model for Dengue Cases in Nakhon Si Thammarat, Southern Thailand
Authors: Manit Pollar
Abstract:
Identifying the dengue epidemic periods early would be helpful to take necessary actions to prevent the dengue outbreaks. Providing an accurate prediction on dengue epidemic seasons will allow sufficient time to take the necessary decisions and actions to safeguard the situation for local authorities. This study aimed to develop a forecasting model on number of dengue incidences in Nakhon Si Thammarat Province, Southern Thailand using time series analysis. We develop Seasonal Autoregressive Moving Average (SARIMA) models on the monthly data collected between 2003-2011 and validated the models using data collected between January-September 2012. The result of this study revealed that the SARIMA(1,1,0)(1,2,1)12 model closely described the trends and seasons of dengue incidence and confirmed the existence of dengue fever cases in Nakhon Si Thammarat for the years between 2003-2011. The study showed that the one-step approach for predicting dengue incidences provided significantly more accurate predictions than the twelve-step approach. The model, even if based purely on statistical data analysis, can provide a useful basis for allocation of resources for disease prevention.Keywords: SARIMA, time series model, dengue cases, Thailand
Procedia PDF Downloads 3561277 Drying Kinects of Soybean Seeds
Authors: Amanda Rithieli Pereira Dos Santos, Rute Quelvia De Faria, Álvaro De Oliveira Cardoso, Anderson Rodrigo Da Silva, Érica Leão Fernandes Araújo
Abstract:
The study of the kinetics of drying has great importance for the mathematical modeling, allowing to know about the processes of transference of heat and mass between the products and to adjust dryers managing new technologies for these processes. The present work had the objective of studying the kinetics of drying of soybean seeds and adjusting different statistical models to the experimental data varying cultivar and temperature. Soybean seeds were pre-dried in a natural environment in order to reduce and homogenize the water content to the level of 14% (b.s.). Then, drying was carried out in a forced air circulation oven at controlled temperatures of 38, 43, 48, 53 and 58 ± 1 ° C, using two soybean cultivars, BRS 8780 and Sambaíba, until reaching a hygroscopic equilibrium. The experimental design was completely randomized in factorial 5 x 2 (temperature x cultivar) with 3 replicates. To the experimental data were adjusted eleven statistical models used to explain the drying process of agricultural products. Regression analysis was performed using the least squares Gauss-Newton algorithm to estimate the parameters. The degree of adjustment was evaluated from the analysis of the coefficient of determination (R²), the adjusted coefficient of determination (R² Aj.) And the standard error (S.E). The models that best represent the drying kinetics of soybean seeds are those of Midilli and Logarítmico.Keywords: curve of drying seeds, Glycine max L., moisture ratio, statistical models
Procedia PDF Downloads 6261276 Diagnostic Performance of Tumor Associated Trypsin Inhibitor in Early Detection of Hepatocellular Carcinoma in Patients with Hepatitis C Virus
Authors: Aml M. El-Sharkawy, Hossam M. Darwesh
Abstract:
Abstract— Background/Aim: Hepatocellular carcinoma (HCC) is often diagnosed at advanced stage where effective therapies are lacking. Identification of new scoring system is needed to discriminate HCC patients from those with chronic liver disease. Based on the link between tumor associated trypsin inhibitor (TATI) and HCC progression, we aimed to develop a novel score based on combination of TATI and routine laboratory tests for early prediction of HCC. Methods: TATI was assayed for HCC group (123), liver cirrhosis group (210) and control group (50) by Enzyme Linked Immunosorbent Assay (ELISA). Data from all groups were retrospectively analyzed including α feto protein (AFP), international normalized ratio (INR), albumin and platelet count, transaminases, and age. Areas under ROC curve were used to develop the score. Results: A novel index named hepatocellular carcinoma-vascular endothelial growth factor score (HCC-TATI score) = 3.1 (numerical constant) + 0.09 ×AFP (U L-1) + 0.067 × TATI (ng ml-1) + 0.16 × INR – 1.17 × Albumin (g l-1) – 0.032 × Platelet count × 109 l-1 was developed. HCC-TATI score produce area under ROC curve of 0.98 for discriminating HCC patients from liver cirrhosis with sensitivity of 91% and specificity of 82% at cut-off 6.5 (ie less than 6.5 considered cirrhosis and greater than 4.4 considered HCC). Conclusion: Hepatocellular carcinoma-TATI score could replace AFP in HCC screening and follow up of cirrhotic patients.Keywords: Hepatocellular carcinoma, cirrhosis, HCV, diagnosis, TATI
Procedia PDF Downloads 3361275 An Empirical Evaluation of Performance of Machine Learning Techniques on Imbalanced Software Quality Data
Authors: Ruchika Malhotra, Megha Khanna
Abstract:
The development of change prediction models can help the software practitioners in planning testing and inspection resources at early phases of software development. However, a major challenge faced during the training process of any classification model is the imbalanced nature of the software quality data. A data with very few minority outcome categories leads to inefficient learning process and a classification model developed from the imbalanced data generally does not predict these minority categories correctly. Thus, for a given dataset, a minority of classes may be change prone whereas a majority of classes may be non-change prone. This study explores various alternatives for adeptly handling the imbalanced software quality data using different sampling methods and effective MetaCost learners. The study also analyzes and justifies the use of different performance metrics while dealing with the imbalanced data. In order to empirically validate different alternatives, the study uses change data from three application packages of open-source Android data set and evaluates the performance of six different machine learning techniques. The results of the study indicate extensive improvement in the performance of the classification models when using resampling method and robust performance measures.Keywords: change proneness, empirical validation, imbalanced learning, machine learning techniques, object-oriented metrics
Procedia PDF Downloads 4181274 Anomaly Detection in a Data Center with a Reconstruction Method Using a Multi-Autoencoders Model
Authors: Victor Breux, Jérôme Boutet, Alain Goret, Viviane Cattin
Abstract:
Early detection of anomalies in data centers is important to reduce downtimes and the costs of periodic maintenance. However, there is little research on this topic and even fewer on the fusion of sensor data for the detection of abnormal events. The goal of this paper is to propose a method for anomaly detection in data centers by combining sensor data (temperature, humidity, power) and deep learning models. The model described in the paper uses one autoencoder per sensor to reconstruct the inputs. The auto-encoders contain Long-Short Term Memory (LSTM) layers and are trained using the normal samples of the relevant sensors selected by correlation analysis. The difference signal between the input and its reconstruction is then used to classify the samples using feature extraction and a random forest classifier. The data measured by the sensors of a data center between January 2019 and May 2020 are used to train the model, while the data between June 2020 and May 2021 are used to assess it. Performances of the model are assessed a posteriori through F1-score by comparing detected anomalies with the data center’s history. The proposed model outperforms the state-of-the-art reconstruction method, which uses only one autoencoder taking multivariate sequences and detects an anomaly with a threshold on the reconstruction error, with an F1-score of 83.60% compared to 24.16%.Keywords: anomaly detection, autoencoder, data centers, deep learning
Procedia PDF Downloads 1921273 A Low Cost Non-Destructive Grain Moisture Embedded System for Food Safety and Quality
Authors: Ritula Thakur, Babankumar S. Bansod, Puneet Mehta, S. Chatterji
Abstract:
Moisture plays an important role in storage, harvesting and processing of food grains and related agricultural products. It is an important characteristic of most agricultural products for maintenance of quality. Accurate knowledge of the moisture content can be of significant value in maintaining quality and preventing contamination of cereal grains. The present work reports the design and development of microcontroller based low cost non-destructive moisture meter, which uses complex impedance measurement method for moisture measurement of wheat using parallel plate capacitor arrangement. Moisture can conveniently be sensed by measuring the complex impedance using a small parallel-plate capacitor sensor filled with the kernels in-between the two plates of sensor, exciting the sensor at 30 KHz and 100 KHz frequencies. The effects of density and temperature variations were compensated by providing suitable compensations in the developed algorithm. The results were compared with standard dry oven technique and the developed method was found to be highly accurate with less than 1% error. The developed moisture meter is low cost, highly accurate, non-destructible method for determining the moisture of grains utilizing the fast computing capabilities of microcontroller.Keywords: complex impedance, moisture content, electrical properties, safety of food
Procedia PDF Downloads 4611272 Strict Stability of Fuzzy Differential Equations by Lyapunov Functions
Authors: Mustafa Bayram Gücen, Coşkun Yakar
Abstract:
In this study, we have investigated the strict stability of fuzzy differential systems and we compare the classical notion of strict stability criteria of ordinary differential equations and the notion of strict stability of fuzzy differential systems. In addition that, we present definitions of stability and strict stability of fuzzy differential equations and also we have some theorems and comparison results. Strict Stability is a different stability definition and this stability type can give us an information about the rate of decay of the solutions. Lyapunov’s second method is a standard technique used in the study of the qualitative behavior of fuzzy differential systems along with a comparison result that allows the prediction of behavior of a fuzzy differential system when the behavior of the null solution of a fuzzy comparison system is known. This method is a usefull for investigating strict stability of fuzzy systems. First of all, we present definitions and necessary background material. Secondly, we discuss and compare the differences between the classical notion of stability and the recent notion of strict stability. And then, we have a comparison result in which the stability properties of the null solution of the comparison system imply the corresponding stability properties of the fuzzy differential system. Consequently, we give the strict stability results and a comparison theorem. We have used Lyapunov second method and we have proved a comparison result with scalar differential equations.Keywords: fuzzy systems, fuzzy differential equations, fuzzy stability, strict stability
Procedia PDF Downloads 2481271 Induction Heating and Electromagnetic Stirring of Bi-Phasic Metal/Glass Molten Bath for Mixed Nuclear Waste Treatment
Authors: P. Charvin, R. Bourrou, F. Lemont, C. Lafon, A. Russello
Abstract:
For nuclear waste treatment and confinement, a specific IN-CAN melting module based on low-frequency induction heating have been designed. The frequency of 50Hz has been chosen to improve penetration length through metal. In this design, the liquid metal, strongly stirred by electromagnetic effects, presents shape of a dome caused by strong Laplace forces developing in the bulk of bath. Because of a lower density, the glass phase is located above the metal phase and is heated and stirred by metal through interface. Electric parameters (Intensity, frequency) give precious information about metal load and composition (resistivity of alloy) through impedance modification. Then, power supply can be adapted to energy transfer efficiency for suitable process supervision. Modeling of this system allows prediction of metal dome shape (in agreement with experimental measurement with a specific device), glass and metal velocity, heat and motion transfer through interface. MHD modeling is achieved with COMSOL and Fluent. First, a simplified model is used to obtain the shape of the metal dome. Then the shape is fixed to calculate the fluid flow and the thermal part.Keywords: electromagnetic stirring, induction heating, interface modeling, metal load
Procedia PDF Downloads 2641270 Thermophysical Properties and Kinetic Study of Dioscorea bulbifera
Authors: Emmanuel Chinagorom Nwadike, Joseph Tagbo Nwabanne, Matthew Ndubuisi Abonyi, Onyemazu Andrew Azaka
Abstract:
This research focused on the modeling of the convective drying of aerial yam using finite element methods. The thermo-gravimetric analyzer was used to determine the thermal stability of the sample. An aerial yam sample of size 30 x 20 x 4 mm was cut with a mold designed for the purpose and dried in a convective dryer set at 4m/s fan speed and temperatures of 68.58 and 60.56°C. The volume shrinkage of the resultant dried sample was determined by immersing the sample in a toluene solution. The finite element analysis was done with PDE tools in Matlab 2015. Seven kinetic models were employed to model the drying process. The result obtained revealed three regions in the thermogravimetric analysis (TGA) profile of aerial yam. The maximum thermal degradation rates of the sample occurred at 432.7°C. The effective thermal diffusivity of the sample increased as the temperature increased from 60.56°C to 68.58°C. The finite element prediction of moisture content of aerial yam at an air temperature of 68.58°C and 60.56°C shows R² of 0.9663 and 0.9155, respectively. There was a good agreement between the finite element predicted moisture content and the measured moisture content, which is indicative of a highly reliable finite element model developed. The result also shows that the best kinetic model for the aerial yam under the given drying conditions was the Logarithmic model with a correlation coefficient of 0.9991.Keywords: aerial yam, finite element, convective, effective, diffusivity
Procedia PDF Downloads 152