Search results for: semantic clinical classification
5481 Collective Intelligence-Based Early Warning Management for Agriculture
Authors: Jarbas Lopes Cardoso Jr., Frederic Andres, Alexandre Guitton, Asanee Kawtrakul, Silvio E. Barbin
Abstract:
The important objective of the CyberBrain Mass Agriculture Alarm Acquisition and Analysis (CBMa4) project is to minimize the impacts of diseases and disasters on rice cultivation. For example, early detection of insects will reduce the volume of insecticides that is applied to the rice fields through the use of CBMa4 platform. In order to reach this goal, two major factors need to be considered: (1) the social network of smart farmers; and (2) the warning data alarm acquisition and analysis component. This paper outlines the process for collecting the warning and improving the decision-making result to the warning. It involves two sub-processes: the warning collection and the understanding enrichment. Human sensors combine basic suitable data processing techniques in order to extract warning related semantic according to collective intelligence. We identify each warning by a semantic content called 'warncons' with multimedia metaphors and metadata related to these metaphors. It is important to describe the metric to measuring the relation among warncons. With this knowledge, a collective intelligence-based decision-making approach determines the action(s) to be launched regarding one or a set of warncons.Keywords: agricultural engineering, warning systems, social network services, context awareness
Procedia PDF Downloads 3825480 Automatic Staging and Subtype Determination for Non-Small Cell Lung Carcinoma Using PET Image Texture Analysis
Authors: Seyhan Karaçavuş, Bülent Yılmaz, Ömer Kayaaltı, Semra İçer, Arzu Taşdemir, Oğuzhan Ayyıldız, Kübra Eset, Eser Kaya
Abstract:
In this study, our goal was to perform tumor staging and subtype determination automatically using different texture analysis approaches for a very common cancer type, i.e., non-small cell lung carcinoma (NSCLC). Especially, we introduced a texture analysis approach, called Law’s texture filter, to be used in this context for the first time. The 18F-FDG PET images of 42 patients with NSCLC were evaluated. The number of patients for each tumor stage, i.e., I-II, III or IV, was 14. The patients had ~45% adenocarcinoma (ADC) and ~55% squamous cell carcinoma (SqCCs). MATLAB technical computing language was employed in the extraction of 51 features by using first order statistics (FOS), gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and Laws’ texture filters. The feature selection method employed was the sequential forward selection (SFS). Selected textural features were used in the automatic classification by k-nearest neighbors (k-NN) and support vector machines (SVM). In the automatic classification of tumor stage, the accuracy was approximately 59.5% with k-NN classifier (k=3) and 69% with SVM (with one versus one paradigm), using 5 features. In the automatic classification of tumor subtype, the accuracy was around 92.7% with SVM one vs. one. Texture analysis of FDG-PET images might be used, in addition to metabolic parameters as an objective tool to assess tumor histopathological characteristics and in automatic classification of tumor stage and subtype.Keywords: cancer stage, cancer cell type, non-small cell lung carcinoma, PET, texture analysis
Procedia PDF Downloads 3265479 Contextual Senses of Ambiguous Words Based on Cognitive Semantics
Authors: Madhavi
Abstract:
All linguistic units are context-dependent. They occur in particular settings, from which they derive much of their import, and are recognized by speakers as distinct entities only through a process of abstraction. Most of the words have several concepts associated with them and convey a number of meanings in different contexts in any language. For instance, there are different uses of the word good as an adjective from English. The adjective good expresses many senses like (1) ‘high quality of someone or something’ (2) ‘efficient’ (3) ‘virtuous’ (4) ‘reliable’ etc. These senses will be analyzed by using cognitive semantics framework. The context has the power to insulate one meaning from all the other meanings in communication. This paper will provide a cognitive semantic analysis. The basic tenet of cognitive semantics is the sense of a word is the way we conceptualize it. Our conceptualization is based on the physical experience we go through. Cognitive semantics tries to capture this conceptualization in terms of some categories like schema, frame, and domain. Cognitive semantics is a subfield of cognitive linguistics. Cognitive linguistics studies the language creation, learning, and usage by the reference to human cognition. The semantic structure is conceptual structure which is related to the concepts which are the elements of reason and constitute the meanings of words and linguistic expressions. Cognitive semantics studies how our mind works for the meaning of any word and how it perceives meaning from the environment through senses and works to map with the knowledge which already exists in our mind through experience. In the present paper, the senses are further classified into some categories.Keywords: cognitive, contexts, semantics, senses
Procedia PDF Downloads 2195478 Systematic Evaluation of Convolutional Neural Network on Land Cover Classification from Remotely Sensed Images
Authors: Eiman Kattan, Hong Wei
Abstract:
In using Convolutional Neural Network (CNN) for classification, there is a set of hyperparameters available for the configuration purpose. This study aims to evaluate the impact of a range of parameters in CNN architecture i.e. AlexNet on land cover classification based on four remotely sensed datasets. The evaluation tests the influence of a set of hyperparameters on the classification performance. The parameters concerned are epoch values, batch size, and convolutional filter size against input image size. Thus, a set of experiments were conducted to specify the effectiveness of the selected parameters using two implementing approaches, named pertained and fine-tuned. We first explore the number of epochs under several selected batch size values (32, 64, 128 and 200). The impact of kernel size of convolutional filters (1, 3, 5, 7, 10, 15, 20, 25 and 30) was evaluated against the image size under testing (64, 96, 128, 180 and 224), which gave us insight of the relationship between the size of convolutional filters and image size. To generalise the validation, four remote sensing datasets, AID, RSD, UCMerced and RSCCN, which have different land covers and are publicly available, were used in the experiments. These datasets have a wide diversity of input data, such as number of classes, amount of labelled data, and texture patterns. A specifically designed interactive deep learning GPU training platform for image classification (Nvidia Digit) was employed in the experiments. It has shown efficiency in both training and testing. The results have shown that increasing the number of epochs leads to a higher accuracy rate, as expected. However, the convergence state is highly related to datasets. For the batch size evaluation, it has shown that a larger batch size slightly decreases the classification accuracy compared to a small batch size. For example, selecting the value 32 as the batch size on the RSCCN dataset achieves the accuracy rate of 90.34 % at the 11th epoch while decreasing the epoch value to one makes the accuracy rate drop to 74%. On the other extreme, setting an increased value of batch size to 200 decreases the accuracy rate at the 11th epoch is 86.5%, and 63% when using one epoch only. On the other hand, selecting the kernel size is loosely related to data set. From a practical point of view, the filter size 20 produces 70.4286%. The last performed image size experiment shows a dependency in the accuracy improvement. However, an expensive performance gain had been noticed. The represented conclusion opens the opportunities toward a better classification performance in various applications such as planetary remote sensing.Keywords: CNNs, hyperparamters, remote sensing, land cover, land use
Procedia PDF Downloads 1695477 Impact of Self-Efficacy, Resilience and Social Support on Vicarious Trauma among Clinical Psychologists, Counselors and Teachers of Special Schools
Authors: Hamna Hamid, Kashmala Zaman
Abstract:
The aim of this study was to evaluate the relationship between self-efficacy, resilience and social support among clinical psychologists, counselors and teachers of special schools. The study also assesses the gender differences on self-efficacy, resilience, social support and vicarious trauma and also vicarious trauma differences among three professions i.e. clinical psychologists, counselors and teachers of special schools. A sample of 150 women and 97 men were handed out a set questionnaire to complete: General Self-Efficacy Scale, Brief Resilience Scale, Multidimensional Scale of Perceived Social Support and Vicarious Trauma Scale. Results showed that there is significant negative correlation between self-efficacy, resilience and vicarious trauma. Women experiences higher levels of vicarious trauma as compared to men. While clinical psychologists and counselors experience higher levels of vicarious trauma as compared to teachers of special schools. Moderation effect of social support is not significant towards resilience and vicarious trauma.Keywords: self-efficacy, resilience, vicarious trauma, social-support
Procedia PDF Downloads 755476 A Convolution Neural Network Approach to Predict Pes-Planus Using Plantar Pressure Mapping Images
Authors: Adel Khorramrouz, Monireh Ahmadi Bani, Ehsan Norouzi, Morvarid Lalenoor
Abstract:
Background: Plantar pressure distribution measurement has been used for a long time to assess foot disorders. Plantar pressure is an important component affecting the foot and ankle function and Changes in plantar pressure distribution could indicate various foot and ankle disorders. Morphologic and mechanical properties of the foot may be important factors affecting the plantar pressure distribution. Accurate and early measurement may help to reduce the prevalence of pes planus. With recent developments in technology, new techniques such as machine learning have been used to assist clinicians in predicting patients with foot disorders. Significance of the study: This study proposes a neural network learning-based flat foot classification methodology using static foot pressure distribution. Methodologies: Data were collected from 895 patients who were referred to a foot clinic due to foot disorders. Patients with pes planus were labeled by an experienced physician based on clinical examination. Then all subjects (with and without pes planus) were evaluated for static plantar pressures distribution. Patients who were diagnosed with the flat foot in both feet were included in the study. In the next step, the leg length was normalized and the network was trained for plantar pressure mapping images. Findings: From a total of 895 image data, 581 were labeled as pes planus. A computational neural network (CNN) ran to evaluate the performance of the proposed model. The prediction accuracy of the basic CNN-based model was performed and the prediction model was derived through the proposed methodology. In the basic CNN model, the training accuracy was 79.14%, and the test accuracy was 72.09%. Conclusion: This model can be easily and simply used by patients with pes planus and doctors to predict the classification of pes planus and prescreen for possible musculoskeletal disorders related to this condition. However, more models need to be considered and compared for higher accuracy.Keywords: foot disorder, machine learning, neural network, pes planus
Procedia PDF Downloads 3625475 Prognostic Value in Meningioma Patients’: A Clinical-Histopathological Study
Authors: Ilham Akbar Rahman, Aflah Dhea Bariz Yasta, Iin Fadhilah Utami Tamasse, Devina Juanita
Abstract:
Meningioma is adult brain tumors originating from the meninges covering the brain and spinal cord. The females have approximately twice higher 2:1 than male in the incidence of meningioma. This study aimed to analyze the histopathological grading and clinical aspect in predicting the prognosis of meningioma patients. An observational study with cross sectional design was used on 53 meningioma patients treated at Dr. Wahidin Sudirohusodo hospital in 2016. The data then were analyzed using SPSS 20.0. Of 53 patients, mostly 41 (77,4%) were female and 12 (22,6%) were male. The distribution of histopathology patients showed the meningothelial meningioma of 18 (43,9%) as the most type found. Fibroplastic meningioma were 8 (19,5%), while atypical meningioma and psammomatous meningioma were 6 (14,6%) each. The rest were malignant meningioma and angiomatous meningioma which found in respectively 2 (4,9%) and 1 (2,4%). Our result found significant finding that mostly male were fibroblastic meningioma (50%), however meningothelial meningioma were found in the majority of female (54,8%) and also seizure comprised only in higher grade meningioma. On the outcome of meningioma patient treated operatively, histopathological grade remained insignificant (p > 0,05). This study can be used as prognostic value of meningioma patients based on gender, histopathological grade, and clinical manifestation. Overall, the outcome of the meningioma’s patients is good and promising as long as it is well managed.Keywords: meningioma, prognostic value, histopathological grading, clinical manifestation
Procedia PDF Downloads 1715474 Nursing Students' Experience of Using Electronic Health Record System in Clinical Placements
Authors: Nurten Tasdemir, Busra Baloglu, Zeynep Cingoz, Can Demirel, Zeki Gezer, Barıs Efe
Abstract:
Student nurses are increasingly exposed to technology in the workplace after graduation with the growing numbers of electric health records (EHRs), handheld computers, barcode scanner medication dispensing systems, and automatic capture of patient data such as vital signs. Internationally, electronic health records (EHRs) systems are being implemented and evaluated. Students will inevitably encounter EHRs in the clinical learning environment and their professional practice. Nursing students must develop competency in the use of EHR. Aim: The study aimed to examine nursing students’ experiences of learning to use electronic health records (EHR) in clinical placements. Method: This study adopted a descriptive approach. The study population consisted of second and third-year nursing students at the Zonguldak School of Health in the West Black Sea Region of Turkey; the study was conducted during the 2015–2016 academic year. The sample consisted of 315 (74.1% of 425 students) nursing students who volunteered to participate. The students, who were involved in clinical practice, were invited to participate in the study Data were collected by a questionnaire designed by the researchers based on the relevant literature. Data were analyzed descriptively using the Statistical Package for Social Sciences (SPSS) for Windows version 16.0. The data are presented as means, standard deviations, and percentages. Approval for the study was obtained from the Ethical Committee of the University (Reg. Number: 29/03/2016/112) and the director of Nursing Department. Findings: A total of 315 students enrolled in this study, for a response rate of 74.1%. The mean age of the sample was 22.24 ± 1.37 (min: 19, max: 32) years, and most participants (79.7%) were female. Most of the nursing students (82.3%) stated that they use information technologies in clinical practice. Nearly half of the students (42.5%) reported that they have not accessed to EHR system. In addition, 61.6% of the students reported that insufficient computers available in clinical placement. Of the students, 84.7% reported that they prefer to have patient information from EHR system, and 63.8% of them found more effective to preparation for the clinical reporting. Conclusion: This survey indicated that nursing students experience to learn about EHR systems in clinical placements. For more effective learning environment nursing education should prepare nursing students for EHR systems in their educational life.Keywords: electronic health record, clinical placement, nursing student, nursing education
Procedia PDF Downloads 2915473 Enhancing Spatial Interpolation: A Multi-Layer Inverse Distance Weighting Model for Complex Regression and Classification Tasks in Spatial Data Analysis
Authors: Yakin Hajlaoui, Richard Labib, Jean-François Plante, Michel Gamache
Abstract:
This study introduces the Multi-Layer Inverse Distance Weighting Model (ML-IDW), inspired by the mathematical formulation of both multi-layer neural networks (ML-NNs) and Inverse Distance Weighting model (IDW). ML-IDW leverages ML-NNs' processing capabilities, characterized by compositions of learnable non-linear functions applied to input features, and incorporates IDW's ability to learn anisotropic spatial dependencies, presenting a promising solution for nonlinear spatial interpolation and learning from complex spatial data. it employ gradient descent and backpropagation to train ML-IDW, comparing its performance against conventional spatial interpolation models such as Kriging and standard IDW on regression and classification tasks using simulated spatial datasets of varying complexity. the results highlight the efficacy of ML-IDW, particularly in handling complex spatial datasets, exhibiting lower mean square error in regression and higher F1 score in classification.Keywords: deep learning, multi-layer neural networks, gradient descent, spatial interpolation, inverse distance weighting
Procedia PDF Downloads 525472 Radar Track-based Classification of Birds and UAVs
Authors: Altilio Rosa, Chirico Francesco, Foglia Goffredo
Abstract:
In recent years, the number of Unmanned Aerial Vehicles (UAVs) has significantly increased. The rapid development of commercial and recreational drones makes them an important part of our society. Despite the growing list of their applications, these vehicles pose a huge threat to civil and military installations: detection, classification and neutralization of such flying objects become an urgent need. Radar is an effective remote sensing tool for detecting and tracking flying objects, but scenarios characterized by the presence of a high number of tracks related to flying birds make especially challenging the drone detection task: operator PPI is cluttered with a huge number of potential threats and his reaction time can be severely affected. Flying birds compared to UAVs show similar velocity, RADAR cross-section and, in general, similar characteristics. Building from the absence of a single feature that is able to distinguish UAVs and birds, this paper uses a multiple features approach where an original feature selection technique is developed to feed binary classifiers trained to distinguish birds and UAVs. RADAR tracks acquired on the field and related to different UAVs and birds performing various trajectories were used to extract specifically designed target movement-related features based on velocity, trajectory and signal strength. An optimization strategy based on a genetic algorithm is also introduced to select the optimal subset of features and to estimate the performance of several classification algorithms (Neural network, SVM, Logistic regression…) both in terms of the number of selected features and misclassification error. Results show that the proposed methods are able to reduce the dimension of the data space and to remove almost all non-drone false targets with a suitable classification accuracy (higher than 95%).Keywords: birds, classification, machine learning, UAVs
Procedia PDF Downloads 2225471 Biorisk Management Education for Undergraduates Studying Clinical Microbiology at University in Japan
Authors: Shuji Fujimoto, Fumiko Kojima, Mika Shigematsu
Abstract:
Biorisk management (Biosafety/Biosecurity) is required for anyone working in a clinical laboratory (including medical/clinical research laboratories) where infectious agents and potentially hazardous biological materials are examined/stored. Proper education and training based on international standards of biorisk management should be provided not only as a part of laboratory safety program in work place but also as a part of introductory training at educational institutions for continuity and to elevate overall baseline of the biorisk management. We reported results of the pilot study of biorisk management education for graduate students majored in laboratory diagnostics previously. However, postgraduate education is still late in their profession and the participants’ interview also revealed importance and demands of earlier biorisk management education for undergraduates. The aim of this study is to identify the need for biosafety/biosecurity education and training program which is designed for undergraduate students who are entering the profession in clinical microbiology. We modified the previous program to include more basic topics and explanations (risk management, principles of safe clinical lab practices, personal protective equipment, disinfection, disposal of biological substances) and provided incorporating in the routine educational system for faculty of medical sciences in Kyushu University. The results of the pre and post examinations showed that the knowledge of the students on biorisk control had developed effectively as a proof of effectiveness of the program even in the undergraduate students. Our study indicates that administrating the basic biorisk management program in the earlier stage of learning will add positive impact to the understanding of biosafety to the health professional education.Keywords: biorisk management, biosafety, biosecurity, clinical microbiology, education for undergraduates
Procedia PDF Downloads 2115470 Evaluation of Clinical Decision Support System in Electronic Medical Record System: A Case of Malawi National Art Electronic Medical Record System
Authors: Pachawo Bisani, Goodall Nyirenda
Abstract:
The Malawi National Antiretroviral Therapy (NART) Electronic Medical Record (EMR) system was designed and developed with guidance from the Ministry of Health through the Department of HIV and AIDS (DHA) with the aim of supporting the management of HIV patient data and reporting in high prevalence ART clinics. As of 2021, the system has been scaled up to over 206 facilities across the country. The system is integrated with the clinical decision support system (CDSS) to assist healthcare providers in making a decision about an individual patient at a particular point in time. Despite NART EMR undergoing several evaluations and assessments, little has been done to evaluate the clinical decision support system in the NART EMR system. Hence, the study aimed to evaluate the use of CDSS in the NART EMR system in Malawi. The study adopted a mixed-method approach, and data was collected through interviews, observations, and questionnaires. The study has revealed that the CDSS tools were integrated into the ART clinic workflow, making it easy for the user to use it. The study has also revealed challenges in system reliability and information accuracy. Despite the challenges, the study further revealed that the system is effective and efficient, and overall, users are satisfied with the system. The study recommends that the implementers focus more on the logic behind the clinical decision-support intervention in order to address some of the concerns and enhance the accuracy of the information supplied. The study further suggests consulting the system's actual users throughout implementation.Keywords: clinical decision support system, electronic medical record system, usability, antiretroviral therapy
Procedia PDF Downloads 1005469 Patient-Reported Adverse Drug Reactions, Medication Adherence and Clinical Outcomes among major depression disorder Patients in Ethiopia: A Prospective Hospital Based Study.
Authors: Tadesse Melaku Abegaz
Abstract:
Background: there was paucity of data on the self-reported adverse drug reactions (ADRs), level of adherence and clinical outcomes with antidepressants among major depressive disorder (MDD) patients in Ethiopia. Hence, the present study sought to determine the level of adherence for and clinical outcome with antidepressants and the magnitude of ADRs. Methods: A prospective cross-sectional study was employed on MDD patients from September 2016 to January 2017 at Gondar university hospital psychiatry clinic. All patients who were available during the study period were included under the study population. The Naranjo adverse drug reaction probability scale was employed to assess the adverse drug reaction. The rate of medication adherence was determined using morisky medication adherence measurement scale eight. Clinical Outcome of patients was measured by using patient health questionnaire. Multivariable logistic carried out to determine factors for adherence and patient outcome. Results: two hundred seventy patients were participated in the study. More than half of the respondents were males 122(56.2%). The mean age of the participants was 30.94 ± 8.853. More than one-half of the subjects had low adherence to their medications 124(57.1%). About 186(85.7%) of patients encountered ADR. The most common ADR was weight gain 29(13.2). Around 198(92.2%) ADRs were probable and 19(8.8%) were possible. Patients with long standing MDD had high risk of non-adherence COR: 2.458[4.413-4.227], AOR: 2.424[1.185-4.961]. More than one-half 125(57.6) of respondents showed improved outcome. Optimal level of medication adherence was found to be associated with reduced risk of progression of the diseases COR: 0.37[0.110-5.379] and AOR: 0.432[0.201-0.909]. Conclusion: Patient reported adverse drug reactions were more prevalent in major depressive disorder patients. Adherence to medications was very poor in the setup. However, the clinical outcome was relatively higher. Long standing depression was associated with non-adherence. In addition, clinical outcome of patients were affected by non-adherence. Therefore, adherence enhancing interventions should be provided to improve medication adherence and patient outcome.Keywords: adverse drug reactions, clinical outcomes, Ethiopia, prospective study, medication adherence
Procedia PDF Downloads 2475468 Deep Graph Embeddings for the Analysis of Short Heartbeat Interval Time Series
Authors: Tamas Madl
Abstract:
Sudden cardiac death (SCD) constitutes a large proportion of cardiovascular mortalities, provides little advance warning, and the risk is difficult to recognize based on ubiquitous, low cost medical equipment such as the standard, 12-lead, ten second ECG. Autonomic abnormalities have been shown to be strongly predictive of SCD risk; yet current methods are not trivially applicable to the brevity and low temporal and electrical resolution of standard ECGs. Here, we build horizontal visibility graph representations of very short inter-beat interval time series, and perform unsuper- vised representation learning in order to convert these variable size objects into fixed-length vectors preserving similarity rela- tions. We show that such representations facilitate classification into healthy vs. at-risk patients on two different datasets, the Mul- tiparameter Intelligent Monitoring in Intensive Care II and the PhysioNet Sudden Cardiac Death Holter Database. Our results suggest that graph representation learning of heartbeat interval time series facilitates robust classification even in sequences as short as ten seconds.Keywords: sudden cardiac death, heart rate variability, ECG analysis, time series classification
Procedia PDF Downloads 2345467 Preparation of Papers - Developing a Leukemia Diagnostic System Based on Hybrid Deep Learning Architectures in Actual Clinical Environments
Authors: Skyler Kim
Abstract:
An early diagnosis of leukemia has always been a challenge to doctors and hematologists. On a worldwide basis, it was reported that there were approximately 350,000 new cases in 2012, and diagnosing leukemia was time-consuming and inefficient because of an endemic shortage of flow cytometry equipment in current clinical practice. As the number of medical diagnosis tools increased and a large volume of high-quality data was produced, there was an urgent need for more advanced data analysis methods. One of these methods was the AI approach. This approach has become a major trend in recent years, and several research groups have been working on developing these diagnostic models. However, designing and implementing a leukemia diagnostic system in real clinical environments based on a deep learning approach with larger sets remains complex. Leukemia is a major hematological malignancy that results in mortality and morbidity throughout different ages. We decided to select acute lymphocytic leukemia to develop our diagnostic system since acute lymphocytic leukemia is the most common type of leukemia, accounting for 74% of all children diagnosed with leukemia. The results from this development work can be applied to all other types of leukemia. To develop our model, the Kaggle dataset was used, which consists of 15135 total images, 8491 of these are images of abnormal cells, and 5398 images are normal. In this paper, we design and implement a leukemia diagnostic system in a real clinical environment based on deep learning approaches with larger sets. The proposed diagnostic system has the function of detecting and classifying leukemia. Different from other AI approaches, we explore hybrid architectures to improve the current performance. First, we developed two independent convolutional neural network models: VGG19 and ResNet50. Then, using both VGG19 and ResNet50, we developed a hybrid deep learning architecture employing transfer learning techniques to extract features from each input image. In our approach, fusing the features from specific abstraction layers can be deemed as auxiliary features and lead to further improvement of the classification accuracy. In this approach, features extracted from the lower levels are combined into higher dimension feature maps to help improve the discriminative capability of intermediate features and also overcome the problem of network gradient vanishing or exploding. By comparing VGG19 and ResNet50 and the proposed hybrid model, we concluded that the hybrid model had a significant advantage in accuracy. The detailed results of each model’s performance and their pros and cons will be presented in the conference.Keywords: acute lymphoblastic leukemia, hybrid model, leukemia diagnostic system, machine learning
Procedia PDF Downloads 1875466 Application of Fuzzy Clustering on Classification Agile Supply Chain Firms
Authors: Hamidreza Fallah Lajimi, Elham Karami, Alireza Arab, Fatemeh Alinasab
Abstract:
Being responsive is an increasingly important skill for firms in today’s global economy; thus firms must be agile. Naturally, it follows that an organization’s agility depends on its supply chain being agile. However, achieving supply chain agility is a function of other abilities within the organization. This paper analyses results from a survey of 71 Iran manufacturing companies in order to identify some of the factors for agile organizations in managing their supply chains. Then we classification this company in four cluster with fuzzy c-mean technique and with Four validations functional determine automatically the optimal number of clusters.Keywords: agile supply chain, clustering, fuzzy clustering, business engineering
Procedia PDF Downloads 7135465 Assessing the Quality of Clinical Photographs Taken for Orthodontic Patients at Queen’s Hospital, Romford
Authors: Maya Agarwala
Abstract:
Objectives: Audit the quality of clinical photographs taken for Orthodontic patients at Queen’s hospital, Romford. Design and setting: All Orthodontic photographs are taken in the Medical Photography Department at Queen’s Hospital. Retrospective audit with data collected between January - March 2023. Gold standard: Institute of Medical Illustrators (IMI) standard 12 photographs: 6 extraoral and 6 intraoral. 100% of patients to have the standard 12 photographs meeting a satisfactory diagnostic quality. Materials and methods: 30 patients randomly selected. All photographs analysed against the IMI gold standard. Results: A total of 360 photographs were analysed. 100% of the photographs had the 12 photographic views. Of which, 93.1% met the gold standard. Of the extraoral photos: 99.4% met the gold standard, 0.6% had incorrect head positioning. Of the intraoral photographs: 87.2% met the gold standard. The most common intraoral errors were: the presence of saliva pooling (7.2%), insufficient soft tissue retraction (3.3%), incomplete occlusal surface visibility (2.2%) and mirror fogging (1.1%). Conclusion: The gold standard was not met, however the overall standard of Orthodontic photographs is high. Further training of the Medical Photography team is needed to improve the quality of photographs. Following the training, the audit will be repeated. High-quality clinical photographs are an important part of clinical record keeping.Keywords: orthodontics, paediatric, photography, audit
Procedia PDF Downloads 965464 Ensemble of Deep CNN Architecture for Classifying the Source and Quality of Teff Cereal
Authors: Belayneh Matebie, Michael Melese
Abstract:
The study focuses on addressing the challenges in classifying and ensuring the quality of Eragrostis Teff, a small and round grain that is the smallest cereal grain. Employing a traditional classification method is challenging because of its small size and the similarity of its environmental characteristics. To overcome this, this study employs a machine learning approach to develop a source and quality classification system for Teff cereal. Data is collected from various production areas in the Amhara regions, considering two types of cereal (high and low quality) across eight classes. A total of 5,920 images are collected, with 740 images for each class. Image enhancement techniques, including scaling, data augmentation, histogram equalization, and noise removal, are applied to preprocess the data. Convolutional Neural Network (CNN) is then used to extract relevant features and reduce dimensionality. The dataset is split into 80% for training and 20% for testing. Different classifiers, including FVGG16, FINCV3, QSCTC, EMQSCTC, SVM, and RF, are employed for classification, achieving accuracy rates ranging from 86.91% to 97.72%. The ensemble of FVGG16, FINCV3, and QSCTC using the Max-Voting approach outperforms individual algorithms.Keywords: Teff, ensemble learning, max-voting, CNN, SVM, RF
Procedia PDF Downloads 555463 Wavelet-Based Classification of Myocardial Ischemia, Arrhythmia, Congestive Heart Failure and Sleep Apnea
Authors: Santanu Chattopadhyay, Gautam Sarkar, Arabinda Das
Abstract:
This paper presents wavelet based classification of various heart diseases. Electrocardiogram signals of different heart patients have been studied. Statistical natures of electrocardiogram signals for different heart diseases have been compared with the statistical nature of electrocardiograms for normal persons. Under this study four different heart diseases have been considered as follows: Myocardial Ischemia (MI), Congestive Heart Failure (CHF), Arrhythmia and Sleep Apnea. Statistical nature of electrocardiograms for each case has been considered in terms of kurtosis values of two types of wavelet coefficients: approximate and detail. Nine wavelet decomposition levels have been considered in each case. Kurtosis corresponding to both approximate and detail coefficients has been considered for decomposition level one to decomposition level nine. Based on significant difference, few decomposition levels have been chosen and then used for classification.Keywords: arrhythmia, congestive heart failure, discrete wavelet transform, electrocardiogram, myocardial ischemia, sleep apnea
Procedia PDF Downloads 1355462 An Experimental Study for Assessing Email Classification Attributes Using Feature Selection Methods
Authors: Issa Qabaja, Fadi Thabtah
Abstract:
Email phishing classification is one of the vital problems in the online security research domain that have attracted several scholars due to its impact on the users payments performed daily online. One aspect to reach a good performance by the detection algorithms in the email phishing problem is to identify the minimal set of features that significantly have an impact on raising the phishing detection rate. This paper investigate three known feature selection methods named Information Gain (IG), Chi-square and Correlation Features Set (CFS) on the email phishing problem to separate high influential features from low influential ones in phishing detection. We measure the degree of influentially by applying four data mining algorithms on a large set of features. We compare the accuracy of these algorithms on the complete features set before feature selection has been applied and after feature selection has been applied. After conducting experiments, the results show 12 common significant features have been chosen among the considered features by the feature selection methods. Further, the average detection accuracy derived by the data mining algorithms on the reduced 12-features set was very slight affected when compared with the one derived from the 47-features set.Keywords: data mining, email classification, phishing, online security
Procedia PDF Downloads 4325461 A Review and Classification of Maritime Disasters: The Case of Saudi Arabia's Coastline
Authors: Arif Almutairi, Monjur Mourshed
Abstract:
Due to varying geographical and tectonic factors, the region of Saudi Arabia has been subjected to numerous natural and man-made maritime disasters during the last two decades. Natural maritime disasters, such as cyclones and tsunamis, have been recorded in coastal areas of the Indian Ocean (including the Arabian Sea and the Gulf of Aden). Therefore, the Indian Ocean is widely recognised as the potential source of future destructive natural disasters that could affect Saudi Arabia’s coastline. Meanwhile, man-made maritime disasters, such as those arising from piracy and oil pollution, are located in the Red Sea and the Arabian Gulf, which are key locations for oil export and transportation between Asia and Europe. This paper provides a brief overview of maritime disasters surrounding Saudi Arabia’s coastline in order to classify them by frequency of occurrence and location, and discuss their future impact the region. Results show that the Arabian Gulf will be more vulnerable to natural maritime disasters because of its location, whereas the Red Sea is more vulnerable to man-made maritime disasters, as it is the key location for transportation between Asia and Europe. The results also show that with the aid of proper classification, effective disaster management can reduce the consequences of maritime disasters.Keywords: disaster classification, maritime disaster, natural disasters, man-made disasters
Procedia PDF Downloads 1895460 Application of Machine Learning Models to Predict Couchsurfers on Free Homestay Platform Couchsurfing
Authors: Yuanxiang Miao
Abstract:
Couchsurfing is a free homestay and social networking service accessible via the website and mobile app. Couchsurfers can directly request free accommodations from others and receive offers from each other. However, it is typically difficult for people to make a decision that accepts or declines a request when they receive it from Couchsurfers because they do not know each other at all. People are expected to meet up with some Couchsurfers who are kind, generous, and interesting while it is unavoidable to meet up with someone unfriendly. This paper utilized classification algorithms of Machine Learning to help people to find out the Good Couchsurfers and Not Good Couchsurfers on the Couchsurfing website. By knowing the prior experience, like Couchsurfer’s profiles, the latest references, and other factors, it became possible to recognize what kind of the Couchsurfers, and furthermore, it helps people to make a decision that whether to host the Couchsurfers or not. The value of this research lies in a case study in Kyoto, Japan in where the author has hosted 54 Couchsurfers, and the author collected relevant data from the 54 Couchsurfers, finally build a model based on classification algorithms for people to predict Couchsurfers. Lastly, the author offered some feasible suggestions for future research.Keywords: Couchsurfing, Couchsurfers prediction, classification algorithm, hospitality tourism platform, hospitality sciences, machine learning
Procedia PDF Downloads 1315459 Contribution to the Study of Automatic Epileptiform Pattern Recognition in Long Term EEG Signals
Authors: Christine F. Boos, Fernando M. Azevedo
Abstract:
Electroencephalogram (EEG) is a record of the electrical activity of the brain that has many applications, such as monitoring alertness, coma and brain death; locating damaged areas of the brain after head injury, stroke and tumor; monitoring anesthesia depth; researching physiology and sleep disorders; researching epilepsy and localizing the seizure focus. Epilepsy is a chronic condition, or a group of diseases of high prevalence, still poorly explained by science and whose diagnosis is still predominantly clinical. The EEG recording is considered an important test for epilepsy investigation and its visual analysis is very often applied for clinical confirmation of epilepsy diagnosis. Moreover, this EEG analysis can also be used to help define the types of epileptic syndrome, determine epileptiform zone, assist in the planning of drug treatment and provide additional information about the feasibility of surgical intervention. In the context of diagnosis confirmation the analysis is made using long term EEG recordings with at least 24 hours long and acquired by a minimum of 24 electrodes in which the neurophysiologists perform a thorough visual evaluation of EEG screens in search of specific electrographic patterns called epileptiform discharges. Considering that the EEG screens usually display 10 seconds of the recording, the neurophysiologist has to evaluate 360 screens per hour of EEG or a minimum of 8,640 screens per long term EEG recording. Analyzing thousands of EEG screens in search patterns that have a maximum duration of 200 ms is a very time consuming, complex and exhaustive task. Because of this, over the years several studies have proposed automated methodologies that could facilitate the neurophysiologists’ task of identifying epileptiform discharges and a large number of methodologies used neural networks for the pattern classification. One of the differences between all of these methodologies is the type of input stimuli presented to the networks, i.e., how the EEG signal is introduced in the network. Five types of input stimuli have been commonly found in literature: raw EEG signal, morphological descriptors (i.e. parameters related to the signal’s morphology), Fast Fourier Transform (FFT) spectrum, Short-Time Fourier Transform (STFT) spectrograms and Wavelet Transform features. This study evaluates the application of these five types of input stimuli and compares the classification results of neural networks that were implemented using each of these inputs. The performance of using raw signal varied between 43 and 84% efficiency. The results of FFT spectrum and STFT spectrograms were quite similar with average efficiency being 73 and 77%, respectively. The efficiency of Wavelet Transform features varied between 57 and 81% while the descriptors presented efficiency values between 62 and 93%. After simulations we could observe that the best results were achieved when either morphological descriptors or Wavelet features were used as input stimuli.Keywords: Artificial neural network, electroencephalogram signal, pattern recognition, signal processing
Procedia PDF Downloads 5285458 High Resolution Image Generation Algorithm for Archaeology Drawings
Authors: Xiaolin Zeng, Lei Cheng, Zhirong Li, Xueping Liu
Abstract:
Aiming at the problem of low accuracy and susceptibility to cultural relic diseases in the generation of high-resolution archaeology drawings by current image generation algorithms, an archaeology drawings generation algorithm based on a conditional generative adversarial network is proposed. An attention mechanism is added into the high-resolution image generation network as the backbone network, which enhances the line feature extraction capability and improves the accuracy of line drawing generation. A dual-branch parallel architecture consisting of two backbone networks is implemented, where the semantic translation branch extracts semantic features from orthophotographs of cultural relics, and the gradient screening branch extracts effective gradient features. Finally, the fusion fine-tuning module combines these two types of features to achieve the generation of high-quality and high-resolution archaeology drawings. Experimental results on the self-constructed archaeology drawings dataset of grotto temple statues show that the proposed algorithm outperforms current mainstream image generation algorithms in terms of pixel accuracy (PA), structural similarity (SSIM), and peak signal-to-noise ratio (PSNR) and can be used to assist in drawing archaeology drawings.Keywords: archaeology drawings, digital heritage, image generation, deep learning
Procedia PDF Downloads 595457 Leveraging Large Language Models to Build a Cutting-Edge French Word Sense Disambiguation Corpus
Authors: Mouheb Mehdoui, Amel Fraisse, Mounir Zrigui
Abstract:
With the increasing amount of data circulating over the Web, there is a growing need to develop and deploy tools aimed at unraveling semantic nuances within text or sentences. The challenges in extracting precise meanings arise from the complexity of natural language, while words usually have multiple interpretations depending on the context. The challenge of precisely interpreting words within a given context is what the task of Word Sense Disambiguation meets. It is a very old domain within the area of Natural Language Processing aimed at determining a word’s meaning that it is going to carry in a particular context, hence increasing the correctness of applications processing the language. Numerous linguistic resources are accessible online, including WordNet, thesauri, and dictionaries, enabling exploration of diverse contextual meanings. However, several limitations persist. These include the scarcity of resources for certain languages, a limited number of examples within corpora, and the challenge of accurately detecting the topic or context covered by text, which significantly impacts word sense disambiguation. This paper will discuss the different approaches to WSD and review corpora available for this task. We will contrast these approaches, highlighting the limitations, which will allow us to build a corpus in French, targeted for WSD.Keywords: semantic enrichment, disambiguation, context fusion, natural language processing, multilingual applications
Procedia PDF Downloads 105456 An Automated System for the Detection of Citrus Greening Disease Based on Visual Descriptors
Authors: Sidra Naeem, Ayesha Naeem, Sahar Rahim, Nadia Nawaz Qadri
Abstract:
Citrus greening is a bacterial disease that causes considerable damage to citrus fruits worldwide. Efficient method for this disease detection must be carried out to minimize the production loss. This paper presents a pattern recognition system that comprises three stages for the detection of citrus greening from Orange leaves: segmentation, feature extraction and classification. Image segmentation is accomplished by adaptive thresholding. The feature extraction stage comprises of three visual descriptors i.e. shape, color and texture. From shape feature we have used asymmetry index, from color feature we have used histogram of Cb component from YCbCr domain and from texture feature we have used local binary pattern. Classification was done using support vector machines and k nearest neighbors. The best performances of the system is Accuracy = 88.02% and AUROC = 90.1% was achieved by automatic segmented images. Our experiments validate that: (1). Segmentation is an imperative preprocessing step for computer assisted diagnosis of citrus greening, and (2). The combination of shape, color and texture features form a complementary set towards the identification of citrus greening disease.Keywords: citrus greening, pattern recognition, feature extraction, classification
Procedia PDF Downloads 1845455 A Qualitative Study to Analyze Clinical Coders’ Decision Making Process of Adverse Drug Event Admissions
Authors: Nisa Mohan
Abstract:
Clinical coding is a feasible method for estimating the national prevalence of adverse drug event (ADE) admissions. However, under-coding of ADE admissions is a limitation of this method. Whilst the under-coding will impact the accurate estimation of the actual burden of ADEs, the feasibility of the coded data in estimating the adverse drug event admissions goes much further compared to the other methods. Therefore, it is necessary to know the reasons for the under-coding in order to improve the clinical coding of ADE admissions. The ability to identify the reasons for the under-coding of ADE admissions rests on understanding the decision-making process of coding ADE admissions. Hence, the current study aimed to explore the decision-making process of clinical coders when coding cases of ADE admissions. Clinical coders from different levels of coding job such as trainee, intermediate and advanced level coders were purposefully selected for the interviews. Thirteen clinical coders were recruited from two Auckland region District Health Board hospitals for the interview study. Semi-structured, one-on-one, face-to-face interviews using open-ended questions were conducted with the selected clinical coders. Interviews were about 20 to 30 minutes long and were audio-recorded with the approval of the participants. The interview data were analysed using a general inductive approach. The interviews with the clinical coders revealed that the coders have targets to meet, and they sometimes hesitate to adhere to the coding standards. Coders deviate from the standard coding processes to make a decision. Coders avoid contacting the doctors for clarifying small doubts such as ADEs and the name of the medications because of the delay in getting a reply from the doctors. They prefer to do some research themselves or take help from their seniors and colleagues for making a decision because they can avoid a long wait to get a reply from the doctors. Coders think of ADE as a small thing. Lack of time for searching for information to confirm an ADE admission, inadequate communication with clinicians, along with coders’ belief that an ADE is a small thing may contribute to the under-coding of the ADE admissions. These findings suggest that further work is needed on interventions to improve the clinical coding of ADE admissions. Providing education to coders about the importance of ADEs, educating clinicians about the importance of clear and confirmed medical records entries, availing pharmacists’ services to improve the detection and clear documentation of ADE admissions, and including a mandatory field in the discharge summary about external causes of diseases may be useful for improving the clinical coding of ADE admissions. The findings of the research will help the policymakers to make informed decisions about the improvements. This study urges the coding policymakers, auditors, and trainers to engage with the unconscious cognitive biases and short-cuts of the clinical coders. This country-specific research conducted in New Zealand may also benefit other countries by providing insight into the clinical coding of ADE admissions and will offer guidance about where to focus changes and improvement initiatives.Keywords: adverse drug events, clinical coders, decision making, hospital admissions
Procedia PDF Downloads 1205454 A General Framework for Knowledge Discovery from Echocardiographic and Natural Images
Authors: S. Nandagopalan, N. Pradeep
Abstract:
The aim of this paper is to propose a general framework for storing, analyzing, and extracting knowledge from two-dimensional echocardiographic images, color Doppler images, non-medical images, and general data sets. A number of high performance data mining algorithms have been used to carry out this task. Our framework encompasses four layers namely physical storage, object identification, knowledge discovery, user level. Techniques such as active contour model to identify the cardiac chambers, pixel classification to segment the color Doppler echo image, universal model for image retrieval, Bayesian method for classification, parallel algorithms for image segmentation, etc., were employed. Using the feature vector database that have been efficiently constructed, one can perform various data mining tasks like clustering, classification, etc. with efficient algorithms along with image mining given a query image. All these facilities are included in the framework that is supported by state-of-the-art user interface (UI). The algorithms were tested with actual patient data and Coral image database and the results show that their performance is better than the results reported already.Keywords: active contour, Bayesian, echocardiographic image, feature vector
Procedia PDF Downloads 4455453 Translation Choices of Logical Meaning from Chinese into English: A Systemic Functional Linguistics Perspective
Authors: Xueying Li
Abstract:
Different from English, it is common to observe Chinese clauses logically related in an implicit way without any conjunctions. This typological difference has posed a great challenge for Chinese-English translators, as 1) translators may interpret logical meaning in different ways when there are no conjunctions in Chinese Source Text (ST); 2) translators may have questions whether to make Chinese implicit logical meaning explicit or to remain implicit in Target Text (TT), and whether other dimensions of logical meaning (e.g., type of logical meaning) should be shifted or not. Against this background, this study examines a comprehensive arrange of Chinese-English translation choices of logical meaning to deal with this challenge in a systematic way. It compiles several ST-TT passages from a set of translation textbooks in a corpus, namely Ying Yu Bi Yi Shi Wu (Er Ji)) [Translation Practice between Chinese and English: Intermediate Level] and its supportive training book, analyzes how logical meaning in ST are translated in TT in texts across different text types with Systemic Functional Linguistics (SFL) as the theoretical framework, and finally draws a system network of translation choices of logical meaning from Chinese into English. Since translators may probably think about semantic meaning rather than lexico-grammatical resources in translation, this study goes away from traditional lexico-grammatical choices, but rather describing translation choices from the semantic level. The findings in this study can provide some help and support for translation practitioners so that they can understand that besides explicitation, there are a variety of possible linguistic choices available for making informed decisions when translating Chinese logical meaning into English.Keywords: Chinese-English translation, logical meaning, systemic functional linguistics, translation choices
Procedia PDF Downloads 1805452 12-Week Comparative Clinical Trial with Low Dose Phentermine/Topiramate with Liraglutide on Obesity in Korea
Authors: Kyu Rae Lee
Abstract:
The aim of the study is to investigate the clinical efficacy of combination therapeutic modalities using liraglutide (1.2mg/d) add on low-dose phentermine (7.5 mg/d)/topiramate (50mg/d) medication on the obese patient in the bariatric clinic. We assessed the retrospective cohort clinical analyses to the clinical efficacy of medication and combination in the patients who visited the bariatric clinic. We measured all participants’ body fat (bioelectric impedance analysis), weight, height, and the cross-sectional areas of adipose tissues (umbilicus level) after keep fasting for 8 hours at 0, 4, 12 weeks. The design of the study was opened, paired t-test and Wilcoxon test were performed using SPSS for windows (ver.18, IL, USA) for comparison of weight, body fat, and adipose tissues. The participants were one hundred twenty-eight subjects aged 44.67 (1.18) years, 28.95 (0.39) kg/m², and female (82.7%). Their body fat was 40.57 (2.23%), and waist to hip ratio was 0.96 (0.01). The mean cross-sectional area of visceral adipose tissue was 142.59 (7.06) mm², and that of subcutaneous adipose was 274.37 (9.18) mm². 73 of them (57.5%) took medication only, 54 of them took medication with liraglutide for 12 weeks. The subjects in the medication group lost 5.4165 kg, 6.8069%, and those of the combination group did 6.2481 kg, 3.564%. The mean cross-sectional areas of visceral, subcutaneous adipose tissue in the medication group significantly decreased (p=.043), even more in the combination group. (p=.028). Further controlled clinical trials should be considered in the future. We conclude that the low dose of phentermine/topiramate with liraglutide therapeutic modalities would be more effective than phentermine/topiramate medication only in obesity treatment for 12 weeks.Keywords: low dose phentermine, topiramate, liraglutide, obesity, efficacy
Procedia PDF Downloads 158