Search results for: computer assisted classification
3879 Information Management Approach in the Prediction of Acute Appendicitis
Authors: Ahmad Shahin, Walid Moudani, Ali Bekraki
Abstract:
This research aims at presenting a predictive data mining model to handle an accurate diagnosis of acute appendicitis with patients for the purpose of maximizing the health service quality, minimizing morbidity/mortality, and reducing cost. However, acute appendicitis is the most common disease which requires timely accurate diagnosis and needs surgical intervention. Although the treatment of acute appendicitis is simple and straightforward, its diagnosis is still difficult because no single sign, symptom, laboratory or image examination accurately confirms the diagnosis of acute appendicitis in all cases. This contributes in increasing morbidity and negative appendectomy. In this study, the authors propose to generate an accurate model in prediction of patients with acute appendicitis which is based, firstly, on the segmentation technique associated to ABC algorithm to segment the patients; secondly, on applying fuzzy logic to process the massive volume of heterogeneous and noisy data (age, sex, fever, white blood cell, neutrophilia, CRP, urine, ultrasound, CT, appendectomy, etc.) in order to express knowledge and analyze the relationships among data in a comprehensive manner; and thirdly, on applying dynamic programming technique to reduce the number of data attributes. The proposed model is evaluated based on a set of benchmark techniques and even on a set of benchmark classification problems of osteoporosis, diabetes and heart obtained from the UCI data and other data sources.Keywords: healthcare management, acute appendicitis, data mining, classification, decision tree
Procedia PDF Downloads 3503878 Return to Work Rates of Participants in Medical Rehabilitation: The Role of Fitness and Health
Authors: Julius Steinkopf, Eric Rost, Aike Hessel, Sonia Lippke
Abstract:
Objective: This study examined possible determinants of return to work (RTW) in individuals who participated in a medical rehabilitation program longitudinally over a time period of six months. Design/methodology/approach: N=1,044 rehabilitants were included in the baseline measurement in terms of completing a questionnaire during their medical rehabilitation. About 30% (n=350) have remained in the study in terms of participating in computer-assisted telephone interviewing (CATI) six months later. Frequency analyses and Regression analyses were run. Findings: About 70% of the rehabilitants returned to work six months after rehabilitation. Regression analyses revealed that the RTW rates were significantly predicted by gender (OR=0.12, men were more likely to return), perceived social support (OR=3.01) and current physical functioning (OR=1.25). Furthermore RTW motives, like expected monetary rewards (OR=25.2) and feelings of being needed (OR=0.18) same as motives for not returning to work (nRTW), like the wish to stop working in order to spend time with the spouse (OR=0.13) or a lack of enjoyment of work (OR=3.81), significantly predicted return to work rates. Life satisfaction, self-efficacy beliefs, mental health, current income, educational background or age did not significantly increase explained variance (all ps > .05). Practical implications: Taking theses predictors into account provides options to increase the effectiveness of interventions aiming at increasing RTW: Medical rehabilitations should not only aim at improving the physical functioning but also to enhance beneficial motives and social support as well as support women specifically in order to improve the effectiveness of medical rehabilitation and public health interventions. Originality/value: Illness-caused work absences are related to high financial costs and individual burden. Despite of the public health and societal implications, this is one of the very few studies investigating systematically fitness and health for the return to work.Keywords: gender, fitness, health, physical functioning
Procedia PDF Downloads 2363877 Breast Cancer Survivability Prediction via Classifier Ensemble
Authors: Mohamed Al-Badrashiny, Abdelghani Bellaachia
Abstract:
This paper presents a classifier ensemble approach for predicting the survivability of the breast cancer patients using the latest database version of the Surveillance, Epidemiology, and End Results (SEER) Program of the National Cancer Institute. The system consists of two main components; features selection and classifier ensemble components. The features selection component divides the features in SEER database into four groups. After that it tries to find the most important features among the four groups that maximizes the weighted average F-score of a certain classification algorithm. The ensemble component uses three different classifiers, each of which models different set of features from SEER through the features selection module. On top of them, another classifier is used to give the final decision based on the output decisions and confidence scores from each of the underlying classifiers. Different classification algorithms have been examined; the best setup found is by using the decision tree, Bayesian network, and Na¨ıve Bayes algorithms for the underlying classifiers and Na¨ıve Bayes for the classifier ensemble step. The system outperforms all published systems to date when evaluated against the exact same data of SEER (period of 1973-2002). It gives 87.39% weighted average F-score compared to 85.82% and 81.34% of the other published systems. By increasing the data size to cover the whole database (period of 1973-2014), the overall weighted average F-score jumps to 92.4% on the held out unseen test set.Keywords: classifier ensemble, breast cancer survivability, data mining, SEER
Procedia PDF Downloads 3283876 Determination of the Effective Economic and/or Demographic Indicators in Classification of European Union Member and Candidate Countries Using Partial Least Squares Discriminant Analysis
Authors: Esra Polat
Abstract:
Partial Least Squares Discriminant Analysis (PLSDA) is a statistical method for classification and consists a classical Partial Least Squares Regression (PLSR) in which the dependent variable is a categorical one expressing the class membership of each observation. PLSDA can be applied in many cases when classical discriminant analysis cannot be applied. For example, when the number of observations is low and when the number of independent variables is high. When there are missing values, PLSDA can be applied on the data that is available. Finally, it is adapted when multicollinearity between independent variables is high. The aim of this study is to determine the economic and/or demographic indicators, which are effective in grouping the 28 European Union (EU) member countries and 7 candidate countries (including potential candidates Bosnia and Herzegovina (BiH) and Kosova) by using the data set obtained from database of the World Bank for 2014. Leaving the political issues aside, the analysis is only concerned with the economic and demographic variables that have the potential influence on country’s eligibility for EU entrance. Hence, in this study, both the performance of PLSDA method in classifying the countries correctly to their pre-defined groups (candidate or member) and the differences between the EU countries and candidate countries in terms of these indicators are analyzed. As a result of the PLSDA, the value of percentage correctness of 100 % indicates that overall of the 35 countries is classified correctly. Moreover, the most important variables that determine the statuses of member and candidate countries in terms of economic indicators are identified as 'external balance on goods and services (% GDP)', 'gross domestic savings (% GDP)' and 'gross national expenditure (% GDP)' that means for the 2014 economical structure of countries is the most important determinant of EU membership. Subsequently, the model validated to prove the predictive ability by using the data set for 2015. For prediction sample, %97,14 of the countries are correctly classified. An interesting result is obtained for only BiH, which is still a potential candidate for EU, predicted as a member of EU by using the indicators data set for 2015 as a prediction sample. Although BiH has made a significant transformation from a war-torn country to a semi-functional state, ethnic tensions, nationalistic rhetoric and political disagreements are still evident, which inhibit Bosnian progress towards the EU.Keywords: classification, demographic indicators, economic indicators, European Union, partial least squares discriminant analysis
Procedia PDF Downloads 2803875 Preparation of Nanocrystalline Mesoporous ThO2 Via Surfactant Assisted Sol-gel Procedure
Authors: N. Mohseni, S. Janitabar, S.J. Ahmadi, M. Roshanzamir, M. Thaghizadeh
Abstract:
There has been proposed a technique for getting thorium dioxide mesoporous nanocrystalline. In this paper thorium dioxide powder was synthesized through the sol-gel method using hydrated thorium nitrate and ammonium hydroxide as starting materials and Triton X100 as surfactant. ThO2 gel was characterized by thermogravimetric (TG), and prepared ThO2 powder was subjected to scanning electron microscopy (SEM), X-ray diffraction (XRD), and Brunauer-Emett-Teller (BET) analyses studies. Detailed analyses show that prepared powder consisted of phase with the space group Fm3m of thoria and its crystalline size was 27 nm. The thoria possesses 16.7 m2/g surface area and the pore volume and size calculated to be 0.0423 cc/g and 1.947 nm, respectively.Keywords: mesoporous, nanocrystalline, sol-gel, thoria
Procedia PDF Downloads 2813874 Detection and Classification Strabismus Using Convolutional Neural Network and Spatial Image Processing
Authors: Anoop T. R., Otman Basir, Robert F. Hess, Eileen E. Birch, Brooke A. Koritala, Reed M. Jost, Becky Luu, David Stager, Ben Thompson
Abstract:
Strabismus refers to a misalignment of the eyes. Early detection and treatment of strabismus in childhood can prevent the development of permanent vision loss due to abnormal development of visual brain areas. We developed a two-stage method for strabismus detection and classification based on photographs of the face. The first stage detects the presence or absence of strabismus, and the second stage classifies the type of strabismus. The first stage comprises face detection using Haar cascade, facial landmark estimation, face alignment, aligned face landmark detection, segmentation of the eye region, and detection of strabismus using VGG 16 convolution neural networks. Face alignment transforms the face to a canonical pose to ensure consistency in subsequent analysis. Using facial landmarks, the eye region is segmented from the aligned face and fed into a VGG 16 CNN model, which has been trained to classify strabismus. The CNN determines whether strabismus is present and classifies the type of strabismus (exotropia, esotropia, and vertical deviation). If stage 1 detects strabismus, the eye region image is fed into stage 2, which starts with the estimation of pupil center coordinates using mask R-CNN deep neural networks. Then, the distance between the pupil coordinates and eye landmarks is calculated along with the angle that the pupil coordinates make with the horizontal and vertical axis. The distance and angle information is used to characterize the degree and direction of the strabismic eye misalignment. This model was tested on 100 clinically labeled images of children with (n = 50) and without (n = 50) strabismus. The True Positive Rate (TPR) and False Positive Rate (FPR) of the first stage were 94% and 6% respectively. The classification stage has produced a TPR of 94.73%, 94.44%, and 100% for esotropia, exotropia, and vertical deviations, respectively. This method also had an FPR of 5.26%, 5.55%, and 0% for esotropia, exotropia, and vertical deviation, respectively. The addition of one more feature related to the location of corneal light reflections may reduce the FPR, which was primarily due to children with pseudo-strabismus (the appearance of strabismus due to a wide nasal bridge or skin folds on the nasal side of the eyes).Keywords: strabismus, deep neural networks, face detection, facial landmarks, face alignment, segmentation, VGG 16, mask R-CNN, pupil coordinates, angle deviation, horizontal and vertical deviation
Procedia PDF Downloads 933873 Deep Feature Augmentation with Generative Adversarial Networks for Class Imbalance Learning in Medical Images
Authors: Rongbo Shen, Jianhua Yao, Kezhou Yan, Kuan Tian, Cheng Jiang, Ke Zhou
Abstract:
This study proposes a generative adversarial networks (GAN) framework to perform synthetic sampling in feature space, i.e., feature augmentation, to address the class imbalance problem in medical image analysis. A feature extraction network is first trained to convert images into feature space. Then the GAN framework incorporates adversarial learning to train a feature generator for the minority class through playing a minimax game with a discriminator. The feature generator then generates features for minority class from arbitrary latent distributions to balance the data between the majority class and the minority class. Additionally, a data cleaning technique, i.e., Tomek link, is employed to clean up undesirable conflicting features introduced from the feature augmentation and thus establish well-defined class clusters for the training. The experiment section evaluates the proposed method on two medical image analysis tasks, i.e., mass classification on mammogram and cancer metastasis classification on histopathological images. Experimental results suggest that the proposed method obtains superior or comparable performance over the state-of-the-art counterparts. Compared to all counterparts, our proposed method improves more than 1.5 percentage of accuracy.Keywords: class imbalance, synthetic sampling, feature augmentation, generative adversarial networks, data cleaning
Procedia PDF Downloads 1273872 Classification of Emotions in Emergency Call Center Conversations
Authors: Magdalena Igras, Joanna Grzybowska, Mariusz Ziółko
Abstract:
The study of emotions expressed in emergency phone call is presented, covering both statistical analysis of emotions configurations and an attempt to automatically classify emotions. An emergency call is a situation usually accompanied by intense, authentic emotions. They influence (and may inhibit) the communication between caller and responder. In order to support responders in their responsible and psychically exhaustive work, we studied when and in which combinations emotions appeared in calls. A corpus of 45 hours of conversations (about 3300 calls) from emergency call center was collected. Each recording was manually tagged with labels of emotions valence (positive, negative or neutral), type (sadness, tiredness, anxiety, surprise, stress, anger, fury, calm, relief, compassion, satisfaction, amusement, joy) and arousal (weak, typical, varying, high) on the basis of perceptual judgment of two annotators. As we concluded, basic emotions tend to appear in specific configurations depending on the overall situational context and attitude of speaker. After performing statistical analysis we distinguished four main types of emotional behavior of callers: worry/helplessness (sadness, tiredness, compassion), alarm (anxiety, intense stress), mistake or neutral request for information (calm, surprise, sometimes with amusement) and pretension/insisting (anger, fury). The frequency of profiles was respectively: 51%, 21%, 18% and 8% of recordings. A model of presenting the complex emotional profiles on the two-dimensional (tension-insecurity) plane was introduced. In the stage of acoustic analysis, a set of prosodic parameters, as well as Mel-Frequency Cepstral Coefficients (MFCC) were used. Using these parameters, complex emotional states were modeled with machine learning techniques including Gaussian mixture models, decision trees and discriminant analysis. Results of classification with several methods will be presented and compared with the state of the art results obtained for classification of basic emotions. Future work will include optimization of the algorithm to perform in real time in order to track changes of emotions during a conversation.Keywords: acoustic analysis, complex emotions, emotion recognition, machine learning
Procedia PDF Downloads 3983871 Topics of Blockchain Technology to Teach at Community College
Authors: Penn P. Wu, Jeannie Jo
Abstract:
Blockchain technology has rapidly gained popularity in industry. This paper attempts to assist academia to answer four questions. First, should community colleges begin offering education to nurture blockchain-literate students for the job market? Second, what are the appropriate topical areas to cover? Third, should it be an individual course? And forth, should it be a technical or management course? This paper starts with identifying the knowledge domains of blockchain technology and the topical areas each domain has, and continues with placing them in appropriate academic territories (Computer Sciences vs. Business) and subjects (programming, management, marketing, and laws), and then develops an evaluation model to determine the appropriate topical area for community colleges to teach. The evaluation is based on seven factors: maturity of technology, impacts on management, real-world applications, subject classification, knowledge prerequisites, textbook readiness, and recommended pedagogies. The evaluation results point to an interesting direction that offering an introductory course is an ideal option to guide students through the learning journey of what blockchain is and how it applies to business. Such an introductory course does not need to engage students in the discussions of mathematics and sciences that make blockchain technologies possible. While it is inevitable to brief technical topics to help students build a solid knowledge foundation of blockchain technologies, community colleges should avoid offering students a course centered on the discussion of developing blockchain applications.Keywords: blockchain, pedagogies, blockchain technologies, blockchain course, blockchain pedagogies
Procedia PDF Downloads 1333870 Methodology for Temporary Analysis of Production and Logistic Systems on the Basis of Distance Data
Authors: M. Mueller, M. Kuehn, M. Voelker
Abstract:
In small and medium-sized enterprises (SMEs), the challenge is to create a well-grounded and reliable basis for process analysis, optimization and planning due to a lack of data. SMEs have limited access to methods with which they can effectively and efficiently analyse processes and identify cause-and-effect relationships in order to generate the necessary database and derive optimization potential from it. The implementation of digitalization within the framework of Industry 4.0 thus becomes a particular necessity for SMEs. For these reasons, the abstract presents an analysis methodology that is subject to the objective of developing an SME-appropriate methodology for efficient, temporarily feasible data collection and evaluation in flexible production and logistics systems as a basis for process analysis and optimization. The overall methodology focuses on retrospective, event-based tracing and analysis of material flow objects. The technological basis consists of Bluetooth low energy (BLE)-based transmitters, so-called beacons, and smart mobile devices (SMD), e.g. smartphones as receivers, between which distance data can be measured and derived motion profiles. The distance is determined using the Received Signal Strength Indicator (RSSI), which is a measure of signal field strength between transmitter and receiver. The focus is the development of a software-based methodology for interpretation of relative movements of transmitters and receivers based on distance data. The main research is on selection and implementation of pattern recognition methods for automatic process recognition as well as methods for the visualization of relative distance data. Due to an existing categorization of the database regarding process types, classification methods (e.g. Support Vector Machine) from the field of supervised learning are used. The necessary data quality requires selection of suitable methods as well as filters for smoothing occurring signal variations of the RSSI, the integration of methods for determination of correction factors depending on possible signal interference sources (columns, pallets) as well as the configuration of the used technology. The parameter settings on which respective algorithms are based have a further significant influence on result quality of the classification methods, correction models and methods for visualizing the position profiles used. The accuracy of classification algorithms can be improved up to 30% by selected parameter variation; this has already been proven in studies. Similar potentials can be observed with parameter variation of methods and filters for signal smoothing. Thus, there is increased interest in obtaining detailed results on the influence of parameter and factor combinations on data quality in this area. The overall methodology is realized with a modular software architecture consisting of independently modules for data acquisition, data preparation and data storage. The demonstrator for initialization and data acquisition is available as mobile Java-based application. The data preparation, including methods for signal smoothing, are Python-based with the possibility to vary parameter settings and to store them in the database (SQLite). The evaluation is divided into two separate software modules with database connection: the achievement of an automated assignment of defined process classes to distance data using selected classification algorithms and the visualization as well as reporting in terms of a graphical user interface (GUI).Keywords: event-based tracing, machine learning, process classification, parameter settings, RSSI, signal smoothing
Procedia PDF Downloads 1313869 Comparing Implications of Manual and ROSA-assisted Total Knee Replacements on Patients and Physicians: A Scoping Review
Authors: Bassem M. Darwish, Robert H. Ablove
Abstract:
Introduction: Total knee arthroscopy (TKA) is a commonly performed procedure in patients with end-stage osteoarthritis and inaccuracy of component alignment in TKA has been shown to have many adverse post-operative outcomes such as accelerated implant wear, reduced functional outcomes, and shorter overall implant survival. Robotic surgical systems have been introduced to try and improve joint alignment and functional outcomes in knee arthroscopy, one recent iteration is the ROSA knee system, released to the market in 2019. The objective of this scoping review is to map the available evidence, identify the current types of evidence, and identify knowledge gaps to guide future studies on patient outcomes following ROSA-assisted total knee arthroplasties. Methods: An electronic search was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) extension for scoping reviews. Search terms included ROSA, knee arthroscopy, osteoarthritis, robotic, and malalignment. Types of study participants included patients with osteoarthritis, ages 18 and older, male or female, who received manual TKA (mTKA) or ROSA-assisted TKA (rTKA), and human patients or cadavers. Published, peer-reviewed controlled trials, observational studies, and case series were included. Case reports were not included in article review. Resulting articles were first screened based on title and abstract. Articles meeting inclusion criteria based on title and abstract review then underwent full-text review by the same reviewer. Results: This scoping review identified 11 total studies, 3 prospective observational studies, and 8 retrospective observational studies - a total of 970 rTKA patients and 1745 mTKA patients. There were no case series or randomized controlled trials comparing rTKA and mTKA. Patient-centered outcomes showed promise for rTKA, where it frequently showed significantly favorable functional outcomes, measured via KOOS-JR, VAS, KSS, OKS, FJS, and PROMIS scores, at various times postoperatively. However, there was much discrepancy about which score yielded significance at which postoperative follow-up. Complication rates, reoperation rates, and LOS were very similar between mTKA and rTKA groups. Studies also showed rTKA had more accurate joint alignment within the 0 ± 3o corridor and had significantly higher rates of achieving postoperative joint angles similar to the preoperative plan. Finally, there was major agreement that rTKA cases take significantly longer time at the start, however, there is a rapid learning curve. Once past the learning curve, rTKA cases are performed in a similar time to mTKA and reduced physician stress and strain. Conclusion: The ROSA knee system represents a promising option for the management of osteoarthritis via total knee arthroscopy. The studies reviewed in this paper favor the patient-centered function outcomes, joint alignments, and physician health implications of the ROSA knee system to conventional total knee arthroscopy. Further study is warranted, however, to better understand recovery periods, longer-term functional outcomes, operative fatigue, and reduction in radiation exposure.Keywords: arthroplasty, knee, robotics, malalignment
Procedia PDF Downloads 293868 Vineyard Soils of Karnataka - Characterization, Classification and Soil Site Suitability Evaluation
Authors: Harsha B. R., K. S. Anil Kumar
Abstract:
Land characterization, classification, and soil suitability evaluation of grapes-growing pedons were assessed at fifteen taluks covering four agro climatic zones of Karnataka. Study on problems and potentials of grapes cultivation in selected agro-climatic zones was carried out along with the plant sample analysis. Twenty soil profiles were excavated as study site based on the dominance of area falling under grapes production and existing spatial variability of soils. The detailed information of profiles and horizon wise soil samples were collected to study the morphological, physical, chemical, and fertility characteristics. Climatic analysis and water retention characteristics of soils of major grapes-growing areas were also done. Based on the characterisation and classification study, it was revealed that soils of Doddaballapur (Bangalore Blue and Wine grapes), Bangalore North (GKVK Farm, Rajankunte, and IIHR Farm), Devanahalli, Magadi, Hoskote, Chikkaballapur (Dilkush and Red globe), Yelaburga, Hagari Bommanahalli, Bagalkot (UHS farm) and Indi fall under the soil order Alfisol. Vijaypur pedon of northern dry zone was keyed out as Vertisols whereas, Jamkhandi and Athani as Inceptisols. Properties of Aridisols were observed in B. Bagewadi (Manikchaman and Thompson Seedless) and Afzalpur. Soil fertility status and its mapping using GIS technique revealed that all the nutrients were found to be in adequate range except nitrogen, potassium, zinc, iron, and boron, which indicated the need for application along with organic matter to improve the SOC status. Varieties differed among themselves in yield and plant nutrient composition depending on their age, climatic, soil, and management requirements. Bangalore North (GKVK farm) and Jamkhandi are having medium soil organic carbon stocks of 6.21 and 6.55 kg m⁻³, respectively. Soils of Bangalore North (Rajankunte) were highly suitable (S1) for grapes cultivation. Under northern Karnataka, Vijayapura, B. Bagewadi, Indi, and Afzalpur vineyards were good performers despite the limitations of fertility and free lime content.Keywords: land characterization, suitability, soil orders, soil organic carbon stock
Procedia PDF Downloads 1133867 The Application of Video Segmentation Methods for the Purpose of Action Detection in Videos
Authors: Nassima Noufail, Sara Bouhali
Abstract:
In this work, we develop a semi-supervised solution for the purpose of action detection in videos and propose an efficient algorithm for video segmentation. The approach is divided into video segmentation, feature extraction, and classification. In the first part, a video is segmented into clips, and we used the K-means algorithm for this segmentation; our goal is to find groups based on similarity in the video. The application of k-means clustering into all the frames is time-consuming; therefore, we started by the identification of transition frames where the scene in the video changes significantly, and then we applied K-means clustering into these transition frames. We used two image filters, the gaussian filter and the Laplacian of Gaussian. Each filter extracts a set of features from the frames. The Gaussian filter blurs the image and omits the higher frequencies, and the Laplacian of gaussian detects regions of rapid intensity changes; we then used this vector of filter responses as an input to our k-means algorithm. The output is a set of cluster centers. Each video frame pixel is then mapped to the nearest cluster center and painted with a corresponding color to form a visual map. The resulting visual map had similar pixels grouped. We then computed a cluster score indicating how clusters are near each other and plotted a signal representing frame number vs. clustering score. Our hypothesis was that the evolution of the signal would not change if semantically related events were happening in the scene. We marked the breakpoints at which the root mean square level of the signal changes significantly, and each breakpoint is an indication of the beginning of a new video segment. In the second part, for each segment from part 1, we randomly selected a 16-frame clip, then we extracted spatiotemporal features using convolutional 3D network C3D for every 16 frames using a pre-trained model. The C3D final output is a 512-feature vector dimension; hence we used principal component analysis (PCA) for dimensionality reduction. The final part is the classification. The C3D feature vectors are used as input to a multi-class linear support vector machine (SVM) for the training model, and we used a multi-classifier to detect the action. We evaluated our experiment on the UCF101 dataset, which consists of 101 human action categories, and we achieved an accuracy that outperforms the state of art by 1.2%.Keywords: video segmentation, action detection, classification, Kmeans, C3D
Procedia PDF Downloads 773866 Computational Fluid Dynamics Simulations of Thermal and Flow Fields inside a Desktop Personal Computer Cabin
Authors: Mohammad Salehi, Mohammad Erfan Doraki
Abstract:
In this paper, airflow analysis inside a desktop computer case is performed by simulating computational fluid dynamics. The purpose is to investigate the cooling process of the central processing unit (CPU) with thermal capacities of 80 and 130 watts. The airflow inside the computer enclosure, selected from the microATX model, consists of the main components of heat production such as CPU, hard disk drive, CD drive, floppy drive, memory card and power supply unit; According to the amount of thermal power produced by the CPU with 80 and 130 watts of power, two different geometries have been used for a direct and radial heat sink. First, the independence of the computational mesh and the validation of the solution were performed, and after ensuring the correctness of the numerical solution, the results of the solution were analyzed. The simulation results showed that changes in CPU temperature and other components linearly increased with increasing CPU heat output. Also, the ambient air temperature has a significant effect on the maximum processor temperature.Keywords: computational fluid dynamics, CPU cooling, computer case simulation, heat sink
Procedia PDF Downloads 1223865 Detection of Internal Mold Infection of Intact Tomatoes by Non-Destructive, Transmittance VIS-NIR Spectroscopy
Authors: K. Petcharaporn
Abstract:
The external characteristics of tomatoes, such as freshness, color and size are typically used in quality control processes for tomatoes sorting. However, the internal mold infection of intact tomato cannot be sorted based on a visible assessment and destructive method alone. In this study, a non-destructive technique was used to predict the internal mold infection of intact tomatoes by using transmittance visible and near infrared (VIS-NIR) spectroscopy. Spectra for 200 samples contained 100 samples for normal tomatoes and 100 samples for mold infected tomatoes were acquired in the wavelength range between 665-955 nm. This data was used in conjunction with partial least squares-discriminant analysis (PLS-DA) method to generate a classification model for tomato quality between groups of internal mold infection of intact tomato samples. For this task, the data was split into two groups, 140 samples were used for a training set and 60 samples were used for a test set. The spectra of both normal and internally mold infected tomatoes showed different features in the visible wavelength range. Combined spectral pretreatments of standard normal variate transformation (SNV) and smoothing (Savitzky-Golay) gave the optimal calibration model in training set, 85.0% (63 out of 71 for the normal samples and 56 out of 69 for the internal mold samples). The classification accuracy of the best model on the test set was 91.7% (29 out of 29 for the normal samples and 26 out of 31 for the internal mold tomato samples). The results from this experiment showed that transmittance VIS-NIR spectroscopy can be used as a non-destructive technique to predict the internal mold infection of intact tomatoes.Keywords: tomato, mold, quality, prediction, transmittance
Procedia PDF Downloads 3633864 A Supervised Approach for Detection of Singleton Spam Reviews
Authors: Atefeh Heydari, Mohammadali Tavakoli, Naomie Salim
Abstract:
In recent years, we have witnessed that online reviews are the most important source of customers’ opinion. They are progressively more used by individuals and organisations to make purchase and business decisions. Unfortunately, for the reason of profit or fame, frauds produce deceptive reviews to hoodwink potential customers. Their activities mislead not only potential customers to make appropriate purchasing decisions and organisations to reshape their business, but also opinion mining techniques by preventing them from reaching accurate results. Spam reviews could be divided into two main groups, i.e. multiple and singleton spam reviews. Detecting a singleton spam review that is the only review written by a user ID is extremely challenging due to lack of clue for detection purposes. Singleton spam reviews are very harmful and various features and proofs used in multiple spam reviews detection are not applicable in this case. Current research aims to propose a novel supervised technique to detect singleton spam reviews. To achieve this, various features are proposed in this study and are to be combined with the most appropriate features extracted from literature and employed in a classifier. In order to compare the performance of different classifiers, SVM and naive Bayes classification algorithms were used for model building. The results revealed that SVM was more accurate than naive Bayes and our proposed technique is capable to detect singleton spam reviews effectively.Keywords: classification algorithms, Naïve Bayes, opinion review spam detection, singleton review spam detection, support vector machine
Procedia PDF Downloads 3093863 Reconstructability Analysis for Landslide Prediction
Authors: David Percy
Abstract:
Landslides are a geologic phenomenon that affects a large number of inhabited places and are constantly being monitored and studied for the prediction of future occurrences. Reconstructability analysis (RA) is a methodology for extracting informative models from large volumes of data that work exclusively with discrete data. While RA has been used in medical applications and social science extensively, we are introducing it to the spatial sciences through applications like landslide prediction. Since RA works exclusively with discrete data, such as soil classification or bedrock type, working with continuous data, such as porosity, requires that these data are binned for inclusion in the model. RA constructs models of the data which pick out the most informative elements, independent variables (IVs), from each layer that predict the dependent variable (DV), landslide occurrence. Each layer included in the model retains its classification data as a primary encoding of the data. Unlike other machine learning algorithms that force the data into one-hot encoding type of schemes, RA works directly with the data as it is encoded, with the exception of continuous data, which must be binned. The usual physical and derived layers are included in the model, and testing our results against other published methodologies, such as neural networks, yields accuracy that is similar but with the advantage of a completely transparent model. The results of an RA session with a data set are a report on every combination of variables and their probability of landslide events occurring. In this way, every combination of informative state combinations can be examined.Keywords: reconstructability analysis, machine learning, landslides, raster analysis
Procedia PDF Downloads 663862 Detection of Internal Mold Infection of Intact For Tomatoes by Non-Destructive, Transmittance VIS-NIR Spectroscopy
Authors: K. Petcharaporn, N. Prathengjit
Abstract:
The external characteristics of tomatoes, such as freshness, color and size are typically used in quality control processes for tomatoes sorting. However, the internal mold infection of intact tomato cannot be sorted based on a visible assessment and destructive method alone. In this study, a non-destructive technique was used to predict the internal mold infection of intact tomatoes by using transmittance visible and near infrared (VIS-NIR) spectroscopy. Spectra for 200 samples contained 100 samples for normal tomatoes and 100 samples for mold infected tomatoes were acquired in the wavelength range between 665-955 nm. This data was used in conjunction with partial least squares-discriminant analysis (PLS-DA) method to generate a classification model for tomato quality between groups of internal mold infection of intact tomato samples. For this task, the data was split into two groups, 140 samples were used for a training set and 60 samples were used for a test set. The spectra of both normal and internally mold infected tomatoes showed different features in the visible wavelength range. Combined spectral pretreatments of standard normal variate transformation (SNV) and smoothing (Savitzky-Golay) gave the optimal calibration model in training set, 85.0% (63 out of 71 for the normal samples and 56 out of 69 for the internal mold samples). The classification accuracy of the best model on the test set was 91.7% (29 out of 29 for the normal samples and 26 out of 31 for the internal mold tomato samples). The results from this experiment showed that transmittance VIS-NIR spectroscopy can be used as a non-destructive technique to predict the internal mold infection of intact tomatoes.Keywords: tomato, mold, quality, prediction, transmittance
Procedia PDF Downloads 5193861 Implementation of the Outputs of Computer Simulation to Support Decision-Making Processes
Authors: Jiri Barta
Abstract:
At the present time, awareness, education, computer simulation and information systems protection are very serious and relevant topics. The article deals with perspectives and possibilities of implementation of emergence or natural hazard threats into the system which is developed for communication among members of crisis management staffs. The Czech Hydro-Meteorological Institute with its System of Integrated Warning Service resents the largest usable base of information. National information systems are connected to foreign systems, especially to flooding emergency systems of neighboring countries, systems of European Union and international organizations where the Czech Republic is a member. Use of outputs of particular information systems and computer simulations on a single communication interface of information system for communication among members of crisis management staff and setting the site interoperability in the net will lead to time savings in decision-making processes in solving extraordinary events and crisis situations. Faster managing of an extraordinary event or a crisis situation will bring positive effects and minimize the impact of negative effects on the environment.Keywords: computer simulation, communication, continuity, critical infrastructure, information systems, safety
Procedia PDF Downloads 3333860 Change Detection of Vegetative Areas Using Land Use Land Cover of Desertification Vulnerable Areas in Nigeria
Authors: T. Garba, Y. Y. Sabo A. Babanyara, K. G. Ilellah, A. K. Mutari
Abstract:
This study used the Normalized Difference Vegetation Index (NDVI) and maps compiled from the classification of Landsat TM and Landsat ETM images of 1986 and 1999 respectively and Nigeria sat 1 images of 2007 to quantify changes in land use and land cover in selected areas of Nigeria covering 143,609 hectares that are threatened by the encroaching Sahara desert. The results of this investigation revealed a decrease in natural vegetation over the three time slices (1986, 1999 and 2007) which was characterised by an increase in high positive pixel values from 0.04 in 1986 to 0.22 and 0.32 in 1999 and 2007 respectively and, a decrease in natural vegetation from 74,411.60ha in 1986 to 28,591.93ha and 21,819.19ha in 1999 and 2007 respectively. The same results also revealed a periodic trend in which there was progressive increase in the cultivated area from 60,191.87ha in 1986 to 104,376.07ha in 1999 and a terminal decrease to 88,868.31ha in 2007. These findings point to expansion of vegetated and cultivated areas in in the initial period between 1988 and 1996 and reversal of these increases in the terminal period between 1988 and 1996. The study also revealed progressive expansion of built-up areas from 1, 681.68ha in 1986 to 2,661.82ha in 1999 and to 3,765.35ha in 2007. These results argue for the urgent need to protect and conserve the depleting natural vegetation by adopting sustainable human resource use practices i.e. intensive farming in order to minimize persistent depletion of natural vegetation.Keywords: changes, classification, desertification, vegetation changes
Procedia PDF Downloads 3873859 Diagnosis of the Heart Rhythm Disorders by Using Hybrid Classifiers
Authors: Sule Yucelbas, Gulay Tezel, Cuneyt Yucelbas, Seral Ozsen
Abstract:
In this study, it was tried to identify some heart rhythm disorders by electrocardiography (ECG) data that is taken from MIT-BIH arrhythmia database by subtracting the required features, presenting to artificial neural networks (ANN), artificial immune systems (AIS), artificial neural network based on artificial immune system (AIS-ANN) and particle swarm optimization based artificial neural network (PSO-NN) classifier systems. The main purpose of this study is to evaluate the performance of hybrid AIS-ANN and PSO-ANN classifiers with regard to the ANN and AIS. For this purpose, the normal sinus rhythm (NSR), atrial premature contraction (APC), sinus arrhythmia (SA), ventricular trigeminy (VTI), ventricular tachycardia (VTK) and atrial fibrillation (AF) data for each of the RR intervals were found. Then these data in the form of pairs (NSR-APC, NSR-SA, NSR-VTI, NSR-VTK and NSR-AF) is created by combining discrete wavelet transform which is applied to each of these two groups of data and two different data sets with 9 and 27 features were obtained from each of them after data reduction. Afterwards, the data randomly was firstly mixed within themselves, and then 4-fold cross validation method was applied to create the training and testing data. The training and testing accuracy rates and training time are compared with each other. As a result, performances of the hybrid classification systems, AIS-ANN and PSO-ANN were seen to be close to the performance of the ANN system. Also, the results of the hybrid systems were much better than AIS, too. However, ANN had much shorter period of training time than other systems. In terms of training times, ANN was followed by PSO-ANN, AIS-ANN and AIS systems respectively. Also, the features that extracted from the data affected the classification results significantly.Keywords: AIS, ANN, ECG, hybrid classifiers, PSO
Procedia PDF Downloads 4423858 Life Stage Customer Segmentation by Fine-Tuning Large Language Models
Authors: Nikita Katyal, Shaurya Uppal
Abstract:
This paper tackles the significant challenge of accurately classifying customers within a retailer’s customer base. Accurate classification is essential for developing targeted marketing strategies that effectively engage this important demographic. To address this issue, we propose a method that utilizes Large Language Models (LLMs). By employing LLMs, we analyze the metadata associated with product purchases derived from historical data to identify key product categories that act as distinguishing factors. These categories, such as baby food, eldercare products, or family-sized packages, offer valuable insights into the likely household composition of customers, including families with babies, families with kids/teenagers, families with pets, households caring for elders, or mixed households. We segment high-confidence customers into distinct categories by integrating historical purchase behavior with LLM-powered product classification. This paper asserts that life stage segmentation can significantly enhance e-commerce businesses’ ability to target the appropriate customers with tailored products and campaigns, thereby augmenting sales and improving customer retention. Additionally, the paper details the data sources, model architecture, and evaluation metrics employed for the segmentation task.Keywords: LLMs, segmentation, product tags, fine-tuning, target segments, marketing communication
Procedia PDF Downloads 233857 Segmentation of Liver Using Random Forest Classifier
Authors: Gajendra Kumar Mourya, Dinesh Bhatia, Akash Handique, Sunita Warjri, Syed Achaab Amir
Abstract:
Nowadays, Medical imaging has become an integral part of modern healthcare. Abdominal CT images are an invaluable mean for abdominal organ investigation and have been widely studied in the recent years. Diagnosis of liver pathologies is one of the major areas of current interests in the field of medical image processing and is still an open problem. To deeply study and diagnose the liver, segmentation of liver is done to identify which part of the liver is mostly affected. Manual segmentation of the liver in CT images is time-consuming and suffers from inter- and intra-observer differences. However, automatic or semi-automatic computer aided segmentation of the Liver is a challenging task due to inter-patient Liver shape and size variability. In this paper, we present a technique for automatic segmenting the liver from CT images using Random Forest Classifier. Random forests or random decision forests are an ensemble learning method for classification that operate by constructing a multitude of decision trees at training time and outputting the class that is the mode of the classes of the individual trees. After comparing with various other techniques, it was found that Random Forest Classifier provide a better segmentation results with respect to accuracy and speed. We have done the validation of our results using various techniques and it shows above 89% accuracy in all the cases.Keywords: CT images, image validation, random forest, segmentation
Procedia PDF Downloads 3133856 The Classification Accuracy of Finance Data through Holder Functions
Authors: Yeliz Karaca, Carlo Cattani
Abstract:
This study focuses on the local Holder exponent as a measure of the function regularity for time series related to finance data. In this study, the attributes of the finance dataset belonging to 13 countries (India, China, Japan, Sweden, France, Germany, Italy, Australia, Mexico, United Kingdom, Argentina, Brazil, USA) located in 5 different continents (Asia, Europe, Australia, North America and South America) have been examined.These countries are the ones mostly affected by the attributes with regard to financial development, covering a period from 2012 to 2017. Our study is concerned with the most important attributes that have impact on the development of finance for the countries identified. Our method is comprised of the following stages: (a) among the multi fractal methods and Brownian motion Holder regularity functions (polynomial, exponential), significant and self-similar attributes have been identified (b) The significant and self-similar attributes have been applied to the Artificial Neuronal Network (ANN) algorithms (Feed Forward Back Propagation (FFBP) and Cascade Forward Back Propagation (CFBP)) (c) the outcomes of classification accuracy have been compared concerning the attributes that have impact on the attributes which affect the countries’ financial development. This study has enabled to reveal, through the application of ANN algorithms, how the most significant attributes are identified within the relevant dataset via the Holder functions (polynomial and exponential function).Keywords: artificial neural networks, finance data, Holder regularity, multifractals
Procedia PDF Downloads 2463855 Differences in Assessing Hand-Written and Typed Student Exams: A Corpus-Linguistic Study
Authors: Jutta Ransmayr
Abstract:
The digital age has long arrived at Austrian schools, so both society and educationalists demand that digital means should be integrated accordingly to day-to-day school routines. Therefore, the Austrian school-leaving exam (A-levels) can now be written either by hand or by using a computer. However, the choice of writing medium (pen and paper or computer) for written examination papers, which are considered 'high-stakes' exams, raises a number of questions that have not yet been adequately investigated and answered until recently, such as: What effects do the different conditions of text production in the written German A-levels have on the component of normative linguistic accuracy? How do the spelling skills of German A-level papers written with a pen differ from those that the students wrote on the computer? And how is the teacher's assessment related to this? Which practical desiderata for German didactics can be derived from this? In a trilateral pilot project of the Austrian Center for Digital Humanities (ACDH) of the Austrian Academy of Sciences and the University of Vienna in cooperation with the Austrian Ministry of Education and the Council for German Orthography, these questions were investigated. A representative Austrian learner corpus, consisting of around 530 German A-level papers from all over Austria (pen and computer written), was set up in order to subject it to a quantitative (corpus-linguistic and statistical) and qualitative investigation with regard to the spelling and punctuation performance of the high school graduates and the differences between pen- and computer-written papers and their assessments. Relevant studies are currently available mainly from the Anglophone world. These have shown that writing on the computer increases the motivation to write, has positive effects on the length of the text, and, in some cases, also on the quality of the text. Depending on the writing situation and other technical aids, better results in terms of spelling and punctuation could also be found in the computer-written texts as compared to the handwritten ones. Studies also point towards a tendency among teachers to rate handwritten texts better than computer-written texts. In this paper, the first comparable results from the German-speaking area are to be presented. Research results have shown that, on the one hand, there are significant differences between handwritten and computer-written work with regard to performance in orthography and punctuation. On the other hand, the corpus linguistic investigation and the subsequent statistical analysis made it clear that not only the teachers' assessments of the students’ spelling performance vary enormously but also the overall assessments of the exam papers – the factor of the production medium (pen and paper or computer) also seems to play a decisive role.Keywords: exam paper assessment, pen and paper or computer, learner corpora, linguistics
Procedia PDF Downloads 1703854 Climate Change and Food Security in Nigeria: The World Bank Assisted Third National Fadama Development Programme (Nfdp Iii) Approach in Rivers State, Niger Delta, Nigeria
Authors: Temple Probyne Abali
Abstract:
Port Harcourt, Rivers State in the Niger Delta region of Nigeria is bedeviled by the phenomenon of climatechange, posing threat to food security and livelihood. This study examined a 4 decadel (1980-2020) trend of climate change as well as its socio-economic impact on food security in the region. Furthermore, to achieve sustainable food security and livelihood amidst the phenomenon, the study adopted the World Bank Assisted Third National Fadama Development Programme approach. The data source for climate change involved secondary data from Nigeria Meteorological Agency (NIMET). Consequently, the results for climate change over the 4decade period were displayed in tables, charts and maps for the expected changes. Data sources on socio-economic impact of food security and livelihood were acquired through questionnairedesign. A purposive random sampling technique was used in selecting 5 coastal communities inthe region known for viable economic potentials for agricultural development and the resultswere analyzed using Analysis of Variance (ANOVA). The Participatory Rural Appraisal (PRA) technique of the World Bank for needs assessment wasadopted in selecting 5 agricultural sub-project proposals/activities based on groups’ commoneconomic interest from a total of 1,000 farmers each drawn from the 5 communities of differentage groups including men, women, youths and the vulnerable. Based on the farmers’ sub-projectinterests, the various groups’ Strength, Weakness, Opportunities and Threats (SWOT), Problem Listing Matrix, Skill Gap Analysis as well as EIAson their sub-project proposals/activities were analyzed with substantialMonitoring and Evaluation (M & E), using the Specific, Measurable, Attribute, Reliable and Time bound (SMART)approach. Based on the findings from the PRA technique, the farmers recorded considerableincreaseinincomeofover200%withinthe5yearprojectplan(2008-2013).Thestudyrecommends capacity building and advisory services on this PRA innovation. By so doing, there would be a sustainable increase in agricultural production and assured food security in an environmental friendly manner, in line with the United Nation’s Sustainable Development Goals(SDGs).Keywords: climate change, food security, fadama, world bank, agriculture, sdgs
Procedia PDF Downloads 933853 Drugstore Control System Design and Realization Based on Programmable Logic Controller (PLC)
Authors: Muhammad Faheem Khakhi, Jian Yu Wang, Salman Muhammad, Muhammad Faisal Shabir
Abstract:
Population growth and Chinese two-child policy will boost pharmaceutical market, and it will continue to maintain the growth for a period of time in the future, the traditional pharmacy dispensary has been unable to meet the growing medical needs of the peoples. Under the strong support of the national policy, the automatic transformation of traditional pharmacies is the inclination of the Times, the new type of intelligent pharmacy system will continue to promote the development of the pharmaceutical industry. Under this background, based on PLC control, the paper proposed an intelligent storage and automatic drug delivery system; complete design of the lower computer's control system and the host computer's software system has been present. The system can be applied to dispensing work for Chinese herbal medicinal and Western medicines. Firstly, the essential of intelligent control system for pharmacy is discussed. After the analysis of the requirements, the overall scheme of the system design is presented. Secondly, introduces the software and hardware design of the lower computer's control system, including the selection of PLC and the selection of motion control system, the problem of the human-computer interaction module and the communication between PC and PLC solves, the program design and development of the PLC control system is completed. The design of the upper computer software management system is described in detail. By analyzing of E-R diagram, built the establish data, the communication protocol between systems is customize, C++ Builder is adopted to realize interface module, supply module, main control module, etc. The paper also gives the implementations of the multi-threaded system and communication method. Lastly, each module of the lower computer control system is tested. Then, after building a test environment, the function test of the upper computer software management system is completed. On this basis, the entire control system accepts the overall test.Keywords: automatic pharmacy, PLC, control system, management system, communication
Procedia PDF Downloads 3103852 Surrogacy in India: Emerging Business or Disguised Human Trafficking
Authors: Priya Sepaha
Abstract:
Commercial Surrogacy refers to a contract in which a woman carries a pregnancy for intended parents. There are two types of surrogacy; first, Traditional Surrogacy, in which, sperm of the donor or father is artificially inseminated in the women and carries the fetus till birth. Second, Gestational Surrogacy, in which the egg and sperm of the intended parent are collected for artificial fertilization through In Vitro Fertilization (IVF) technique and after the embryo formation, it is transferred into the womb of a surrogate mother with the help of Assisted Reproductive Technique. Surrogacy has become so widespread in India that it has now been nicknamed the "rent-a-womb" capital of the world due to relatively low cost and lack of stringent regulatory legalisation. The legal aspects surrounding surrogacy are complex, diverse and mostly unsettled. Although this appears to be beneficial for the parties concerned, there are certain sensitive issues which need to be addressed to ensure ample protection to all stakeholders. Commercial surrogacy is an emerging business and a new means of human trafficking particularly in India. Poor and illiterate women are often lured in such deals by their spouse or broker for earning easy money. Traffickers also use force, fraud, or coercion at times to intimidate the probable surrogate mothers. A major chunk of money received from covert surrogacy agreement is taken away by the brokers. The Law Commission of India has specifically reviewed the issue as India is emerging as a major global surrogacy destination. The Supreme Court of India held in the Manji's case in 2008, that commercial surrogacy can be permitted with certain restrictions but had directed the Legislature to pass an appropriate Law for governing Surrogacy in India. The draft Assisted Reproductive Technique (ART) Bill, 2010 is still pending for approval. At present, the Surrogacy Contract between the parties and the ART Clinics Guidelines are perhaps the only guiding force. The Immoral Trafficking Prevention Act (ITPA), 1956 and Sections 366(A) and 372 of the Indian Penal Code, 1860 are perhaps the only existing laws, which deal with human trafficking. Yet, none of these provisions specifically deal with the serious issue of trafficking for the purpose of Commercial Surrogacy. India remains one of the few countries that still allow commercial surrogacy. International Surrogacy involves bilateral issues, where the laws of both the nations have to be at par in order to ensure that the concerns and interests of parties involved get amicably resolved. There is urgent need to pass a comprehensive law by incorporating the latest developments in this field in order to make it ethical on the one hand and to curb disguised human trafficking on the other.Keywords: business, human trafficking, legal, surrogacy
Procedia PDF Downloads 3433851 Change Detection and Analysis of Desertification Processes in Semi Arid Land in Algeria Using Landsat Data
Authors: Zegrar Ahmed, Ghabi Mohamed
Abstract:
The degradation of arid and semi-arid ecosystems in Algeria has become a palpable fact that only hinders progress and rural development. In these exceptionally fragile environments, the decline of vegetation is done according to an alarming increase and wind erosion dominates. The ecosystem is subjected to a long hot dry season and low annual average rainfall. The urgency of the fight against desertification is imposed by the very nature of the process that tends to self-accelerate, resulting when human intervention is not forthcoming the irreversibility situations, preventing any possibility of restoration state of these zones. These phenomena have led to different degradation processes, such as the destruction of vegetation, soil erosion, and deterioration of the physical environment. In this study, the work is mainly based on the criteria for classification and identification of physical parameters for spatial analysis and multi-sources to determine the vulnerability of major steppe formations and their impact on desertification. we used Landsat data with two different dates March 2010 and November 2014 in order to determine the changes in land cover, sand moving and land degradation for the diagnosis of the desertification Phenomenon. The application, through specific processes, including the supervised classification was used to characterize the main steppe formations. An analysis of the vulnerability of plant communities was conducted to assign weights and identify areas most susceptible to desertification. Vegetation indices are used to characterize the steppe formations to determine changes in land use.Keywords: remote sensing, SIG, ecosystem, degradation, desertification
Procedia PDF Downloads 3393850 Navigating Government Finance Statistics: Effortless Retrieval and Comparative Analysis through Data Science and Machine Learning
Authors: Kwaku Damoah
Abstract:
This paper presents a methodology and software application (App) designed to empower users in accessing, retrieving, and comparatively exploring data within the hierarchical network framework of the Government Finance Statistics (GFS) system. It explores the ease of navigating the GFS system and identifies the gaps filled by the new methodology and App. The GFS, embodies a complex Hierarchical Network Classification (HNC) structure, encapsulating institutional units, revenues, expenses, assets, liabilities, and economic activities. Navigating this structure demands specialized knowledge, experience, and skill, posing a significant challenge for effective analytics and fiscal policy decision-making. Many professionals encounter difficulties deciphering these classifications, hindering confident utilization of the system. This accessibility barrier obstructs a vast number of professionals, students, policymakers, and the public from leveraging the abundant data and information within the GFS. Leveraging R programming language, Data Science Analytics and Machine Learning, an efficient methodology enabling users to access, navigate, and conduct exploratory comparisons was developed. The machine learning Fiscal Analytics App (FLOWZZ) democratizes access to advanced analytics through its user-friendly interface, breaking down expertise barriers.Keywords: data science, data wrangling, drilldown analytics, government finance statistics, hierarchical network classification, machine learning, web application.
Procedia PDF Downloads 70