Search results for: non-formal learning contexts
4028 Integrating Wound Location Data with Deep Learning for Improved Wound Classification
Authors: Mouli Banga, Chaya Ravindra
Abstract:
Wound classification is a crucial step in wound diagnosis. An effective classifier can aid wound specialists in identifying wound types with reduced financial and time investments, facilitating the determination of optimal treatment procedures. This study presents a deep neural network-based classifier that leverages wound images and their corresponding locations to categorize wounds into various classes, such as diabetic, pressure, surgical, and venous ulcers. By incorporating a developed body map, the process of tagging wound locations is significantly enhanced, providing healthcare specialists with a more efficient tool for wound analysis. We conducted a comparative analysis between two prominent convolutional neural network models, ResNet50 and MobileNetV2, utilizing a dataset of 730 images. Our findings reveal that the RestNet50 outperforms MovileNetV2, achieving an accuracy of approximately 90%, compared to MobileNetV2’s 83%. This disparity highlights the superior capability of ResNet50 in the context of this dataset. The results underscore the potential of integrating deep learning with spatial data to improve the precision and efficiency of wound diagnosis, ultimately contributing to better patient outcomes and reducing healthcare costs.Keywords: wound classification, MobileNetV2, ResNet50, multimodel
Procedia PDF Downloads 324027 Feature Engineering Based Detection of Buffer Overflow Vulnerability in Source Code Using Deep Neural Networks
Authors: Mst Shapna Akter, Hossain Shahriar
Abstract:
One of the most important challenges in the field of software code audit is the presence of vulnerabilities in software source code. Every year, more and more software flaws are found, either internally in proprietary code or revealed publicly. These flaws are highly likely exploited and lead to system compromise, data leakage, or denial of service. C and C++ open-source code are now available in order to create a largescale, machine-learning system for function-level vulnerability identification. We assembled a sizable dataset of millions of opensource functions that point to potential exploits. We developed an efficient and scalable vulnerability detection method based on deep neural network models that learn features extracted from the source codes. The source code is first converted into a minimal intermediate representation to remove the pointless components and shorten the dependency. Moreover, we keep the semantic and syntactic information using state-of-the-art word embedding algorithms such as glove and fastText. The embedded vectors are subsequently fed into deep learning networks such as LSTM, BilSTM, LSTM-Autoencoder, word2vec, BERT, and GPT-2 to classify the possible vulnerabilities. Furthermore, we proposed a neural network model which can overcome issues associated with traditional neural networks. Evaluation metrics such as f1 score, precision, recall, accuracy, and total execution time have been used to measure the performance. We made a comparative analysis between results derived from features containing a minimal text representation and semantic and syntactic information. We found that all of the deep learning models provide comparatively higher accuracy when we use semantic and syntactic information as the features but require higher execution time as the word embedding the algorithm puts on a bit of complexity to the overall system.Keywords: cyber security, vulnerability detection, neural networks, feature extraction
Procedia PDF Downloads 894026 Machine Learning Prediction of Diabetes Prevalence in the U.S. Using Demographic, Physical, and Lifestyle Indicators: A Study Based on NHANES 2009-2018
Authors: Oluwafunmibi Omotayo Fasanya, Augustine Kena Adjei
Abstract:
To develop a machine learning model to predict diabetes (DM) prevalence in the U.S. population using demographic characteristics, physical indicators, and lifestyle habits, and to analyze how these factors contribute to the likelihood of diabetes. We analyzed data from 23,546 participants aged 20 and older, who were non-pregnant, from the 2009-2018 National Health and Nutrition Examination Survey (NHANES). The dataset included key demographic (age, sex, ethnicity), physical (BMI, leg length, total cholesterol [TCHOL], fasting plasma glucose), and lifestyle indicators (smoking habits). A weighted sample was used to account for NHANES survey design features such as stratification and clustering. A classification machine learning model was trained to predict diabetes status. The target variable was binary (diabetes or non-diabetes) based on fasting plasma glucose measurements. The following models were evaluated: Logistic Regression (baseline), Random Forest Classifier, Gradient Boosting Machine (GBM), Support Vector Machine (SVM). Model performance was assessed using accuracy, F1-score, AUC-ROC, and precision-recall metrics. Feature importance was analyzed using SHAP values to interpret the contributions of variables such as age, BMI, ethnicity, and smoking status. The Gradient Boosting Machine (GBM) model outperformed other classifiers with an AUC-ROC score of 0.85. Feature importance analysis revealed the following key predictors: Age: The most significant predictor, with diabetes prevalence increasing with age, peaking around the 60s for males and 70s for females. BMI: Higher BMI was strongly associated with a higher risk of diabetes. Ethnicity: Black participants had the highest predicted prevalence of diabetes (14.6%), followed by Mexican-Americans (13.5%) and Whites (10.6%). TCHOL: Diabetics had lower total cholesterol levels, particularly among White participants (mean decline of 23.6 mg/dL). Smoking: Smoking showed a slight increase in diabetes risk among Whites (0.2%) but had a limited effect in other ethnic groups. Using machine learning models, we identified key demographic, physical, and lifestyle predictors of diabetes in the U.S. population. The results confirm that diabetes prevalence varies significantly across age, BMI, and ethnic groups, with lifestyle factors such as smoking contributing differently by ethnicity. These findings provide a basis for more targeted public health interventions and resource allocation for diabetes management.Keywords: diabetes, NHANES, random forest, gradient boosting machine, support vector machine
Procedia PDF Downloads 84025 Bounded Rational Heterogeneous Agents in Artificial Stock Markets: Literature Review and Research Direction
Authors: Talal Alsulaiman, Khaldoun Khashanah
Abstract:
In this paper, we provided a literature survey on the artificial stock problem (ASM). The paper began by exploring the complexity of the stock market and the needs for ASM. ASM aims to investigate the link between individual behaviors (micro level) and financial market dynamics (macro level). The variety of patterns at the macro level is a function of the AFM complexity. The financial market system is a complex system where the relationship between the micro and macro level cannot be captured analytically. Computational approaches, such as simulation, are expected to comprehend this connection. Agent-based simulation is a simulation technique commonly used to build AFMs. The paper proceeds by discussing the components of the ASM. We consider the roles of behavioral finance (BF) alongside the traditionally risk-averse assumption in the construction of agent's attributes. Also, the influence of social networks in the developing of agents’ interactions is addressed. Network topologies such as a small world, distance-based, and scale-free networks may be utilized to outline economic collaborations. In addition, the primary methods for developing agents learning and adaptive abilities have been summarized. These incorporated approach such as Genetic Algorithm, Genetic Programming, Artificial neural network and Reinforcement Learning. In addition, the most common statistical properties (the stylized facts) of stock that are used for calibration and validation of ASM are discussed. Besides, we have reviewed the major related previous studies and categorize the utilized approaches as a part of these studies. Finally, research directions and potential research questions are argued. The research directions of ASM may focus on the macro level by analyzing the market dynamic or on the micro level by investigating the wealth distributions of the agents.Keywords: artificial stock markets, market dynamics, bounded rationality, agent based simulation, learning, interaction, social networks
Procedia PDF Downloads 3544024 General Mood and Emotional Regulation as Predictors of Bullying Behaviors among Adolescent Males: Basis for a Proposed Bullying Intervention Program
Authors: Angelyn Del Mundo
Abstract:
Bullying cases are a proliferating issue that schools need to address. This calls for a challenge in providing effective measures to reduce bullying. The study aimed to determine which among the socio-emotional aspects of adolescent males could predict bullying. The respondents of the study were the grades 10 and 11 level and the selection of the respondents was based on the names listed by the teachers and guidance counselors through the Student Nomination Questionnaire. The Bullying Survey Questionnaire Checklist was answered by the respondents to be able to identify their most observed bullying behavior. On the other hand, the level of their mental ability was measured through the use of Otis-Lennon School Ability Test, while their socio-emotional aspects was is classified into 2 contexts: emotional intelligence and personality traits which were determined with the use of Bar-On Emotional Quotient Inventory: Youth Version (BarOn EQ-i:YV) and the Five-Factor Personality Inventory-Children (FFPI-C). Results indicated that majority of the respondents have average level of mental ability and socio-emotional aspects. However, many students have low to markedly low level interpersonal scale. Furthermore, general mood and emotional regulation were found as predictors of bullying behaviors. These findings became the basis for a proposed bullying intervention program.Keywords: bullying, emotional intelligence, mental ability, personality traits
Procedia PDF Downloads 2824023 Education for Sustainability Using PBL on an Engineering Course at the National University of Colombia
Authors: Hernán G. Cortés-Mora, José I. Péna-Reyes, Alfonso Herrera-Jiménez
Abstract:
This article describes the implementation experience of Project-Based Learning (PBL) in an engineering course of the Universidad Nacional de Colombia, with the aim of strengthening student skills necessary for the exercise of their profession under a sustainability framework. Firstly, we present a literature review on the education for sustainability field, emphasizing the skills and knowledge areas required for its development, as well as the commitment of the Faculty of Engineering of the Universidad Nacional de Colombia, and other engineering faculties of the country, regarding education for sustainability. This article covers the general aspects of the course, describes how students team were formed, and how their experience was during the first semester of 2017. During this period two groups of students decided to develop their course project aiming to solve a problem regarding a Non-Governmental Organization (NGO) that works with head-of-household mothers in a low-income neighborhood in Bogota (Colombia). Subsequently, we show how sustainability is involved in the course, how tools are provided to students, and how activities are developed as to strengthen their abilities, which allows them to incorporate sustainability in their projects while also working on the methodology used to develop said projects. Finally, we introduce the results obtained by the students who sent the prototypes of their projects to the community they were working on and the conclusions reached by them regarding the course experience.Keywords: sustainability, project-based learning, engineering education, higher education for sustainability
Procedia PDF Downloads 3514022 Clustering and Modelling Electricity Conductors from 3D Point Clouds in Complex Real-World Environments
Authors: Rahul Paul, Peter Mctaggart, Luke Skinner
Abstract:
Maintaining public safety and network reliability are the core objectives of all electricity distributors globally. For many electricity distributors, managing vegetation clearances from their above ground assets (poles and conductors) is the most important and costly risk mitigation control employed to meet these objectives. Light Detection And Ranging (LiDAR) is widely used by utilities as a cost-effective method to inspect their spatially-distributed assets at scale, often captured using high powered LiDAR scanners attached to fixed wing or rotary aircraft. The resulting 3D point cloud model is used by these utilities to perform engineering grade measurements that guide the prioritisation of vegetation cutting programs. Advances in computer vision and machine-learning approaches are increasingly applied to increase automation and reduce inspection costs and time; however, real-world LiDAR capture variables (e.g., aircraft speed and height) create complexity, noise, and missing data, reducing the effectiveness of these approaches. This paper proposes a method for identifying each conductor from LiDAR data via clustering methods that can precisely reconstruct conductors in complex real-world configurations in the presence of high levels of noise. It proposes 3D catenary models for individual clusters fitted to the captured LiDAR data points using a least square method. An iterative learning process is used to identify potential conductor models between pole pairs. The proposed method identifies the optimum parameters of the catenary function and then fits the LiDAR points to reconstruct the conductors.Keywords: point cloud, LİDAR data, machine learning, computer vision, catenary curve, vegetation management, utility industry
Procedia PDF Downloads 994021 The Art and Science of Trauma-Informed Psychotherapy: Guidelines for Inter-Disciplinary Clinicians
Authors: Daphne Alroy-Thiberge
Abstract:
Trauma-impacted individuals present unique treatment challenges that include high reactivity, hyper-and hypo-arousal, poor adherence to therapy, as well as powerful transference and counter-transference experiences in therapy. This work provides an overview of the clinical tenets most often encountered in trauma-impacted individuals. Further, it provides readily applicable clinical techniques to optimize therapeutic rapport and facilitate accelerated positive mental health outcomes. Finally, integrated neuroscience and clinical evidence-based data are discussed to shed new light on crisis states in trauma-impacted individuals. This knowledge is utilized to provide effective and concrete interventions towards rapid and successful de-escalation of the impacted individual. A highly interactive, adult-learning-principles-based modality is utilized to provide an organic learning experience for participants. The information and techniques learned aim to increase clinical effectiveness, reduce staff injuries and burnout, and significantly enhance positive mental health outcomes and self-determination for the trauma-impacted individuals treated.Keywords: clinical competencies, crisis interventions, psychotherapy techniques, trauma informed care
Procedia PDF Downloads 1084020 Improving the Students’ Writing Skill by Using Brainstorming Technique
Authors: M. Z. Abdul Rofiq Badril Rizal
Abstract:
This research is aimed to know the improvement of students’ English writing skill by using brainstorming technique. The technique used in writing is able to help the students’ difficulties in generating ideas and to lead the students to arrange the ideas well as well as to focus on the topic developed in writing. The research method used is classroom action research. The data sources of the research are an English teacher who acts as an observer and the students of class X.MIA5 consist of 35 students. The test result and observation are collected as the data in this research. Based on the research result in cycle one, the percentage of students who reach minimum accomplishment criteria (MAC) is 76.31%. It shows that the cycle must be continued to cycle two because the aim of the research has not accomplished, all of the students’ scores have not reached MAC yet. After continuing the research to cycle two and the weaknesses are improved, the process of teaching and learning runs better. At the test which is conducted in the end of learning process in cycle two, all of the students reach the minimum score and above 76 based on the minimum accomplishment criteria. It means the research has been successful and the percentage of students who reach minimum accomplishment criteria is 100%. Therefore, the writer concludes that brainstorming technique is able to improve the students’ English writing skill at the tenth grade of SMAN 2 Jember.Keywords: brainstorming technique, improving, writing skill, knowledge and innovation engineering
Procedia PDF Downloads 3674019 Autism Spectrum Disorder Classification Algorithm Using Multimodal Data Based on Graph Convolutional Network
Authors: Yuntao Liu, Lei Wang, Haoran Xia
Abstract:
Machine learning has shown extensive applications in the development of classification models for autism spectrum disorder (ASD) using neural image data. This paper proposes a fusion multi-modal classification network based on a graph neural network. First, the brain is segmented into 116 regions of interest using a medical segmentation template (AAL, Anatomical Automatic Labeling). The image features of sMRI and the signal features of fMRI are extracted, which build the node and edge embedding representations of the brain map. Then, we construct a dynamically updated brain map neural network and propose a method based on a dynamic brain map adjacency matrix update mechanism and learnable graph to further improve the accuracy of autism diagnosis and recognition results. Based on the Autism Brain Imaging Data Exchange I dataset(ABIDE I), we reached a prediction accuracy of 74% between ASD and TD subjects. Besides, to study the biomarkers that can help doctors analyze diseases and interpretability, we used the features by extracting the top five maximum and minimum ROI weights. This work provides a meaningful way for brain disorder identification.Keywords: autism spectrum disorder, brain map, supervised machine learning, graph network, multimodal data, model interpretability
Procedia PDF Downloads 664018 Unraveling the Phonosignological Foundations of Human Language and Semantic Analysis of Linguistic Elements in Cross-Cultural Contexts
Authors: Mahmudjon Kuchkarov, Marufjon Kuchkarov, Mukhayyo Sobirjanova
Abstract:
The origins of human language remain a profound scientific mystery, characterized by speculative theories often lacking empirical support. This study presents findings that may illuminate the genesis of human language, emphasizing its roots in natural, systematic, and repetitive sound patterns. Also, this paper presents the phonosignological and semantic analysis of linguistic elements across various languages and cultures. By utilizing the principles of the "Human Language" theory, we analyze the symbolic, phonetic, and semantic characteristics of elements such as "A", "L", "I", "F", and "四" (pronounced /si/ in Chinese and /shi/ in Japanese). Our findings reveal that natural sounds and their symbolic representations form the foundation of language, with significant implications for understanding religious and secular myths. This paper explores the intricate relationships between these elements and their cultural connotations, particularly focusing on the concept of "descent" in the context of the phonetic sequence "A, L, I, F," and the symbolic associations of the number four with death.Keywords: empirical research, human language, phonosignology, semantics, sound patterns, symbolism, body shape, body language, coding, Latin alphabet, merging method, natural sound, origin of language, pairing, phonetics, sound and shape production, word origin, word semantic
Procedia PDF Downloads 374017 Design and Implementation of Machine Learning Model for Short-Term Energy Forecasting in Smart Home Management System
Authors: R. Ramesh, K. K. Shivaraman
Abstract:
The main aim of this paper is to handle the energy requirement in an efficient manner by merging the advanced digital communication and control technologies for smart grid applications. In order to reduce user home load during peak load hours, utility applies several incentives such as real-time pricing, time of use, demand response for residential customer through smart meter. However, this method provides inconvenience in the sense that user needs to respond manually to prices that vary in real time. To overcome these inconvenience, this paper proposes a convolutional neural network (CNN) with k-means clustering machine learning model which have ability to forecast energy requirement in short term, i.e., hour of the day or day of the week. By integrating our proposed technique with home energy management based on Bluetooth low energy provides predicted value to user for scheduling appliance in advanced. This paper describes detail about CNN configuration and k-means clustering algorithm for short-term energy forecasting.Keywords: convolutional neural network, fuzzy logic, k-means clustering approach, smart home energy management
Procedia PDF Downloads 3054016 Service Information Integration Platform as Decision Making Tools for the Service Industry Supply Chain-Indonesia Service Integration Project
Authors: Haikal Achmad Thaha, Pujo Laksono, Dhamma Nibbana Putra
Abstract:
Customer service is one of the core interest in a service sector of a company, whether as the core business or as service part of the operation. Most of the time, the people and the previous research in service industry is focused on finding the best business model solution for the service sector, usually to decide between total in house customer service, outsourcing, or something in between. Conventionally, to take this decision is some important part of the management job, and this is a process that usually takes some time and staff effort, meanwhile market condition and overall company needs may change and cause loss of income and temporary disturbance in the companies operation . However, in this paper we have offer a new concept model to assist decision making process in service industry. This model will featured information platform as central tool to integrate service industry operation. The result is service information model which would ideally increase response time and effectivity of the decision making. it will also help service industry in switching the service solution system quickly through machine learning when the companies growth and the service solution needed are changing.Keywords: service industry, customer service, machine learning, decision making, information platform
Procedia PDF Downloads 6224015 Prediction of Survival Rate after Gastrointestinal Surgery Based on The New Japanese Association for Acute Medicine (JAAM Score) With Neural Network Classification Method
Authors: Ayu Nabila Kusuma Pradana, Aprinaldi Jasa Mantau, Tomohiko Akahoshi
Abstract:
The incidence of Disseminated intravascular coagulation (DIC) following gastrointestinal surgery has a poor prognosis. Therefore, it is important to determine the factors that can predict the prognosis of DIC. This study will investigate the factors that may influence the outcome of DIC in patients after gastrointestinal surgery. Eighty-one patients were admitted to the intensive care unit after gastrointestinal surgery in Kyushu University Hospital from 2003 to 2021. Acute DIC scores were estimated using the new Japanese Association for Acute Medicine (JAAM) score from before and after surgery from day 1, day 3, and day 7. Acute DIC scores will be compared with The Sequential Organ Failure Assessment (SOFA) score, platelet count, lactate level, and a variety of biochemical parameters. This study applied machine learning algorithms to predict the prognosis of DIC after gastrointestinal surgery. The results of this study are expected to be used as an indicator for evaluating patient prognosis so that it can increase life expectancy and reduce mortality from cases of DIC patients after gastrointestinal surgery.Keywords: the survival rate, gastrointestinal surgery, JAAM score, neural network, machine learning, disseminated intravascular coagulation (DIC)
Procedia PDF Downloads 2604014 On the Utility of Bidirectional Transformers in Gene Expression-Based Classification
Authors: Babak Forouraghi
Abstract:
A genetic circuit is a collection of interacting genes and proteins that enable individual cells to implement and perform vital biological functions such as cell division, growth, death, and signaling. In cell engineering, synthetic gene circuits are engineered networks of genes specifically designed to implement functionalities that are not evolved by nature. These engineered networks enable scientists to tackle complex problems such as engineering cells to produce therapeutics within the patient's body, altering T cells to target cancer-related antigens for treatment, improving antibody production using engineered cells, tissue engineering, and production of genetically modified plants and livestock. Construction of computational models to realize genetic circuits is an especially challenging task since it requires the discovery of the flow of genetic information in complex biological systems. Building synthetic biological models is also a time-consuming process with relatively low prediction accuracy for highly complex genetic circuits. The primary goal of this study was to investigate the utility of a pre-trained bidirectional encoder transformer that can accurately predict gene expressions in genetic circuit designs. The main reason behind using transformers is their innate ability (attention mechanism) to take account of the semantic context present in long DNA chains that are heavily dependent on the spatial representation of their constituent genes. Previous approaches to gene circuit design, such as CNN and RNN architectures, are unable to capture semantic dependencies in long contexts, as required in most real-world applications of synthetic biology. For instance, RNN models (LSTM, GRU), although able to learn long-term dependencies, greatly suffer from vanishing gradient and low-efficiency problem when they sequentially process past states and compresses contextual information into a bottleneck with long input sequences. In other words, these architectures are not equipped with the necessary attention mechanisms to follow a long chain of genes with thousands of tokens. To address the above-mentioned limitations, a transformer model was built in this work as a variation to the existing DNA Bidirectional Encoder Representations from Transformers (DNABERT) model. It is shown that the proposed transformer is capable of capturing contextual information from long input sequences with an attention mechanism. In previous works on genetic circuit design, the traditional approaches to classification and regression, such as Random Forrest, Support Vector Machine, and Artificial Neural Networks, were able to achieve reasonably high R2 accuracy levels of 0.95 to 0.97. However, the transformer model utilized in this work, with its attention-based mechanism, was able to achieve a perfect accuracy level of 100%. Further, it is demonstrated that the efficiency of the transformer-based gene expression classifier is not dependent on the presence of large amounts of training examples, which may be difficult to compile in many real-world gene circuit designs.Keywords: machine learning, classification and regression, gene circuit design, bidirectional transformers
Procedia PDF Downloads 614013 Utilising Sociodrama as Classroom Intervention to Develop Sensory Integration in Adolescents who Present with Mild Impaired Learning
Authors: Talita Veldsman, Elzette Fritz
Abstract:
Many children attending special education present with sensory integration difficulties that hamper their learning and behaviour. These learners can benefit from therapeutic interventions as part of their classroom curriculum that can address sensory development and allow for holistic development to take place. A research study was conducted by utilizing socio-drama as a therapeutic intervention in the classroom in order to develop sensory integration skills. The use of socio-drama as therapeutic intervention proved to be a successful multi-disciplinary approach where education and psychology could build a bridge of growth and integration. The paper describes how socio-drama was used in the classroom and how these sessions were designed. The research followed a qualitative approach and involved six Afrikaans-speaking children attending special secondary school in the age group 12-14 years. Data collection included observations during the session, reflective art journals, semi-structured interviews with the teacher and informal interviews with the adolescents. The analysis found improved self-confidence, better social relationships, sensory awareness and self-regulation in the participants after a period of a year.Keywords: education, sensory integration, sociodrama, classroom intervention, psychology
Procedia PDF Downloads 5774012 Data Refinement Enhances The Accuracy of Short-Term Traffic Latency Prediction
Authors: Man Fung Ho, Lap So, Jiaqi Zhang, Yuheng Zhao, Huiyang Lu, Tat Shing Choi, K. Y. Michael Wong
Abstract:
Nowadays, a tremendous amount of data is available in the transportation system, enabling the development of various machine learning approaches to make short-term latency predictions. A natural question is then the choice of relevant information to enable accurate predictions. Using traffic data collected from the Taiwan Freeway System, we consider the prediction of short-term latency of a freeway segment with a length of 17 km covering 5 measurement points, each collecting vehicle-by-vehicle data through the electronic toll collection system. The processed data include the past latencies of the freeway segment with different time lags, the traffic conditions of the individual segments (the accumulations, the traffic fluxes, the entrance and exit rates), the total accumulations, and the weekday latency profiles obtained by Gaussian process regression of past data. We arrive at several important conclusions about how data should be refined to obtain accurate predictions, which have implications for future system-wide latency predictions. (1) We find that the prediction of median latency is much more accurate and meaningful than the prediction of average latency, as the latter is plagued by outliers. This is verified by machine-learning prediction using XGBoost that yields a 35% improvement in the mean square error of the 5-minute averaged latencies. (2) We find that the median latency of the segment 15 minutes ago is a very good baseline for performance comparison, and we have evidence that further improvement is achieved by machine learning approaches such as XGBoost and Long Short-Term Memory (LSTM). (3) By analyzing the feature importance score in XGBoost and calculating the mutual information between the inputs and the latencies to be predicted, we identify a sequence of inputs ranked in importance. It confirms that the past latencies are most informative of the predicted latencies, followed by the total accumulation, whereas inputs such as the entrance and exit rates are uninformative. It also confirms that the inputs are much less informative of the average latencies than the median latencies. (4) For predicting the latencies of segments composed of two or three sub-segments, summing up the predicted latencies of each sub-segment is more accurate than the one-step prediction of the whole segment, especially with the latency prediction of the downstream sub-segments trained to anticipate latencies several minutes ahead. The duration of the anticipation time is an increasing function of the traveling time of the upstream segment. The above findings have important implications to predicting the full set of latencies among the various locations in the freeway system.Keywords: data refinement, machine learning, mutual information, short-term latency prediction
Procedia PDF Downloads 1694011 Teacher Professionalisation and Professionalism Discourses in Teacher Unions: A Case Study of New Zealand
Authors: Huidan Niu
Abstract:
Existing research has focused extensively on teachers’ professional experience in education reforms. However, there is a lack of research on the role and influence of teacher unions in education policy. This study aimed to examine how teacher unions frame teacher professionalisation and professionalism discourses. Critical education policy scholarship study was adopted. This study positioned teacher professionalisation and professionalism discourses within their socio-political contexts to explore how the meanings of teacher professionalisation and professionalism are constructed, as well as how teacher unions, as collective actors, shape these discourses. This study examined the development of professionalisation and professionalism discourses in the two main teacher unions in Aotearoa, New Zealand, the New Zealand Educational Institute, TeRiuRoa (NZEI), and the New Zealand Post-Primary Teachers’ Association, TeWehengarua (PPTA). The data were collected from documents and archival material, as well as elite interviews. Twenty-four union leaders, including national presidents, secretaries, executives, and senior union officials, participated in the study. The data analysis followed a grounded theory method: from codes to themes. The findings of the study suggest that the teacher unions, as teachers’ collective (powerful) voices, appeared to highlight tension and confrontation between the teaching profession and governments with respect to the meanings of teacher professionalisation and professionalism.Keywords: critical education policy scholarship, governments, teacher professionalisation, teacher professionalism, teacher unions
Procedia PDF Downloads 1294010 Automatic Adult Age Estimation Using Deep Learning of the ResNeXt Model Based on CT Reconstruction Images of the Costal Cartilage
Authors: Ting Lu, Ya-Ru Diao, Fei Fan, Ye Xue, Lei Shi, Xian-e Tang, Meng-jun Zhan, Zhen-hua Deng
Abstract:
Accurate adult age estimation (AAE) is a significant and challenging task in forensic and archeology fields. Attempts have been made to explore optimal adult age metrics, and the rib is considered a potential age marker. The traditional way is to extract age-related features designed by experts from macroscopic or radiological images followed by classification or regression analysis. Those results still have not met the high-level requirements for practice, and the limitation of using feature design and manual extraction methods is loss of information since the features are likely not designed explicitly for extracting information relevant to age. Deep learning (DL) has recently garnered much interest in imaging learning and computer vision. It enables learning features that are important without a prior bias or hypothesis and could be supportive of AAE. This study aimed to develop DL models for AAE based on CT images and compare their performance to the manual visual scoring method. Chest CT data were reconstructed using volume rendering (VR). Retrospective data of 2500 patients aged 20.00-69.99 years were obtained between December 2019 and September 2021. Five-fold cross-validation was performed, and datasets were randomly split into training and validation sets in a 4:1 ratio for each fold. Before feeding the inputs into networks, all images were augmented with random rotation and vertical flip, normalized, and resized to 224×224 pixels. ResNeXt was chosen as the DL baseline due to its advantages of higher efficiency and accuracy in image classification. Mean absolute error (MAE) was the primary parameter. Independent data from 100 patients acquired between March and April 2022 were used as a test set. The manual method completely followed the prior study, which reported the lowest MAEs (5.31 in males and 6.72 in females) among similar studies. CT data and VR images were used. The radiation density of the first costal cartilage was recorded using CT data on the workstation. The osseous and calcified projections of the 1 to 7 costal cartilages were scored based on VR images using an eight-stage staging technique. According to the results of the prior study, the optimal models were the decision tree regression model in males and the stepwise multiple linear regression equation in females. Predicted ages of the test set were calculated separately using different models by sex. A total of 2600 patients (training and validation sets, mean age=45.19 years±14.20 [SD]; test set, mean age=46.57±9.66) were evaluated in this study. Of ResNeXt model training, MAEs were obtained with 3.95 in males and 3.65 in females. Based on the test set, DL achieved MAEs of 4.05 in males and 4.54 in females, which were far better than the MAEs of 8.90 and 6.42 respectively, for the manual method. Those results showed that the DL of the ResNeXt model outperformed the manual method in AAE based on CT reconstruction of the costal cartilage and the developed system may be a supportive tool for AAE.Keywords: forensic anthropology, age determination by the skeleton, costal cartilage, CT, deep learning
Procedia PDF Downloads 734009 Analysis and Detection of Facial Expressions in Autism Spectrum Disorder People Using Machine Learning
Authors: Muhammad Maisam Abbas, Salman Tariq, Usama Riaz, Muhammad Tanveer, Humaira Abdul Ghafoor
Abstract:
Autism Spectrum Disorder (ASD) refers to a developmental disorder that impairs an individual's communication and interaction ability. Individuals feel difficult to read facial expressions while communicating or interacting. Facial Expression Recognition (FER) is a unique method of classifying basic human expressions, i.e., happiness, fear, surprise, sadness, disgust, neutral, and anger through static and dynamic sources. This paper conducts a comprehensive comparison and proposed optimal method for a continued research project—a system that can assist people who have Autism Spectrum Disorder (ASD) in recognizing facial expressions. Comparison has been conducted on three supervised learning algorithms EigenFace, FisherFace, and LBPH. The JAFFE, CK+, and TFEID (I&II) datasets have been used to train and test the algorithms. The results were then evaluated based on variance, standard deviation, and accuracy. The experiments showed that FisherFace has the highest accuracy for all datasets and is considered the best algorithm to be implemented in our system.Keywords: autism spectrum disorder, ASD, EigenFace, facial expression recognition, FisherFace, local binary pattern histogram, LBPH
Procedia PDF Downloads 1744008 Using Deep Learning in Lyme Disease Diagnosis
Authors: Teja Koduru
Abstract:
Untreated Lyme disease can lead to neurological, cardiac, and dermatological complications. Rapid diagnosis of the erythema migrans (EM) rash, a characteristic symptom of Lyme disease is therefore crucial to early diagnosis and treatment. In this study, we aim to utilize deep learning frameworks including Tensorflow and Keras to create deep convolutional neural networks (DCNN) to detect images of acute Lyme Disease from images of erythema migrans. This study uses a custom database of erythema migrans images of varying quality to train a DCNN capable of classifying images of EM rashes vs. non-EM rashes. Images from publicly available sources were mined to create an initial database. Machine-based removal of duplicate images was then performed, followed by a thorough examination of all images by a clinician. The resulting database was combined with images of confounding rashes and regular skin, resulting in a total of 683 images. This database was then used to create a DCNN with an accuracy of 93% when classifying images of rashes as EM vs. non EM. Finally, this model was converted into a web and mobile application to allow for rapid diagnosis of EM rashes by both patients and clinicians. This tool could be used for patient prescreening prior to treatment and lead to a lower mortality rate from Lyme disease.Keywords: Lyme, untreated Lyme, erythema migrans rash, EM rash
Procedia PDF Downloads 2404007 The Role of Organizational Identity in Disaster Response, Recovery and Prevention: A Case Study of an Italian Multi-Utility Company
Authors: Shanshan Zhou, Massimo Battaglia
Abstract:
Identity plays a critical role when an organization faces disasters. Individuals reflect on their working identities and identify themselves with the group and the organization, which facilitate collective sensemaking under crisis situations and enable coordinated actions to respond to and recover from disasters. In addition, an organization’s identity links it to its regional community, which fosters the mobilization of resources and contributes to rapid recovery. However, identity is also problematic for disaster prevention because of its persistence. An organization’s ego-defenses system prohibits the rethink of its identity and a rigid identity obstructs disaster prevention. This research aims to tackle the ‘problem’ of identity by study in-depth a case of an Italian multi–utility which experienced the 2012 Northern Italy earthquakes. Collecting data from 11 interviews with top managers and key players in the local community and archived materials, we find that the earthquakes triggered the rethink of the organization’s identity, which got reinforced afterward. This research highlighted the importance of identity in disaster response and recovery. More importantly, it explored the solution of overcoming the barrier of ego-defense that is to transform the organization into a learning organization which constantly rethinks its identity.Keywords: community identity, disaster, identity, organizational learning
Procedia PDF Downloads 7324006 Single Imputation for Audiograms
Authors: Sarah Beaver, Renee Bryce
Abstract:
Audiograms detect hearing impairment, but missing values pose problems. This work explores imputations in an attempt to improve accuracy. This work implements Linear Regression, Lasso, Linear Support Vector Regression, Bayesian Ridge, K Nearest Neighbors (KNN), and Random Forest machine learning techniques to impute audiogram frequencies ranging from 125Hz to 8000Hz. The data contains patients who had or were candidates for cochlear implants. Accuracy is compared across two different Nested Cross-Validation k values. Over 4000 audiograms were used from 800 unique patients. Additionally, training on data combines and compares left and right ear audiograms versus single ear side audiograms. The accuracy achieved using Root Mean Square Error (RMSE) values for the best models for Random Forest ranges from 4.74 to 6.37. The R\textsuperscript{2} values for the best models for Random Forest ranges from .91 to .96. The accuracy achieved using RMSE values for the best models for KNN ranges from 5.00 to 7.72. The R\textsuperscript{2} values for the best models for KNN ranges from .89 to .95. The best imputation models received R\textsuperscript{2} between .89 to .96 and RMSE values less than 8dB. We also show that the accuracy of classification predictive models performed better with our best imputation models versus constant imputations by a two percent increase.Keywords: machine learning, audiograms, data imputations, single imputations
Procedia PDF Downloads 824005 Exploring the Applications of Neural Networks in the Adaptive Learning Environment
Authors: Baladitya Swaika, Rahul Khatry
Abstract:
Computer Adaptive Tests (CATs) is one of the most efficient ways for testing the cognitive abilities of students. CATs are based on Item Response Theory (IRT) which is based on item selection and ability estimation using statistical methods of maximum information selection/selection from posterior and maximum-likelihood (ML)/maximum a posteriori (MAP) estimators respectively. This study aims at combining both classical and Bayesian approaches to IRT to create a dataset which is then fed to a neural network which automates the process of ability estimation and then comparing it to traditional CAT models designed using IRT. This study uses python as the base coding language, pymc for statistical modelling of the IRT and scikit-learn for neural network implementations. On creation of the model and on comparison, it is found that the Neural Network based model performs 7-10% worse than the IRT model for score estimations. Although performing poorly, compared to the IRT model, the neural network model can be beneficially used in back-ends for reducing time complexity as the IRT model would have to re-calculate the ability every-time it gets a request whereas the prediction from a neural network could be done in a single step for an existing trained Regressor. This study also proposes a new kind of framework whereby the neural network model could be used to incorporate feature sets, other than the normal IRT feature set and use a neural network’s capacity of learning unknown functions to give rise to better CAT models. Categorical features like test type, etc. could be learnt and incorporated in IRT functions with the help of techniques like logistic regression and can be used to learn functions and expressed as models which may not be trivial to be expressed via equations. This kind of a framework, when implemented would be highly advantageous in psychometrics and cognitive assessments. This study gives a brief overview as to how neural networks can be used in adaptive testing, not only by reducing time-complexity but also by being able to incorporate newer and better datasets which would eventually lead to higher quality testing.Keywords: computer adaptive tests, item response theory, machine learning, neural networks
Procedia PDF Downloads 1754004 Using Deep Learning Neural Networks and Candlestick Chart Representation to Predict Stock Market
Authors: Rosdyana Mangir Irawan Kusuma, Wei-Chun Kao, Ho-Thi Trang, Yu-Yen Ou, Kai-Lung Hua
Abstract:
Stock market prediction is still a challenging problem because there are many factors that affect the stock market price such as company news and performance, industry performance, investor sentiment, social media sentiment, and economic factors. This work explores the predictability in the stock market using deep convolutional network and candlestick charts. The outcome is utilized to design a decision support framework that can be used by traders to provide suggested indications of future stock price direction. We perform this work using various types of neural networks like convolutional neural network, residual network and visual geometry group network. From stock market historical data, we converted it to candlestick charts. Finally, these candlestick charts will be feed as input for training a convolutional neural network model. This convolutional neural network model will help us to analyze the patterns inside the candlestick chart and predict the future movements of the stock market. The effectiveness of our method is evaluated in stock market prediction with promising results; 92.2% and 92.1 % accuracy for Taiwan and Indonesian stock market dataset respectively.Keywords: candlestick chart, deep learning, neural network, stock market prediction
Procedia PDF Downloads 4474003 A Review on Parametric Optimization of Casting Processes Using Optimization Techniques
Authors: Bhrugesh Radadiya, Jaydeep Shah
Abstract:
In Indian foundry industry, there is a need of defect free casting with minimum production cost in short lead time. Casting defect is a very large issue in foundry shop which increases the rejection rate of casting and wastage of materials. The various parameters influences on casting process such as mold machine related parameters, green sand related parameters, cast metal related parameters, mold related parameters and shake out related parameters. The mold related parameters are most influences on casting defects in sand casting process. This paper review the casting produced by foundry with shrinkage and blow holes as a major defects was analyzed and identified that mold related parameters such as mold temperature, pouring temperature and runner size were not properly set in sand casting process. These parameters were optimized using different optimization techniques such as Taguchi method, Response surface methodology, Genetic algorithm and Teaching-learning based optimization algorithm. Finally, concluded that a Teaching-learning based optimization algorithm give better result than other optimization techniques.Keywords: casting defects, genetic algorithm, parametric optimization, Taguchi method, TLBO algorithm
Procedia PDF Downloads 7284002 Trade-Offs between Verb Frequency and Syntactic Complexity in Children with Developmental Language Disorder
Authors: Pui I. Chao, Shanju Lin
Abstract:
Purpose: Children with developmental language disorder (DLD) have persistent language difficulties and often face great challenges when demands are high. The aim of this study was to investigate whether verb frequency would trade-off with syntactic complexity when they talk. Method: Forty-five children with DLD, 45 chronological age matches with TD (AGE), and 45 MLU-matches with TD (MLU) who were Mandarin speakers were selected from the previous study. Language samples were collected under three contexts: conversation about children’s family and school, story retelling, and free play. MLU, verb density, utterance length difference, verb density difference, and average verb frequency were calculated and further analyzed by ANOVAs. Results: Children with DLD and their MLU matches produced shorter utterances and used fewer verbs in expressions than the AGE matches. Compared to their AGE matches, the DLD group used more verbs and verbs with lower frequency in shorter utterances but used fewer verbs and verbs with higher frequency in longer utterances. Conclusion: Mandarin-speaking children with DLD showed difficulties in verb usage and were more vulnerable to trade-offs than their age-matched peers in utterances with high demand. As a result, task demand should be taken into account as speech-language pathologists assess whether children with DLD have adequate abilities in verb usage.Keywords: developmental language disorder, syntactic complexity, trade-offs, verb frequency
Procedia PDF Downloads 1544001 Evaluation of Technology Tools for Mathematics Instruction by Novice Elementary Teachers
Authors: Christopher J. Johnston
Abstract:
This paper presents the finding of a research study in which novice (first and second year) elementary teachers (grades Kindergarten – six) evaluated various mathematics Virtual Manipulatives, websites, and Applets (tools) for use in mathematics instruction. Participants identified the criteria they used for evaluating these types of resources and provided recommendations for or against five pre-selected tools. During the study, participants participated in three data collection activities: (1) A brief Likert-scale survey which gathered information about their attitudes toward technology use; (2) An identification of criteria for evaluating technology tools; and (3) A review of five pre-selected technology tools in light of their self-identified criteria. Data were analyzed qualitatively using four theoretical categories (codes): Software Features (41%), Mathematics (26%), Learning (22%), and Motivation (11%). These four theoretical categories were then grouped into two broad categories: Content and Instruction (Mathematics and Learning), and Surface Features (Software Features and Motivation). These combined, broad categories suggest novice teachers place roughly the same weight on pedagogical features as they do technological features. Implications for mathematics teacher educators are discussed, and suggestions for future research are provided.Keywords: mathematics education, novice teachers, technology, virtual manipulatives
Procedia PDF Downloads 1334000 Effects of Gamification on Lower Secondary School Students’ Motivation and Engagement
Authors: Goh Yung Hong, Mona Masood
Abstract:
This paper explores the effects of gamification on lower secondary school students’ motivation and engagement in the classroom. Two-group posttest-only experimental design were employed to study the influence of gamification teaching method (GTM) when compared with conventional teaching method (CTM) on 60 lower secondary school students. The Student Engagement Instrument (SEI) and Intrinsic Motivation Inventory (IMI) were used to assess students’ intrinsic motivation and engagement level towards the respective teaching method. Finding indicates that students who completed the GTM lesson were significantly higher in intrinsic motivation to learn than those from the CTM. Although the result were insignificant and only marginal difference in the engagement mean, GTM still show better potential in raising student’s engagement in class when compared with CTM. This finding proves that the GTM is likely to solve the current issue of low motivation to learn and low engagement in class among lower secondary school students in Malaysia. On the other hand, despite being not significant, higher mean indicates that CTM positively contribute to higher peer support for learning and better teacher and student relationship when compared with GTM. As a conclusion, gamification approach is flexible and can be adapted into many learning content to enhance the intrinsic motivation to learn and to some extent, encourage better student engagement in class.Keywords: conventional teaching method, gamification teaching method, motivation, engagement
Procedia PDF Downloads 5263999 Effectiveness of the Model in the Development of Teaching Materials for Malay Language in Primary Schools in Singapore
Authors: Salha Mohamed Hussain
Abstract:
As part of the review on the Malay Language curriculum and pedagogy in Singapore conducted in 2010, some recommendations were made to nurture active learners who are able to use the Malay Language efficiently in their daily lives. In response to the review, a new Malay Language teaching and learning package for primary school, called CEKAP (Cungkil – Elicit; Eksplorasi – Exploration; Komunikasi – Communication; Aplikasi – Application; Penilaian – Assessment), was developed from 2012 and implemented for Primary 1 in all primary schools from 2015. Resources developed in this package include the text book, activity book, teacher’s guide, big books, small readers, picture cards, flash cards, a game kit and Information and Communication Technology (ICT) resources. The development of the CEKAP package is continuous until 2020. This paper will look at a model incorporated in the development of the teaching materials in the new Malay Language Curriculum for Primary Schools and the rationale for each phase of development to ensure that the resources meet the needs of every pupil in the teaching and learning of Malay Language in the primary schools. This paper will also focus on the preliminary findings of the effectiveness of the model based on the feedback given by members of the working and steering committees. These members are academicians and educators who were appointed by the Ministry of Education to provide professional input on the soundness of pedagogical approach proposed in the revised syllabus and to make recommendations on the content of the new instructional materials. Quantitative data is derived from the interviews held with these members to gather their input on the model. Preliminary findings showed that the members provided positive feedback on the model and that the comprehensive process has helped to develop good and effective instructional materials for the schools. Some recommendations were also gathered from the interview sessions. This research hopes to provide useful information to those involved in the planning of materials development for teaching and learning.Keywords: Malay language, materials development, model, primary school
Procedia PDF Downloads 112