Search results for: explorative learning
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7267

Search results for: explorative learning

3247 Enabling Oral Communication and Accelerating Recovery: The Creation of a Novel Low-Cost Electroencephalography-Based Brain-Computer Interface for the Differently Abled

Authors: Rishabh Ambavanekar

Abstract:

Expressive Aphasia (EA) is an oral disability, common among stroke victims, in which the Broca’s area of the brain is damaged, interfering with verbal communication abilities. EA currently has no technological solutions and its only current viable solutions are inefficient or only available to the affluent. This prompts the need for an affordable, innovative solution to facilitate recovery and assist in speech generation. This project proposes a novel concept: using a wearable low-cost electroencephalography (EEG) device-based brain-computer interface (BCI) to translate a user’s inner dialogue into words. A low-cost EEG device was developed and found to be 10 to 100 times less expensive than any current EEG device on the market. As part of the BCI, a machine learning (ML) model was developed and trained using the EEG data. Two stages of testing were conducted to analyze the effectiveness of the device: a proof-of-concept and a final solution test. The proof-of-concept test demonstrated an average accuracy of above 90% and the final solution test demonstrated an average accuracy of above 75%. These two successful tests were used as a basis to demonstrate the viability of BCI research in developing lower-cost verbal communication devices. Additionally, the device proved to not only enable users to verbally communicate but has the potential to also assist in accelerated recovery from the disorder.

Keywords: neurotechnology, brain-computer interface, neuroscience, human-machine interface, BCI, HMI, aphasia, verbal disability, stroke, low-cost, machine learning, ML, image recognition, EEG, signal analysis

Procedia PDF Downloads 119
3246 Classifier for Liver Ultrasound Images

Authors: Soumya Sajjan

Abstract:

Liver cancer is the most common cancer disease worldwide in men and women, and is one of the few cancers still on the rise. Liver disease is the 4th leading cause of death. According to new NHS (National Health Service) figures, deaths from liver diseases have reached record levels, rising by 25% in less than a decade; heavy drinking, obesity, and hepatitis are believed to be behind the rise. In this study, we focus on Development of Diagnostic Classifier for Ultrasound liver lesion. Ultrasound (US) Sonography is an easy-to-use and widely popular imaging modality because of its ability to visualize many human soft tissues/organs without any harmful effect. This paper will provide an overview of underlying concepts, along with algorithms for processing of liver ultrasound images Naturaly, Ultrasound liver lesion images are having more spackle noise. Developing classifier for ultrasound liver lesion image is a challenging task. We approach fully automatic machine learning system for developing this classifier. First, we segment the liver image by calculating the textural features from co-occurrence matrix and run length method. For classification, Support Vector Machine is used based on the risk bounds of statistical learning theory. The textural features for different features methods are given as input to the SVM individually. Performance analysis train and test datasets carried out separately using SVM Model. Whenever an ultrasonic liver lesion image is given to the SVM classifier system, the features are calculated, classified, as normal and diseased liver lesion. We hope the result will be helpful to the physician to identify the liver cancer in non-invasive method.

Keywords: segmentation, Support Vector Machine, ultrasound liver lesion, co-occurance Matrix

Procedia PDF Downloads 411
3245 Web-Based Cognitive Writing Instruction (WeCWI): A Theoretical-and-Pedagogical e-Framework for Language Development

Authors: Boon Yih Mah

Abstract:

Web-based Cognitive Writing Instruction (WeCWI)’s contribution towards language development can be divided into linguistic and non-linguistic perspectives. In linguistic perspective, WeCWI focuses on the literacy and language discoveries, while the cognitive and psychological discoveries are the hubs in non-linguistic perspective. In linguistic perspective, WeCWI draws attention to free reading and enterprises, which are supported by the language acquisition theories. Besides, the adoption of process genre approach as a hybrid guided writing approach fosters literacy development. Literacy and language developments are interconnected in the communication process; hence, WeCWI encourages meaningful discussion based on the interactionist theory that involves input, negotiation, output, and interactional feedback. Rooted in the e-learning interaction-based model, WeCWI promotes online discussion via synchronous and asynchronous communications, which allows interactions happened among the learners, instructor, and digital content. In non-linguistic perspective, WeCWI highlights on the contribution of reading, discussion, and writing towards cognitive development. Based on the inquiry models, learners’ critical thinking is fostered during information exploration process through interaction and questioning. Lastly, to lower writing anxiety, WeCWI develops the instructional tool with supportive features to facilitate the writing process. To bring a positive user experience to the learner, WeCWI aims to create the instructional tool with different interface designs based on two different types of perceptual learning style.

Keywords: WeCWI, literacy discovery, language discovery, cognitive discovery, psychological discovery

Procedia PDF Downloads 561
3244 Critical Pedagogy and Ecoliteracy in the Teaching of Foreign Languages

Authors: Anita De Melo

Abstract:

Today we live in a crucial time of ecological crisis, of environmental catastrophes worldwide, and this scenario is, arrogantly, overlooked by powerful economic forces and their politics. Thus, a critical pedagogy that leads to action and that fosters ecoliteracy, environment education, is now inevitable, and it must become an integral part of the school curriculum across the disciplines, including the social sciences and the humanities. One of the most important contemporary and emerging movement of today is ecopedagogy, a movement that blends theory and ethics towards a curriculum that focus on an environmental education that will promote ecological justice, respect, and care by educating students to become planetary citizens. This paper aims, first, to emphasize the need for discussions and investigations regarding ecoliteracy within our field of teaching foreign languages, which will consider, among others, the of role language in stimulating sustainability, and the role of second language proficiency in fostering positive transnational dialogues conducive to fighting our current planetary crisis. Second, this paper suggests and discusses some critical ecopedagogical practices -- in the form of project-based learning, service-learning and environmental-oriented study abroad programs – apropos to ecoliteracy. These interdisciplinary projects can and should bring students in contact with communities speaking the target language, and such encounter would facilitate cultural exchanges and promote positive language proficiency whilst it would also give students the opportunity to work with finding ideas/projects to fight our current ecological catastrophe.

Keywords: critical pedagogy, ecoliteracy, ecopedagogy, planetary crisis

Procedia PDF Downloads 250
3243 Forecast Combination for Asset Classes: Insights on Market Efficiency and Arbitrage

Authors: Rodrigo Baggi Prieto Alvarez, Jorge Miguel Bravo

Abstract:

The Exchange-Traded Funds (ETFs) have transformed asset allocation, allowing investors to gain exposure to diverse asset classes with a single instrument. In turn, forecast combination models have emerged as advantageous methods for improving prediction accuracy. While the Efficient Market Hypothesis (EMH) posits that prices fluctuate randomly, making abnormal returns unattainable, empirical evidence reveals autocorrelation in stock returns, challenging the EMH's strict interpretation. This raises the question of whether econometric models, machine learning methods and forecast combinations can predict asset prices more effectively. Also, comparing forecasts with futures market prices may reveal potential arbitrage opportunities, offering insights into market inefficiencies. Using ETFs indices from January 1st, 2015, to September 30th, 2024, across equity markets (S&P 500, Russell 2000, MSCI Developed Markets and MSCI Emerging Markets), fixed income (7-10 Year Treasury Bond, Developed Markets Treasury Bond, Emerging Markets Treasury Bond and U.S. Corporate Bonds), commodity (Gold Shares ETF) and crypto (ProShares Bitcoin ETF), this paper tests the predictive accuracy of traditional econometric models (ARIMA, ETS), machine learning (SVM, Random Forest, XGBoost) and forecast combinations (ARIMA-SVR, ARIMA-ANN, Ridge Regression and LASSO). Preliminary results suggest that ensemble methods can indeed outperform simple models, indicating that combinations like the Ridge Regression and LASSO are superior to econometric and machine learning models individually. Also, prediction accuracy is better for fixed income ETFs, aligned with the lower volatility of these assets, while models show higher forecast error for crypto and equity ETFs. Finally, initial comparisons between forecasts and the futures market prices reveal potential inefficiencies, suggesting opportunities for spot-futures index arbitrage. Providing empirical evidence on the application of forecasting models to a significant group of financial assets, these findings contribute to discussions on market efficiency and highlight the role of ensemble methods in improving asset price predictability and portfolio management.

Keywords: ETF, asset prediction, forecast combination, EMH, spot-futures index arbitrage

Procedia PDF Downloads 4
3242 A Comparative Study on Deep Learning Models for Pneumonia Detection

Authors: Hichem Sassi

Abstract:

Pneumonia, being a respiratory infection, has garnered global attention due to its rapid transmission and relatively high mortality rates. Timely detection and treatment play a crucial role in significantly reducing mortality associated with pneumonia. Presently, X-ray diagnosis stands out as a reasonably effective method. However, the manual scrutiny of a patient's X-ray chest radiograph by a proficient practitioner usually requires 5 to 15 minutes. In situations where cases are concentrated, this places immense pressure on clinicians for timely diagnosis. Relying solely on the visual acumen of imaging doctors proves to be inefficient, particularly given the low speed of manual analysis. Therefore, the integration of artificial intelligence into the clinical image diagnosis of pneumonia becomes imperative. Additionally, AI recognition is notably rapid, with convolutional neural networks (CNNs) demonstrating superior performance compared to human counterparts in image identification tasks. To conduct our study, we utilized a dataset comprising chest X-ray images obtained from Kaggle, encompassing a total of 5216 training images and 624 test images, categorized into two classes: normal and pneumonia. Employing five mainstream network algorithms, we undertook a comprehensive analysis to classify these diseases within the dataset, subsequently comparing the results. The integration of artificial intelligence, particularly through improved network architectures, stands as a transformative step towards more efficient and accurate clinical diagnoses across various medical domains.

Keywords: deep learning, computer vision, pneumonia, models, comparative study

Procedia PDF Downloads 64
3241 The Effect of Self and Peer Assessment Activities in Second Language Writing: A Washback Effect Study on the Writing Growth during the Revision Phase in the Writing Process: Learners’ Perspective

Authors: Musbah Abdussayed

Abstract:

The washback effect refers to the influence of assessment on teaching and learning, and this washback effect can either be positive or negative. This study implemented, sequentially, self-assessment (SA) and peer assessment (PA) and examined the washback effect of self and peer assessment (SPA) activities on the writing growth during the revision phase in the writing process. Twenty advanced Arabic as a second language learners from a private school in the USA participated in the study. The participants composed and then revised a short Arabic story as a part of a midterm grade. Qualitative data was collected, analyzed, and synthesized from ten interviews with the learners and from the twenty learners’ post-reflective journals. The findings indicate positive washback effects on the learners’ writing growth. The PA activity enhanced descriptions and meaning, promoted creativity, and improved textual coherence, whereas the SA activity led to detecting editing issues. Furthermore, both SPA activities had washback effects in common, including helping the learners meet the writing genre conventions and developing metacognitive awareness. However, the findings also demonstrate negative washback effects on the learners’ attitudes during the revision phase in the writing process, including bias toward self-evaluation during the SA activity and reluctance to rate peers’ writing performance during the PA activity. The findings suggest that self-and peer assessment activities are essential teaching and learning tools that can be utilized sequentially to help learners tackle multiple writing areas during the revision phase in the writing process.

Keywords: self assessment, peer assessment, washback effect, second language writing, writing process

Procedia PDF Downloads 68
3240 Machine Learning Approach for Stress Detection Using Wireless Physical Activity Tracker

Authors: B. Padmaja, V. V. Rama Prasad, K. V. N. Sunitha, E. Krishna Rao Patro

Abstract:

Stress is a psychological condition that reduces the quality of sleep and affects every facet of life. Constant exposure to stress is detrimental not only for mind but also body. Nevertheless, to cope with stress, one should first identify it. This paper provides an effective method for the cognitive stress level detection by using data provided from a physical activity tracker device Fitbit. This device gathers people’s daily activities of food, weight, sleep, heart rate, and physical activities. In this paper, four major stressors like physical activities, sleep patterns, working hours and change in heart rate are used to assess the stress levels of individuals. The main motive of this system is to use machine learning approach in stress detection with the help of Smartphone sensor technology. Individually, the effect of each stressor is evaluated using logistic regression and then combined model is built and assessed using variants of ordinal logistic regression models like logit, probit and complementary log-log. Then the quality of each model is evaluated using Akaike Information Criterion (AIC) and probit is assessed as the more suitable model for our dataset. This system is experimented and evaluated in a real time environment by taking data from adults working in IT and other sectors in India. The novelty of this work lies in the fact that stress detection system should be less invasive as possible for the users.

Keywords: physical activity tracker, sleep pattern, working hours, heart rate, smartphone sensor

Procedia PDF Downloads 256
3239 A TgCNN-Based Surrogate Model for Subsurface Oil-Water Phase Flow under Multi-Well Conditions

Authors: Jian Li

Abstract:

The uncertainty quantification and inversion problems of subsurface oil-water phase flow usually require extensive repeated forward calculations for new runs with changed conditions. To reduce the computational time, various forms of surrogate models have been built. Related research shows that deep learning has emerged as an effective surrogate model, while most surrogate models with deep learning are purely data-driven, which always leads to poor robustness and abnormal results. To guarantee the model more consistent with the physical laws, a coupled theory-guided convolutional neural network (TgCNN) based surrogate model is built to facilitate computation efficiency under the premise of satisfactory accuracy. The model is a convolutional neural network based on multi-well reservoir simulation. The core notion of this proposed method is to bridge two separate blocks on top of an overall network. They underlie the TgCNN model in a coupled form, which reflects the coupling nature of pressure and water saturation in the two-phase flow equation. The model is driven by not only labeled data but also scientific theories, including governing equations, stochastic parameterization, boundary, and initial conditions, well conditions, and expert knowledge. The results show that the TgCNN-based surrogate model exhibits satisfactory accuracy and efficiency in subsurface oil-water phase flow under multi-well conditions.

Keywords: coupled theory-guided convolutional neural network, multi-well conditions, surrogate model, subsurface oil-water phase

Procedia PDF Downloads 86
3238 Localising the Alien: Language, Literature and Theory in the Indian Classroom

Authors: Asima Ranjan Parhi

Abstract:

English language teaching-learning in higher education departments in Indian and Asian contexts needs to be one of innovation and experimentation rather than rigid prescription. The communicative language teaching has been proposing the context to be of primary importance in this process. Today, English print and electronic media have flooded the market with plenty of material suitable to the classroom context. The entries are poetic, catchy and contain a deliberate method in them which could be utilized to teach not only English language but literature, literary terms and the theory of literature. The Bollywood movies, especially through their songs have been propagating a package which may be useful to teach language and even theory in the sub-continent. While investigating, one may be fascinated to see how such material in the body of media (print and electronic), movies and popular songs generate a data for our classroom in our context, thereby developing a mass language with huge pedagogical implications. Harping on the four skills of teaching and learning of a language in general and English language in particular appears stale and mechanical in a decontextualised, matter of fact classroom. So this discussion visualizes a model beyond these skills as well as the conventional theory, literature, language classroom practices in order to build up a systematic pattern stressing the factors responsible in the particular context, that of specific language, society and culture in tune with language-literature teaching. This study intends to examine certain catchy use of the language entries in mass media which could be in the direction of inviting more such investigations in the Asian context in order to develop a common platform of decolonized pedagogy.

Keywords: pedagogy, electronic media, Bollywood, decolonized, mass media

Procedia PDF Downloads 275
3237 Cosmetic Recommendation Approach Using Machine Learning

Authors: Shakila N. Senarath, Dinesh Asanka, Janaka Wijayanayake

Abstract:

The necessity of cosmetic products is arising to fulfill consumer needs of personality appearance and hygiene. A cosmetic product consists of various chemical ingredients which may help to keep the skin healthy or may lead to damages. Every chemical ingredient in a cosmetic product does not perform on every human. The most appropriate way to select a healthy cosmetic product is to identify the texture of the body first and select the most suitable product with safe ingredients. Therefore, the selection process of cosmetic products is complicated. Consumer surveys have shown most of the time, the selection process of cosmetic products is done in an improper way by consumers. From this study, a content-based system is suggested that recommends cosmetic products for the human factors. To such an extent, the skin type, gender and price range will be considered as human factors. The proposed system will be implemented by using Machine Learning. Consumer skin type, gender and price range will be taken as inputs to the system. The skin type of consumer will be derived by using the Baumann Skin Type Questionnaire, which is a value-based approach that includes several numbers of questions to derive the user’s skin type to one of the 16 skin types according to the Bauman Skin Type indicator (BSTI). Two datasets are collected for further research proceedings. The user data set was collected using a questionnaire given to the public. Those are the user dataset and the cosmetic dataset. Product details are included in the cosmetic dataset, which belongs to 5 different kinds of product categories (Moisturizer, Cleanser, Sun protector, Face Mask, Eye Cream). An alternate approach of TF-IDF (Term Frequency – Inverse Document Frequency) is applied to vectorize cosmetic ingredients in the generic cosmetic products dataset and user-preferred dataset. Using the IF-IPF vectors, each user-preferred products dataset and generic cosmetic products dataset can be represented as sparse vectors. The similarity between each user-preferred product and generic cosmetic product will be calculated using the cosine similarity method. For the recommendation process, a similarity matrix can be used. Higher the similarity, higher the match for consumer. Sorting a user column from similarity matrix in a descending order, the recommended products can be retrieved in ascending order. Even though results return a list of similar products, and since the user information has been gathered, such as gender and the price ranges for product purchasing, further optimization can be done by considering and giving weights for those parameters once after a set of recommended products for a user has been retrieved.

Keywords: content-based filtering, cosmetics, machine learning, recommendation system

Procedia PDF Downloads 134
3236 Chatbots as Language Teaching Tools for L2 English Learners

Authors: Feiying Wu

Abstract:

Chatbots are computer programs that attempt to engage a human in a dialogue, which originated in the 1960s with MIT's Eliza. However, they have become widespread more recently as advances in language technology have produced chatbots with increasing linguistic quality and sophistication, leading to their potential to serve as a tool for Computer-Assisted Language Learning(CALL). The aim of this article is to assess the feasibility of using two chatbots, Mitsuku and CleverBot, as pedagogical tools for learning English as a second language by stimulating L2 learners with distinct English proficiencies. Speaking of the input of stimulated learners, they are measured by AntWordProfiler to match the user's expected vocabulary proficiency. Totally, there are four chat sessions as each chatbot will converse with both beginners and advanced learners. For evaluation, it focuses on chatbots' responses from a linguistic standpoint, encompassing vocabulary and sentence levels. The vocabulary level is determined by the vocabulary range and the reaction to misspelled words. Grammatical accuracy and responsiveness to poorly formed sentences are assessed for the sentence level. In addition, the assessment of this essay sets 25% lexical and grammatical incorrect input to determine chatbots' corrective ability towards different linguistic forms. Based on statistical evidence and illustration of examples, despite the small sample size, neither Mitsuku nor CleverBot is ideal as educational tools based on their performance through word range, grammatical accuracy, topic range, and corrective feedback for incorrect words and sentences, but rather as a conversational tool for beginners of L2 English.

Keywords: chatbots, CALL, L2, corrective feedback

Procedia PDF Downloads 78
3235 Effects of the Supplementary for Understanding and Preventing Plagiarism on EFL Students’ Writing

Authors: Surichai Butcha, Dararat Khampusaen

Abstract:

As the Internet is recognized as a high potential and powerful educational tool to access sources of knowledge, plagiarism is an increasing unethical issue found in students’ writing. This paper is deriving from the 1st phase of an on-going study investigating the effects of the supplementary on citing sources on undergraduate students’ writing. The 40 participants were divided into 1 experimental group and 1 control group. Both groups were administered with a questionnaire on knowledge and an interview on attitude related to using sources in writing. Only the experimental group undertook the 4 lessons focusing on using outside sources and citing the original work (quoting, synthesizing, summarizing and paraphrasing) were delivered to them via e-learning tools throughout a semester. Participants were required to produce 4 writing tasks after each lesson. The results were concerned with types and factors on using outside sources in writing of Thai undergraduate EFL students from the survey. The interview results supported and clarified the survey result. In addition, the writing rubrics confirmed the types of plagiarism frequently occurred in students’ writing. The results revealed the types and factors on plagiarism including their perceptions on using the outside sources in their writing from the interview. The discussion shed the lights on cultural dimensions of plagiarism in student writing, roles of teachers, library, and university policy on the rate of plagiarism. Also, the findings promoted the awareness on ethics in writing and prevented the rate of potential unintentional plagiarism. Additionally, the results of this phase of study could lead to the appropriate contents to be considered for inclusion in the supplementary on using sources for writing for future research.

Keywords: citing source, EFL writing, e-learning, Internet, plagiarism

Procedia PDF Downloads 149
3234 A Computationally Intelligent Framework to Support Youth Mental Health in Australia

Authors: Nathaniel Carpenter

Abstract:

Web-enabled systems for supporting youth mental health management in Australia are pioneering in their field; however, with their success, these systems are experiencing exponential growth in demand which is straining an already stretched service. Supporting youth mental is critical as the lack of support is associated with significant and lasting negative consequences. To meet this growing demand, and provide critical support, investigations are needed on evaluating and improving existing online support services. Improvements should focus on developing frameworks capable of augmenting and scaling service provisions. There are few investigations informing best-practice frameworks when implementing e-mental health support systems for youth mental health; there are fewer which implement machine learning or artificially intelligent systems to facilitate the delivering of services. This investigation will use a case study methodology to highlight the design features which are important for systems to enable young people to self-manage their mental health. The investigation will also highlight the current information system challenges, to include challenges associated with service quality, provisioning, and scaling. This work will propose methods of meeting these challenges through improved design, service augmentation and automation, service quality, and through artificially intelligent inspired solutions. The results of this study will inform a framework for supporting youth mental health with intelligent and scalable web-enabled technologies to support an ever-growing user base.

Keywords: artificial intelligence, information systems, machine learning, youth mental health

Procedia PDF Downloads 110
3233 Protecting the Democracy of Children through Sustainable Risk Management: An Investigation into Risk Assessment and Nature-Based Play

Authors: Molly Gerrish

Abstract:

This work explores the physical, emotional, social, and cognitive risks and benefits related to nature-based teaching and highlights the importance of promoting a sustainable workforce within early childhood programs. Assessing and managing risks can help programs reimagine their approach to teaching, learning, recruitment, family connectivity, and staff motivation. The importance of staff sustainability and motivation/engagement related to social justice and the environment will be discussed. We will explore ways to manage fears and limitations faced by early childhood programs regarding nature experiences and risky play in a variety of locations using a lens of place-based learning. We will also examine the alignment of sustainability and social-emotional development, mental health supports, social awareness, and risk assessment. The work will discuss the varied perceptions of risk in diverse areas and the impact on the early childhood workforce. Motivational theory and compassion resiliency are hallmarks of both recruiting and retaining high-quality early childhood educators; the work will discuss how to balance programmatic constraints and healthy motivation for students and teachers while empowering individuals to advocate for their mental health and well-being. Finally, the work will highlight the positive impact of nature-based teaching practices and the overall benefit to young children and their educators.

Keywords: child’s rights, inclusion, nature-based education, risk assessment

Procedia PDF Downloads 60
3232 Integrated Free Space Optical Communication and Optical Sensor Network System with Artificial Intelligence Techniques

Authors: Yibeltal Chanie Manie, Zebider Asire Munyelet

Abstract:

5G and 6G technology offers enhanced quality of service with high data transmission rates, which necessitates the implementation of the Internet of Things (IoT) in 5G/6G architecture. In this paper, we proposed the integration of free space optical communication (FSO) with fiber sensor networks for IoT applications. Recently, free-space optical communications (FSO) are gaining popularity as an effective alternative technology to the limited availability of radio frequency (RF) spectrum. FSO is gaining popularity due to flexibility, high achievable optical bandwidth, and low power consumption in several applications of communications, such as disaster recovery, last-mile connectivity, drones, surveillance, backhaul, and satellite communications. Hence, high-speed FSO is an optimal choice for wireless networks to satisfy the full potential of 5G/6G technology, offering 100 Gbit/s or more speed in IoT applications. Moreover, machine learning must be integrated into the design, planning, and optimization of future optical wireless communication networks in order to actualize this vision of intelligent processing and operation. In addition, fiber sensors are important to achieve real-time, accurate, and smart monitoring in IoT applications. Moreover, we proposed deep learning techniques to estimate the strain changes and peak wavelength of multiple Fiber Bragg grating (FBG) sensors using only the spectrum of FBGs obtained from the real experiment.

Keywords: optical sensor, artificial Intelligence, Internet of Things, free-space optics

Procedia PDF Downloads 63
3231 The Effects of Teacher Efficacy, Instructional Leadership and Professional Learning Communities on Student Achievement in Literacy and Numeracy: A Look at Primary Schools within Sibu Division

Authors: Jarrod Sio Jyh Lih

Abstract:

This paper discusses the factors contributing to student achievement in literacy and numeracy in primary schools within Sibu division. The study involved 694 level 1 primary schoolteachers. Using descriptive statistics, the study observed high levels of practice for teacher efficacy, instructional leadership and professional learning communities (PLCs). The differences between gender, teaching experience and academic qualification were analyzed using the t-test and one-way analysis of variance (ANOVA). The study reported significant differences in respondent perceptions based on teaching experience vis-à-vis teacher efficacy. Here, the post hoc Tukey test revealed that efficaciousness grows with experience. A correlation test observed positive and significant correlations between all independent variables. Binary logistic regression was applied to predict the independent variables’ influence on student achievement. The findings revealed that a dimension of instructional leadership – ‘monitoring student progress’ - emerged as the best predictor of student achievement for literacy and numeracy. The result indicated the students were more than 4 times more likely to achieve the national key performance index for both literacy and numeracy when student progress was monitored. In conclusion, ‘monitoring student progress’ had a positive influence on students’ achievement for literacy and numeracy, hence making it a possible course of action for school heads. However, more comprehensive studies are needed to ascertain its consistency within the context of Malaysia.

Keywords: efficacy, instructional, literacy, numeracy

Procedia PDF Downloads 261
3230 AI-Powered Models for Real-Time Fraud Detection in Financial Transactions to Improve Financial Security

Authors: Shanshan Zhu, Mohammad Nasim

Abstract:

Financial fraud continues to be a major threat to financial institutions across the world, causing colossal money losses and undermining public trust. Fraud prevention techniques, based on hard rules, have become ineffective due to evolving patterns of fraud in recent times. Against such a background, the present study probes into distinct methodologies that exploit emergent AI-driven techniques to further strengthen fraud detection. We would like to compare the performance of generative adversarial networks and graph neural networks with other popular techniques, like gradient boosting, random forests, and neural networks. To this end, we would recommend integrating all these state-of-the-art models into one robust, flexible, and smart system for real-time anomaly and fraud detection. To overcome the challenge, we designed synthetic data and then conducted pattern recognition and unsupervised and supervised learning analyses on the transaction data to identify which activities were fishy. With the use of actual financial statistics, we compare the performance of our model in accuracy, speed, and adaptability versus conventional models. The results of this study illustrate a strong signal and need to integrate state-of-the-art, AI-driven fraud detection solutions into frameworks that are highly relevant to the financial domain. It alerts one to the great urgency that banks and related financial institutions must rapidly implement these most advanced technologies to continue to have a high level of security.

Keywords: AI-driven fraud detection, financial security, machine learning, anomaly detection, real-time fraud detection

Procedia PDF Downloads 42
3229 Investigation of the Opinions and Recommendations of Participants Related to Operating Room Nursing Certified Course Program

Authors: Zehra Gencel Efe, Fatma Susam Ozsayın, Satı Tas

Abstract:

Background and Aim: It is not possible to teach all the knowledge related to operating room nursing in the nursing education process. Certified courses are organized by the Ministry of Health to compensate the lack of postgraduate training and the theoretical and practical training needs of working nurses. In this study; It is aimed to investigate the participants’ opinions and recommendations attending the certified course of operating room nursing that organized in İKCU AtaturkTraining and Research Hospital. Method: Two operating room nursing courses were organized in 2016. The 1st Operating Room Nursing Certified Course Program was organized between March 07, 2016 and April 6, 2016and the 2nd Operating Room Nursing Certified Course Program was organized between 07 November 2016 - 06 December 2016 at the İKCU Ataturk Training and Research Hospital. The first program was accepted for 29 participants, the second program was accepted for 30 participants. In the collection of the data, the 'Operating Room Nursing Certified Training Program Evaluation Form', 'Operating Room Nursing Certified Training Program Theoretical Training Evaluation Form' were used. Three point Likert-type scale is used for responses in the 'Operating Room Nursing Certified Training Program Evaluation Form’ (1=verygood, 2=good, 3=poor). Data is collected in five areas related to training program, operation room practice, communication, responsibility, experiences of learning. Four point Likert-type scale is used for responses in the 'Operating Room Nursing Certified Training Program Theoretical Training Evaluation Form' (1=verysatisfied, 2=quitesatisfied, 3=satisfied, 4=dissatisfied). Data is collected in two areas include presentation and content. Data were analyzed with SPSS 16 program. Findings and Conclusion: It was found that 93,22% of participants were female in addition, 62,7% had bachelor degree. It was seen that 33,87% of the work group had 1-5 years of experience in their field. It was found that; 88% of trainees participating in the first group to the operating room nursing-certified course program stated the training program was very good, 12% of them stated the training program was good. Nobody was signed the ‘poor’ choice. 81% of the trainees who participated in the 2nd group to the operating room nursing-certified course program stated the training program was very good, 19% of them stated the training program was good. Nobody was signed the ‘poor’ choice. It was found that there was no meaningful difference between the achievement ratios of the trainees and the learning status of the trainees when compared with the t test in the groups with success level of the operating room nursing certified course program according to the learning status of the participants (p ˃ 0,05). The trainees noted that the course was satisfied with theoretical and practical steps but the support services (lunch, coffee breaks etc.) were in adequate.

Keywords: certified courses, nursing certified courses, operating room nursing, training program

Procedia PDF Downloads 216
3228 Technological Tool-Use as an Online Learner Strategy in a Synchronous Speaking Task

Authors: J. Knight, E. Barberà

Abstract:

Language learning strategies have been defined as thoughts and actions, consciously chosen and operationalized by language learners, to help them in carrying out a multiplicity of tasks from the very outset of learning to the most advanced levels of target language performance. While research in the field of Second Language Acquisition has focused on ‘good’ language learners, the effectiveness of strategy-use and orchestration by effective learners in face-to-face classrooms much less research has attended to learner strategies in online contexts, particular strategies in relation to technological tool use which can be part of a task design. In addition, much research on learner strategies and strategy use has been explored focusing on cognitive, attitudinal and metacognitive behaviour with less research focusing on the social aspect of strategies. This study focuses on how learners mediate with a technological tool designed to support synchronous spoken interaction and how this shape their spoken interaction in the opening of their talk. A case study approach is used incorporating notions from communities of practice theory to analyse and understand learner strategies of dyads carrying out a role play task. The study employs analysis of transcripts of spoken interaction in the openings of the talk along with log files of tool use. The study draws on results of previous studies pertaining to the same tool as a form of triangulation. Findings show how learners gain pre-task planning time through technological tool control. The strategies involving learners’ choices to enter and exit the tool shape their spoken interaction qualitatively, with some cases demonstrating long silences whilst others appearing to start the pedagogical task immediately. Who/what learners orientate to in the openings of the talk: an audience (i.e. the teacher), each other and/or screen-based signifiers in the opening moments of the talk also becomes a focus. The study highlights how tool use as a social practice should be considered a learning strategy in online contexts whereby different usages may be understood in the light of the more usual asynchronous social practices of the online community. The teachers’ role in the community is also problematised as the evaluator of the practices of that community. Results are pertinent for task design for synchronous speaking tasks. The use of community of practice theory supports an understanding of strategy use that involves both metacognition alongside social context revealing how tool-use strategies may need to be orally (socially) negotiated by learners and may also differ from an online language community.

Keywords: learner strategy, tool use, community of practice, speaking task

Procedia PDF Downloads 342
3227 Automated Weight Painting: Using Deep Neural Networks to Adjust 3D Mesh Skeletal Weights

Authors: John Gibbs, Benjamin Flanders, Dylan Pozorski, Weixuan Liu

Abstract:

Weight Painting–adjusting the influence a skeletal joint has on a given vertex in a character mesh–is an arduous and time con- suming part of the 3D animation pipeline. This process generally requires a trained technical animator and many hours of work to complete. Our skiNNer plug-in, which works within Autodesk’s Maya 3D animation software, uses Machine Learning and data pro- cessing techniques to create a deep neural network model that can accomplish the weight painting task in seconds rather than hours for bipedal quasi-humanoid character meshes. In order to create a properly trained network, a number of challenges were overcome, including curating an appropriately large data library, managing an arbitrary 3D mesh size, handling arbitrary skeletal architectures, accounting for extreme numeric values (most data points are near 0 or 1 for weight maps), and constructing an appropriate neural network model that can properly capture the high frequency alter- ation between high weight values (near 1.0) and low weight values (near 0.0). The arrived at neural network model is a cross between a traditional CNN, deep residual network, and fully dense network. The resultant network captures the unusually hard-edged features of a weight map matrix, and produces excellent results on many bipedal models.

Keywords: 3d animation, animation, character, rigging, skinning, weight painting, machine learning, artificial intelligence, neural network, deep neural network

Procedia PDF Downloads 273
3226 A Paradigm Model of Educational Policy Review Strategies to Develop Professional Schools

Authors: Farhad Shafiepour Motlagh, Narges Salehi

Abstract:

Purpose: The aim of the present study was a paradigm model of educational policy review strategies for the development of Professional schools in Iran. Research Methodology: The research method was based on Grounded theory. The statistical population included all articles of the ten years 2022-2010 and the method of sampling in a purposeful manner to the extent of theoretical saturation to 31 articles. For data analysis, open coding, axial coding and selective coding were used. Results: The results showed that causal conditions include social requirements (social expectations, educational justice, social justice); technology requirements (use of information and communication technology, use of new learning methods), educational requirements (development of educational territory, Development of educational tools and development of learning methods), contextual conditions including dual dimensions (motivational-psychological context, context of participation and cooperation), strategic conditions including (decentralization, delegation, organizational restructuring), intervention conditions (poor knowledge) Human resources, centralized system governance) and outcomes (school productivity, school professionalism, graduate entry into the labor market) were obtained. Conclusion: A review of educational policy is necessary to develop Iran's Professional schools, and this depends on decentralization, delegation, and, of course, empowerment of school principals.

Keywords: school productivity, professional schools, educational policy, paradigm

Procedia PDF Downloads 207
3225 Interpretation and Prediction of Geotechnical Soil Parameters Using Ensemble Machine Learning

Authors: Goudjil kamel, Boukhatem Ghania, Jlailia Djihene

Abstract:

This paper delves into the development of a sophisticated desktop application designed to calculate soil bearing capacity and predict limit pressure. Drawing from an extensive review of existing methodologies, the study meticulously examines various approaches employed in soil bearing capacity calculations, elucidating their theoretical foundations and practical applications. Furthermore, the study explores the burgeoning intersection of artificial intelligence (AI) and geotechnical engineering, underscoring the transformative potential of AI- driven solutions in enhancing predictive accuracy and efficiency.Central to the research is the utilization of cutting-edge machine learning techniques, including Artificial Neural Networks (ANN), XGBoost, and Random Forest, for predictive modeling. Through comprehensive experimentation and rigorous analysis, the efficacy and performance of each method are rigorously evaluated, with XGBoost emerging as the preeminent algorithm, showcasing superior predictive capabilities compared to its counterparts. The study culminates in a nuanced understanding of the intricate dynamics at play in geotechnical analysis, offering valuable insights into optimizing soil bearing capacity calculations and limit pressure predictions. By harnessing the power of advanced computational techniques and AI-driven algorithms, the paper presents a paradigm shift in the realm of geotechnical engineering, promising enhanced precision and reliability in civil engineering projects.

Keywords: limit pressure of soil, xgboost, random forest, bearing capacity

Procedia PDF Downloads 22
3224 Exploring Teachers’ Beliefs about Diagnostic Language Assessment Practices in a Large-Scale Assessment Program

Authors: Oluwaseun Ijiwade, Chris Davison, Kelvin Gregory

Abstract:

In Australia, like other parts of the world, the debate on how to enhance teachers using assessment data to inform teaching and learning of English as an Additional Language (EAL, Australia) or English as a Foreign Language (EFL, United States) have occupied the centre of academic scholarship. Traditionally, this approach was conceptualised as ‘Formative Assessment’ and, in recent times, ‘Assessment for Learning (AfL)’. The central problem is that teacher-made tests are limited in providing data that can inform teaching and learning due to variability of classroom assessments, which are hindered by teachers’ characteristics and assessment literacy. To address this concern, scholars in language education and testing have proposed a uniformed large-scale computer-based assessment program to meet the needs of teachers and promote AfL in language education. In Australia, for instance, the Victoria state government commissioned a large-scale project called 'Tools to Enhance Assessment Literacy (TEAL) for Teachers of English as an additional language'. As part of the TEAL project, a tool called ‘Reading and Vocabulary assessment for English as an Additional Language (RVEAL)’, as a diagnostic language assessment (DLA), was developed by language experts at the University of New South Wales for teachers in Victorian schools to guide EAL pedagogy in the classroom. Therefore, this study aims to provide qualitative evidence for understanding beliefs about the diagnostic language assessment (DLA) among EAL teachers in primary and secondary schools in Victoria, Australia. To realize this goal, this study raises the following questions: (a) How do teachers use large-scale assessment data for diagnostic purposes? (b) What skills do language teachers think are necessary for using assessment data for instruction in the classroom? and (c) What factors, if any, contribute to teachers’ beliefs about diagnostic assessment in a large-scale assessment? Semi-structured interview method was used to collect data from at least 15 professional teachers who were selected through a purposeful sampling. The findings from the resulting data analysis (thematic analysis) provide an understanding of teachers’ beliefs about DLA in a classroom context and identify how these beliefs are crystallised in language teachers. The discussion shows how the findings can be used to inform professional development processes for language teachers as well as informing important factor of teacher cognition in the pedagogic processes of language assessment. This, hopefully, will help test developers and testing organisations to align the outcome of this study with their test development processes to design assessment that can enhance AfL in language education.

Keywords: beliefs, diagnostic language assessment, English as an additional language, teacher cognition

Procedia PDF Downloads 199
3223 Challenges Faced by Teachers during Teaching with Developmental Disable Students at Primary Level in Lahore

Authors: Zikra Faiz, Nisar Abid, Muhammad Waqas

Abstract:

This study aim to examine the challenges faced by teachers during teaching to those students who are intellectually disable, suffering from autism spectrum disorder, learning disability, and ADHD at the primary level. The descriptive research design of quantitative approach was adopted to conduct this study; a cross-sectional survey method was used to collect data. The sample was comprised of 258 (43 male and 215 female) teachers who teach at special education institutes of Lahore district selected through proportionate stratified random sampling technique. Self-developed questionnaire was used which was comprised of 22 closed-ended items. Collected data were analyzed through descriptive and inferential statistical techniques by using Statistical Package for Social Sciences (SPSS) version 21. Results show that teachers faced problems during group activities, to handle bad behavior and different disabilities of students. It is concluded that there was a significant difference between male and female teachers perceptions about challenges faced during teaching with developmental disable students. Furthermore, there was a significant difference exist in the perceptions of teachers regarding challenges faced during teaching to students with developmental disabilities in term of teachers’ age and area of specialization. It is recommended that developmentally disable student require extra attention so that, teacher should trained through pre-service and in-service training to teach developmentally disabled students.

Keywords: intellectual disability, autism spectrum disorder, ADHD, learning disability

Procedia PDF Downloads 139
3222 Predictive Analytics Algorithms: Mitigating Elementary School Drop Out Rates

Authors: Bongs Lainjo

Abstract:

Educational institutions and authorities that are mandated to run education systems in various countries need to implement a curriculum that considers the possibility and existence of elementary school dropouts. This research focuses on elementary school dropout rates and the ability to replicate various predictive models carried out globally on selected Elementary Schools. The study was carried out by comparing the classical case studies in Africa, North America, South America, Asia and Europe. Some of the reasons put forward for children dropping out include the notion of being successful in life without necessarily going through the education process. Such mentality is coupled with a tough curriculum that does not take care of all students. The system has completely led to poor school attendance - truancy which continuously leads to dropouts. In this study, the focus is on developing a model that can systematically be implemented by school administrations to prevent possible dropout scenarios. At the elementary level, especially the lower grades, a child's perception of education can be easily changed so that they focus on the better future that their parents desire. To deal effectively with the elementary school dropout problem, strategies that are put in place need to be studied and predictive models are installed in every educational system with a view to helping prevent an imminent school dropout just before it happens. In a competency-based curriculum that most advanced nations are trying to implement, the education systems have wholesome ideas of learning that reduce the rate of dropout.

Keywords: elementary school, predictive models, machine learning, risk factors, data mining, classifiers, dropout rates, education system, competency-based curriculum

Procedia PDF Downloads 175
3221 Biofilm Text Classifiers Developed Using Natural Language Processing and Unsupervised Learning Approach

Authors: Kanika Gupta, Ashok Kumar

Abstract:

Biofilms are dense, highly hydrated cell clusters that are irreversibly attached to a substratum, to an interface or to each other, and are embedded in a self-produced gelatinous matrix composed of extracellular polymeric substances. Research in biofilm field has become very significant, as biofilm has shown high mechanical resilience and resistance to antibiotic treatment and constituted as a significant problem in both healthcare and other industry related to microorganisms. The massive information both stated and hidden in the biofilm literature are growing exponentially therefore it is not possible for researchers and practitioners to automatically extract and relate information from different written resources. So, the current work proposes and discusses the use of text mining techniques for the extraction of information from biofilm literature corpora containing 34306 documents. It is very difficult and expensive to obtain annotated material for biomedical literature as the literature is unstructured i.e. free-text. Therefore, we considered unsupervised approach, where no annotated training is necessary and using this approach we developed a system that will classify the text on the basis of growth and development, drug effects, radiation effects, classification and physiology of biofilms. For this, a two-step structure was used where the first step is to extract keywords from the biofilm literature using a metathesaurus and standard natural language processing tools like Rapid Miner_v5.3 and the second step is to discover relations between the genes extracted from the whole set of biofilm literature using pubmed.mineR_v1.0.11. We used unsupervised approach, which is the machine learning task of inferring a function to describe hidden structure from 'unlabeled' data, in the above-extracted datasets to develop classifiers using WinPython-64 bit_v3.5.4.0Qt5 and R studio_v0.99.467 packages which will automatically classify the text by using the mentioned sets. The developed classifiers were tested on a large data set of biofilm literature which showed that the unsupervised approach proposed is promising as well as suited for a semi-automatic labeling of the extracted relations. The entire information was stored in the relational database which was hosted locally on the server. The generated biofilm vocabulary and genes relations will be significant for researchers dealing with biofilm research, making their search easy and efficient as the keywords and genes could be directly mapped with the documents used for database development.

Keywords: biofilms literature, classifiers development, text mining, unsupervised learning approach, unstructured data, relational database

Procedia PDF Downloads 170
3220 Electricity Price Forecasting: A Comparative Analysis with Shallow-ANN and DNN

Authors: Fazıl Gökgöz, Fahrettin Filiz

Abstract:

Electricity prices have sophisticated features such as high volatility, nonlinearity and high frequency that make forecasting quite difficult. Electricity price has a volatile and non-random character so that, it is possible to identify the patterns based on the historical data. Intelligent decision-making requires accurate price forecasting for market traders, retailers, and generation companies. So far, many shallow-ANN (artificial neural networks) models have been published in the literature and showed adequate forecasting results. During the last years, neural networks with many hidden layers, which are referred to as DNN (deep neural networks) have been using in the machine learning community. The goal of this study is to investigate electricity price forecasting performance of the shallow-ANN and DNN models for the Turkish day-ahead electricity market. The forecasting accuracy of the models has been evaluated with publicly available data from the Turkish day-ahead electricity market. Both shallow-ANN and DNN approach would give successful result in forecasting problems. Historical load, price and weather temperature data are used as the input variables for the models. The data set includes power consumption measurements gathered between January 2016 and December 2017 with one-hour resolution. In this regard, forecasting studies have been carried out comparatively with shallow-ANN and DNN models for Turkish electricity markets in the related time period. The main contribution of this study is the investigation of different shallow-ANN and DNN models in the field of electricity price forecast. All models are compared regarding their MAE (Mean Absolute Error) and MSE (Mean Square) results. DNN models give better forecasting performance compare to shallow-ANN. Best five MAE results for DNN models are 0.346, 0.372, 0.392, 0,402 and 0.409.

Keywords: deep learning, artificial neural networks, energy price forecasting, turkey

Procedia PDF Downloads 292
3219 Communication Tools Used in Teaching and Their Effects: An Empirical Study on the T. C. Selcuk University Samples

Authors: Sedat Simsek, Tugay Arat

Abstract:

Today's communication concept, which has a great revolution with the printing press which has been found by Gutenberg, has no boundary thanks to advanced communication devices and the internet. It is possible to take advantage in many areas, such as from medicine to social sciences or from mathematics to education, from the computers that was first produced for the purpose of military services. The use of these developing technologies in the field of education has created a great vision changes in both training and having education. Materials, which can be considered as basic communication resources and used in traditional education has begun to lose its significance, and some technologies have begun to replace them such as internet, computers, smart boards, projection devices and mobile phone. On the other hand, the programs and applications used in these technologies have also been developed. University students use virtual books instead of the traditional printed book, use cell phones instead of note books, use the internet and virtual databases instead of the library to research. They even submit their homework with interactive methods rather than printed materials. The traditional education system, these technologies, which increase productivity, have brought a new dimension to education. The aim of this study is to determine the influence of technologies in the learning process of students and to find whether is there any similarities and differences that arise from the their faculty that they have been educated and and their learning process. In addition to this, it is aimed to determine the level of ICT usage of students studying at the university level. In this context, the advantages and conveniences of the technology used by students are also scrutinized. In this study, we used surveys to collect data. The data were analyzed by using SPSS 16 statistical program with the appropriate testing.

Keywords: education, communication technologies, role of technology, teaching

Procedia PDF Downloads 303
3218 Identification of Damage Mechanisms in Interlock Reinforced Composites Using a Pattern Recognition Approach of Acoustic Emission Data

Authors: M. Kharrat, G. Moreau, Z. Aboura

Abstract:

The latest advances in the weaving industry, combined with increasingly sophisticated means of materials processing, have made it possible to produce complex 3D composite structures. Mainly used in aeronautics, composite materials with 3D architecture offer better mechanical properties than 2D reinforced composites. Nevertheless, these materials require a good understanding of their behavior. Because of the complexity of such materials, the damage mechanisms are multiple, and the scenario of their appearance and evolution depends on the nature of the exerted solicitations. The AE technique is a well-established tool for discriminating between the damage mechanisms. Suitable sensors are used during the mechanical test to monitor the structural health of the material. Relevant AE-features are then extracted from the recorded signals, followed by a data analysis using pattern recognition techniques. In order to better understand the damage scenarios of interlock composite materials, a multi-instrumentation was set-up in this work for tracking damage initiation and development, especially in the vicinity of the first significant damage, called macro-damage. The deployed instrumentation includes video-microscopy, Digital Image Correlation, Acoustic Emission (AE) and micro-tomography. In this study, a multi-variable AE data analysis approach was developed for the discrimination between the different signal classes representing the different emission sources during testing. An unsupervised classification technique was adopted to perform AE data clustering without a priori knowledge. The multi-instrumentation and the clustered data served to label the different signal families and to build a learning database. This latter is useful to construct a supervised classifier that can be used for automatic recognition of the AE signals. Several materials with different ingredients were tested under various solicitations in order to feed and enrich the learning database. The methodology presented in this work was useful to refine the damage threshold for the new generation materials. The damage mechanisms around this threshold were highlighted. The obtained signal classes were assigned to the different mechanisms. The isolation of a 'noise' class makes it possible to discriminate between the signals emitted by damages without resorting to spatial filtering or increasing the AE detection threshold. The approach was validated on different material configurations. For the same material and the same type of solicitation, the identified classes are reproducible and little disturbed. The supervised classifier constructed based on the learning database was able to predict the labels of the classified signals.

Keywords: acoustic emission, classifier, damage mechanisms, first damage threshold, interlock composite materials, pattern recognition

Procedia PDF Downloads 155