Search results for: computer game-based learning
5715 Exploring the Impact of Artificial Intelligence (AI) in the Context of English as a Foreign Language (EFL): A Comprehensive Bibliometric Study
Authors: Kate Benedicta Amenador, Dianjian Wang, Bright Nkrumah
Abstract:
This extensive bibliometric study explores the dynamic influence of artificial intelligence in the field of English as a Foreign Language (EFL) between 2012 and 2024. The study, which examined 4,500 articles from Google Scholar, Modern Language Association Linguistics Abstracts, Web of Science, Scopus, Researchgate, and library genesis databases, indicates that AI integration in EFL is on the rise. This notable increase is ascribed to a variety of transformative events, including increased academic funding for higher education and the COVID-19 epidemic. The results of the study identify leading contributors, prominent authors, publishers and sources, with the United States, China and the United Kingdom emerging as key contributors. The co-occurrence analysis of key terms reveals five clusters highlighting patterns in AI-enhanced language instruction and learning, including evaluation strategies, educational technology, learning motivation, EFL teaching aspects, and learner feedback. The study also discusses the impact of various AIs in enhancing EFL writing skills with software such as Grammarly, Quilbot, and Chatgpt. The current study recognizes limitations in database selection and linguistic constraints. Nevertheless, the results provide useful insights for educators, researchers and policymakers, inspiring and guiding a cross-disciplinary collaboration and creative pedagogical techniques and approaches to teaching and learning in the future.Keywords: artificial intelligence, bibliometrics study, VOSviewer visualization, English as a foreign language
Procedia PDF Downloads 335714 Machine Learning Models for the Prediction of Heating and Cooling Loads of a Residential Building
Authors: Aaditya U. Jhamb
Abstract:
Due to the current energy crisis that many countries are battling, energy-efficient buildings are the subject of extensive research in the modern technological era because of growing worries about energy consumption and its effects on the environment. The paper explores 8 factors that help determine energy efficiency for a building: (relative compactness, surface area, wall area, roof area, overall height, orientation, glazing area, and glazing area distribution), with Tsanas and Xifara providing a dataset. The data set employed 768 different residential building models to anticipate heating and cooling loads with a low mean squared error. By optimizing these characteristics, machine learning algorithms may assess and properly forecast a building's heating and cooling loads, lowering energy usage while increasing the quality of people's lives. As a result, the paper studied the magnitude of the correlation between these input factors and the two output variables using various statistical methods of analysis after determining which input variable was most closely associated with the output loads. The most conclusive model was the Decision Tree Regressor, which had a mean squared error of 0.258, whilst the least definitive model was the Isotonic Regressor, which had a mean squared error of 21.68. This paper also investigated the KNN Regressor and the Linear Regression, which had to mean squared errors of 3.349 and 18.141, respectively. In conclusion, the model, given the 8 input variables, was able to predict the heating and cooling loads of a residential building accurately and precisely.Keywords: energy efficient buildings, heating load, cooling load, machine learning models
Procedia PDF Downloads 965713 Studying the Relationship Between Washback Effects of IELTS Test on Iranian Language Teachers, Teaching Strategies and Candidates
Authors: Afsaneh Jasmine Majidi
Abstract:
Language testing is an important part of language teaching experience and language learning process as it presents assessment strategies for teachers to evaluate the efficiency of teaching and for learners to examine their outcomes. However, language testing is demanding and challenging because it should provide the opportunity for proper and objective decision. In addition to all the efforts test designers put to design valid and reliable tests, there are some other determining factors which are even more complex and complicated. These factors affect the educational system, individuals, and society, and the impact of the tests vary according to the scope of the test. Seemingly, the impact of a simple classroom assessment is not the same as that of high stake tests such as International English Language Testing System (IELTS). As the importance of the test increases, it affects wider domain. Accordingly, the impacts of high stake tests are reflected not only in teaching, learning strategies but also in society. Testing experts use the term ‘washback’ or ‘impact’ to define the different effects of a test on teaching, learning, and community. This paper first looks at the theoretical background of ‘washback’ and ‘impact’ in language testing by reviewing of relevant literature in the field and then investigates washback effects of IELTS test of on Iranian IELTS teachers and students. The study found significant relationship between the washback effect of IELTS test and teaching strategies of Iranian IELTS teachers as well as performance of Iranian IELTS candidates and their community.Keywords: high stake tests, IELTS, Iranian Candidates, language testing, test impact, washback
Procedia PDF Downloads 3275712 Mining Educational Data to Support Students’ Major Selection
Authors: Kunyanuth Kularbphettong, Cholticha Tongsiri
Abstract:
This paper aims to create the model for student in choosing an emphasized track of student majoring in computer science at Suan Sunandha Rajabhat University. The objective of this research is to develop the suggested system using data mining technique to analyze knowledge and conduct decision rules. Such relationships can be used to demonstrate the reasonableness of student choosing a track as well as to support his/her decision and the system is verified by experts in the field. The sampling is from student of computer science based on the system and the questionnaire to see the satisfaction. The system result is found to be satisfactory by both experts and student as well.Keywords: data mining technique, the decision support system, knowledge and decision rules, education
Procedia PDF Downloads 4235711 A Book Review of Inside the Battle of Algiers, by Zohra Drif: A Thematic Analysis on Women’s Agency
Authors: W. Zekri
Abstract:
This paper explores Zohra Drif’s memoir, Inside the Battle of Algiers, which narrates her desires as a student to become a revolutionary activist. She exemplified, in her narrative, the different roles, she and her fellows performed as combatants in the Casbah during the Algerian Revolution 1954-1962. This book review aims to evaluate the concept of women’s agency through education and language learning, and its impact on empowering women’s desires. Close-reading method and thematic analysis are used to explore the text. The analysis identified themes that refine the meaning of agency which are social and cultural supports, education, and language proficiency. These themes aim to contribute to the representation in Inside the Battle of Algiers of a woman guerrilla who engaged herself to perform national acts of resistance.Keywords: agency, education, learning, women
Procedia PDF Downloads 1765710 Predicting Data Center Resource Usage Using Quantile Regression to Conserve Energy While Fulfilling the Service Level Agreement
Authors: Ahmed I. Alutabi, Naghmeh Dezhabad, Sudhakar Ganti
Abstract:
Data centers have been growing in size and dema nd continuously in the last two decades. Planning for the deployment of resources has been shallow and always resorted to over-provisioning. Data center operators try to maximize the availability of their services by allocating multiple of the needed resources. One resource that has been wasted, with little thought, has been energy. In recent years, programmable resource allocation has paved the way to allow for more efficient and robust data centers. In this work, we examine the predictability of resource usage in a data center environment. We use a number of models that cover a wide spectrum of machine learning categories. Then we establish a framework to guarantee the client service level agreement (SLA). Our results show that using prediction can cut energy loss by up to 55%.Keywords: machine learning, artificial intelligence, prediction, data center, resource allocation, green computing
Procedia PDF Downloads 1085709 Developing an Out-of-Distribution Generalization Model Selection Framework through Impurity and Randomness Measurements and a Bias Index
Authors: Todd Zhou, Mikhail Yurochkin
Abstract:
Out-of-distribution (OOD) detection is receiving increasing amounts of attention in the machine learning research community, boosted by recent technologies, such as autonomous driving and image processing. This newly-burgeoning field has called for the need for more effective and efficient methods for out-of-distribution generalization methods. Without accessing the label information, deploying machine learning models to out-of-distribution domains becomes extremely challenging since it is impossible to evaluate model performance on unseen domains. To tackle this out-of-distribution detection difficulty, we designed a model selection pipeline algorithm and developed a model selection framework with different impurity and randomness measurements to evaluate and choose the best-performing models for out-of-distribution data. By exploring different randomness scores based on predicted probabilities, we adopted the out-of-distribution entropy and developed a custom-designed score, ”CombinedScore,” as the evaluation criterion. This proposed score was created by adding labeled source information into the judging space of the uncertainty entropy score using harmonic mean. Furthermore, the prediction bias was explored through the equality of opportunity violation measurement. We also improved machine learning model performance through model calibration. The effectiveness of the framework with the proposed evaluation criteria was validated on the Folktables American Community Survey (ACS) datasets.Keywords: model selection, domain generalization, model fairness, randomness measurements, bias index
Procedia PDF Downloads 1245708 A Digital Environment for Developing Mathematical Abilities in Children with Autism Spectrum Disorder
Authors: M. Isabel Santos, Ana Breda, Ana Margarida Almeida
Abstract:
Research on academic abilities of individuals with autism spectrum disorder (ASD) underlines the importance of mathematics interventions. Yet the proposal of digital applications for children and youth with ASD continues to attract little attention, namely, regarding the development of mathematical reasoning, being the use of the digital technologies an area of great interest for individuals with this disorder and its use is certainly a facilitative strategy in the development of their mathematical abilities. The use of digital technologies can be an effective way to create innovative learning opportunities to these students and to develop creative, personalized and constructive environments, where they can develop differentiated abilities. The children with ASD often respond well to learning activities involving information presented visually. In this context, we present the digital Learning Environment on Mathematics for Autistic children (LEMA) that was a research project conducive to a PhD in Multimedia in Education and was developed by the Thematic Line Geometrix, located in the Department of Mathematics, in a collaboration effort with DigiMedia Research Center, of the Department of Communication and Art (University of Aveiro, Portugal). LEMA is a digital mathematical learning environment which activities are dynamically adapted to the user’s profile, towards the development of mathematical abilities of children aged 6–12 years diagnosed with ASD. LEMA has already been evaluated with end-users (both students and teacher’s experts) and based on the analysis of the collected data readjustments were made, enabling the continuous improvement of the prototype, namely considering the integration of universal design for learning (UDL) approaches, which are of most importance in ASD, due to its heterogeneity. The learning strategies incorporated in LEMA are: (i) provide options to custom choice of math activities, according to user’s profile; (ii) integrates simple interfaces with few elements, presenting only the features and content needed for the ongoing task; (iii) uses a simple visual and textual language; (iv) uses of different types of feedbacks (auditory, visual, positive/negative reinforcement, hints with helpful instructions including math concept definitions, solved math activities using split and easier tasks and, finally, the use of videos/animations that show a solution to the proposed activity); (v) provides information in multiple representation, such as text, video, audio and image for better content and vocabulary understanding in order to stimulate, motivate and engage users to mathematical learning, also helping users to focus on content; (vi) avoids using elements that distract or interfere with focus and attention; (vii) provides clear instructions and orientation about tasks to ease the user understanding of the content and the content language, in order to stimulate, motivate and engage the user; and (viii) uses buttons, familiarly icons and contrast between font and background. Since these children may experience little sensory tolerance and may have an impaired motor skill, besides the user to have the possibility to interact with LEMA through the mouse (point and click with a single button), the user has the possibility to interact with LEMA through Kinect device (using simple gesture moves).Keywords: autism spectrum disorder, digital technologies, inclusion, mathematical abilities, mathematical learning activities
Procedia PDF Downloads 1165707 Effect of Personality Traits on Classification of Political Orientation
Authors: Vesile Evrim, Aliyu Awwal
Abstract:
Today as in the other domains, there are an enormous number of political transcripts available in the Web which is waiting to be mined and used for various purposes such as statistics and recommendations. Therefore, automatically determining the political orientation on these transcripts becomes crucial. The methodologies used by machine learning algorithms to do the automatic classification are based on different features such as Linguistic. Considering the ideology differences between Liberals and Conservatives, in this paper, the effect of Personality Traits on political orientation classification is studied. This is done by considering the correlation between LIWC features and the BIG Five Personality Traits. Several experiments are conducted on Convote U.S. Congressional-Speech dataset with seven benchmark classification algorithms. The different methodologies are applied on selecting different feature sets that constituted by 8 to 64 varying number of features. While Neuroticism is obtained to be the most differentiating personality trait on classification of political polarity, when its top 10 representative features are combined with several classification algorithms, it outperformed the results presented in previous research.Keywords: politics, personality traits, LIWC, machine learning
Procedia PDF Downloads 4955706 Predicting Personality and Psychological Distress Using Natural Language Processing
Authors: Jihee Jang, Seowon Yoon, Gaeun Son, Minjung Kang, Joon Yeon Choeh, Kee-Hong Choi
Abstract:
Background: Self-report multiple choice questionnaires have been widely utilized to quantitatively measure one’s personality and psychological constructs. Despite several strengths (e.g., brevity and utility), self-report multiple-choice questionnaires have considerable limitations in nature. With the rise of machine learning (ML) and Natural language processing (NLP), researchers in the field of psychology are widely adopting NLP to assess psychological constructs to predict human behaviors. However, there is a lack of connections between the work being performed in computer science and that psychology due to small data sets and unvalidated modeling practices. Aims: The current article introduces the study method and procedure of phase II, which includes the interview questions for the five-factor model (FFM) of personality developed in phase I. This study aims to develop the interview (semi-structured) and open-ended questions for the FFM-based personality assessments, specifically designed with experts in the field of clinical and personality psychology (phase 1), and to collect the personality-related text data using the interview questions and self-report measures on personality and psychological distress (phase 2). The purpose of the study includes examining the relationship between natural language data obtained from the interview questions, measuring the FFM personality constructs, and psychological distress to demonstrate the validity of the natural language-based personality prediction. Methods: The phase I (pilot) study was conducted on fifty-nine native Korean adults to acquire the personality-related text data from the interview (semi-structured) and open-ended questions based on the FFM of personality. The interview questions were revised and finalized with the feedback from the external expert committee, consisting of personality and clinical psychologists. Based on the established interview questions, a total of 425 Korean adults were recruited using a convenience sampling method via an online survey. The text data collected from interviews were analyzed using natural language processing. The results of the online survey, including demographic data, depression, anxiety, and personality inventories, were analyzed together in the model to predict individuals’ FFM of personality and the level of psychological distress (phase 2).Keywords: personality prediction, psychological distress prediction, natural language processing, machine learning, the five-factor model of personality
Procedia PDF Downloads 795705 Investigation of Topic Modeling-Based Semi-Supervised Interpretable Document Classifier
Authors: Dasom Kim, William Xiu Shun Wong, Yoonjin Hyun, Donghoon Lee, Minji Paek, Sungho Byun, Namgyu Kim
Abstract:
There have been many researches on document classification for classifying voluminous documents automatically. Through document classification, we can assign a specific category to each unlabeled document on the basis of various machine learning algorithms. However, providing labeled documents manually requires considerable time and effort. To overcome the limitations, the semi-supervised learning which uses unlabeled document as well as labeled documents has been invented. However, traditional document classifiers, regardless of supervised or semi-supervised ones, cannot sufficiently explain the reason or the process of the classification. Thus, in this paper, we proposed a methodology to visualize major topics and class components of each document. We believe that our methodology for visualizing topics and classes of each document can enhance the reliability and explanatory power of document classifiers.Keywords: data mining, document classifier, text mining, topic modeling
Procedia PDF Downloads 4035704 Proactive Approach to Innovation Management
Authors: Andrus Pedai, Igor Astrov
Abstract:
The focus of this paper is to compare common approaches for Systems of Innovation (SI) and identify proactive alternatives for driving the innovation. Proactive approaches will also consider short and medium term perspectives with developments in the field of Computer Technology and Artificial Intelligence. Concerning computer technology and large connected information systems, it is reasonable to predict that during current or the next century, intelligence and innovation will be separated from the constraints of human-driven management. After this happens, humans will no longer be driving the innovation and there is possibility that SI for new intelligent systems will set its own targets and exclude humans. Over long time scale, these developments could result in a scenario, which will lead to the development of larger, cross galactic (universal) proactive SI and Intelligence.Keywords: artificial intelligence, DARPA, Moore’s law, proactive innovation, singularity, systems of innovation
Procedia PDF Downloads 4795703 Play-Based Early Education and Teachers’ Professional Development: Impact on Vulnerable Children
Authors: Chirine Dannaoui, Maya Antoun
Abstract:
This paper explores the intricate dynamics of play-based early childhood education (ECE) and the impact of professional development on teachers implementing play-based pedagogy, particularly in the context of vulnerable Syrian refugee children in Lebanon. By utilizing qualitative methodologies, including classroom observations and in-depth interviews with five early childhood educators and a field manager, this study delves into the challenges and transformations experienced by teachers in adopting play-based learning strategies. The research unveils the critical role of continuous and context-specific professional development in empowering teachers to implement play-based pedagogies effectively. When appropriately supported, it emphasizes how such educational approaches significantly enhance children's cognitive, social, and emotional development in crisis-affected environments. Key findings indicate that despite diverse educational backgrounds, teachers show considerable growth in their pedagogical skills through targeted professional development. This growth is vital for fostering a learning environment where vulnerable children can thrive, particularly in humanitarian settings. The paper also addresses educators' challenges, including adapting to play-based methodologies, resource limitations, and balancing curricular requirements with the need for holistic child development. This study contributes to the discourse on early childhood education in crisis contexts, emphasizing the need for sustainable, well-structured professional development programs. It underscores the potential of play-based learning to bridge educational gaps and contribute to the healing process of children facing calamity. The study highlights significant implications for policymakers, educators, schools, and not-for-profit organizations engaged in early childhood education in humanitarian contexts, stressing the importance of investing in teacher capacity and curriculum reform to enhance the quality of education for children in general and vulnerable ones in particular.Keywords: play-based learning, professional development, vulnerable children, early childhood education
Procedia PDF Downloads 605702 Automatic Classification for the Degree of Disc Narrowing from X-Ray Images Using CNN
Authors: Kwangmin Joo
Abstract:
Automatic detection of lumbar vertebrae and classification method is proposed for evaluating the degree of disc narrowing. Prior to classification, deep learning based segmentation is applied to detect individual lumbar vertebra. M-net is applied to segment five lumbar vertebrae and fine-tuning segmentation is employed to improve the accuracy of segmentation. Using the features extracted from previous step, clustering technique, k-means clustering, is applied to estimate the degree of disc space narrowing under four grade scoring system. As preliminary study, techniques proposed in this research could help building an automatic scoring system to diagnose the severity of disc narrowing from X-ray images.Keywords: Disc space narrowing, Degenerative disc disorders, Deep learning based segmentation, Clustering technique
Procedia PDF Downloads 1255701 A Custom Convolutional Neural Network with Hue, Saturation, Value Color for Malaria Classification
Authors: Ghazala Hcini, Imen Jdey, Hela Ltifi
Abstract:
Malaria disease should be considered and handled as a potential restorative catastrophe. One of the most challenging tasks in the field of microscopy image processing is due to differences in test design and vulnerability of cell classifications. In this article, we focused on applying deep learning to classify patients by identifying images of infected and uninfected cells. We performed multiple forms, counting a classification approach using the Hue, Saturation, Value (HSV) color space. HSV is used since of its superior ability to speak to image brightness; at long last, for classification, a convolutional neural network (CNN) architecture is created. Clusters of focus were used to deliver the classification. The highlights got to be forbidden, and a few more clamor sorts are included in the information. The suggested method has a precision of 99.79%, a recall value of 99.55%, and provides 99.96% accuracy.Keywords: deep learning, convolutional neural network, image classification, color transformation, HSV color, malaria diagnosis, malaria cells images
Procedia PDF Downloads 895700 Identification of Breast Anomalies Based on Deep Convolutional Neural Networks and K-Nearest Neighbors
Authors: Ayyaz Hussain, Tariq Sadad
Abstract:
Breast cancer (BC) is one of the widespread ailments among females globally. The early prognosis of BC can decrease the mortality rate. Exact findings of benign tumors can avoid unnecessary biopsies and further treatments of patients under investigation. However, due to variations in images, it is a tough job to isolate cancerous cases from normal and benign ones. The machine learning technique is widely employed in the classification of BC pattern and prognosis. In this research, a deep convolution neural network (DCNN) called AlexNet architecture is employed to get more discriminative features from breast tissues. To achieve higher accuracy, K-nearest neighbor (KNN) classifiers are employed as a substitute for the softmax layer in deep learning. The proposed model is tested on a widely used breast image database called MIAS dataset for experimental purposes and achieved 99% accuracy.Keywords: breast cancer, DCNN, KNN, mammography
Procedia PDF Downloads 1365699 Simulating an Interprofessional Hospital Day Shift: A Student Interprofessional (IP) Collaborative Learning Activity
Authors: Fiona Jensen, Barb Goodwin, Nancy Kleiman, Rhonda Usunier
Abstract:
Background: Clinical simulation is now a common component in many health profession curricula in preparation for clinical practice. In the Rady Faculty of Health Sciences (RFHS) college leads in simulation and interprofessional (IP) education, planned an eight hour simulated hospital day shift, where seventy students from six health professions across two campuses, learned with each other in a safe, realistic environment. Learning about interprofessional collaboration, an expected competency for many health professions upon graduation, was a primary focus of the simulation event. Method: Faculty representatives from the Colleges of Nursing, Medicine, Pharmacy and Rehabilitation Sciences (Physical Therapy, Occupation Therapy, Respiratory Therapy) and Pharmacy worked together to plan the IP event in a simulation facility in the College of Nursing. Each college provided a faculty mentor to guide the same profession students. Students were placed in interprofessional teams consisting of a nurse, physician, pharmacist, and then sharing respiratory, occupational, and physical therapists across the team depending on the needs of the patients. Eight patient scenarios were role played by health profession students, who had been provided with their patient’s story shortly before the event. Each team was guided by a facilitator. Results and Outcomes: On the morning of the event, all students gathered in a large group to meet mentors and facilitators and have a brief overview of the six competencies for effective collaboration and the session objectives. The students assuming their same profession roles were provided with their patient’s chart at the beginning of the shift, met with their team, and then completed professional specific assessments. Shortly into the shift, IP team rounds began, facilitated by the team facilitator. During the shift, each patient role-played a spontaneous health incident, which required collaboration between the IP team members for assessment and management. The afternoon concluded with team rounds, a collaborative management plan, and a facilitated de-brief. Conclusions: During the de-brief sessions, students responded to set questions related to the session learning objectives and expressed many positive learning moments. We believe that we have a sustainable simulation IP collaborative learning opportunity, which can be embedded into curricula, and has the capacity to grow to include more health profession faculties and students. Opportunities are being explored in the RFHS at the administrative level, to offer this event more frequently in the academic year to reach more students. In addition, a formally structured event evaluation tool would provide important feedback and inform the qualitative feedback to event organizers and the colleges about the significance of the simulation event to student learning.Keywords: simulation, collaboration, teams, interprofessional
Procedia PDF Downloads 1315698 TechWhiz: Empowering Deaf Students through Inclusive Education
Authors: Paula Escudeiro, Nuno Escudeiro, Márcia Campos, Francisca Escudeiro
Abstract:
In today's world, technical and scientific knowledge plays a vital role in education, research, and employment. Deaf students face unique challenges in educational settings, particularly when it comes to understanding technical and scientific terminology. The reliance on written and spoken languages can create barriers for deaf individuals who primarily communicate using sign language. This lack of accessibility can hinder their learning experience and compromise equity in education. To address this issue, the TechWhiz project has been developed as a comprehensive glossary of scientific and technical concepts explained in sign language. By providing deaf students with access to education in their first language, TechWhiz aims to enhance their learning achievements and promote inclusivity while also fostering equity in education for all students.Keywords: deaf students, technical and scientific knowledge, automatic sign language, inclusive education
Procedia PDF Downloads 685697 Image Processing-Based Maize Disease Detection Using Mobile Application
Authors: Nathenal Thomas
Abstract:
In the food chain and in many other agricultural products, corn, also known as maize, which goes by the scientific name Zea mays subsp, is a widely produced agricultural product. Corn has the highest adaptability. It comes in many different types, is employed in many different industrial processes, and is more adaptable to different agro-climatic situations. In Ethiopia, maize is among the most widely grown crop. Small-scale corn farming may be a household's only source of food in developing nations like Ethiopia. The aforementioned data demonstrates that the country's requirement for this crop is excessively high, and conversely, the crop's productivity is very low for a variety of reasons. The most damaging disease that greatly contributes to this imbalance between the crop's supply and demand is the corn disease. The failure to diagnose diseases in maize plant until they are too late is one of the most important factors influencing crop output in Ethiopia. This study will aid in the early detection of such diseases and support farmers during the cultivation process, directly affecting the amount of maize produced. The diseases in maize plants, such as northern leaf blight and cercospora leaf spot, have distinct symptoms that are visible. This study aims to detect the most frequent and degrading maize diseases using the most efficiently used subset of machine learning technology, deep learning so, called Image Processing. Deep learning uses networks that can be trained from unlabeled data without supervision (unsupervised). It is a feature that simulates the exercises the human brain goes through when digesting data. Its applications include speech recognition, language translation, object classification, and decision-making. Convolutional Neural Network (CNN) for Image Processing, also known as convent, is a deep learning class that is widely used for image classification, image detection, face recognition, and other problems. it will also use this algorithm as the state-of-the-art for my research to detect maize diseases by photographing maize leaves using a mobile phone.Keywords: CNN, zea mays subsp, leaf blight, cercospora leaf spot
Procedia PDF Downloads 745696 Vision-Based Hand Segmentation Techniques for Human-Computer Interaction
Abstract:
This work is the part of vision based hand gesture recognition system for Natural Human Computer Interface. Hand tracking and segmentation are the primary steps for any hand gesture recognition system. The aim of this paper is to develop robust and efficient hand segmentation algorithm such as an input to another system which attempt to bring the HCI performance nearby the human-human interaction, by modeling an intelligent sign language recognition system based on prediction in the context of dialogue between the system (avatar) and the interlocutor. For the purpose of hand segmentation, an overcoming occlusion approach has been proposed for superior results for detection of hand from an image.Keywords: HCI, sign language recognition, object tracking, hand segmentation
Procedia PDF Downloads 4125695 Design of Speed Bump Recognition System Integrated with Adjustable Shock Absorber Control
Authors: Ming-Yen Chang, Sheng-Hung Ke
Abstract:
This research focuses on the development of a speed bump identification system for real-time control of adjustable shock absorbers in vehicular suspension systems. The study initially involved the collection of images of various speed bumps, and rubber speed bump profiles found on roadways. These images were utilized for training and recognition purposes through the deep learning object detection algorithm YOLOv5. Subsequently, the trained speed bump identification program was integrated with an in-vehicle camera system for live image capture during driving. These images were instantly transmitted to a computer for processing. Using the principles of monocular vision ranging, the distance between the vehicle and an approaching speed bump was determined. The appropriate control distance was established through both practical vehicle measurements and theoretical calculations. Collaboratively, with the electronically adjustable shock absorbers equipped in the vehicle, a shock absorber control system was devised to dynamically adapt the damping force just prior to encountering a speed bump. This system effectively mitigates passenger discomfort and enhances ride quality.Keywords: adjustable shock absorbers, image recognition, monocular vision ranging, ride
Procedia PDF Downloads 675694 Muscle: The Tactile Texture Designed for the Blind
Authors: Chantana Insra
Abstract:
The research objective focuses on creating a prototype media of the tactile texture of muscles for educational institutes to help visually impaired students learn massage extra learning materials further than the ordinary curriculum. This media is designed as an extra learning material. The population in this study was 30 blinded students between 4th - 6th grades who were able to read Braille language. The research was conducted during the second semester in 2012 at The Bangkok School for the Blind. The method in choosing the population in the study was purposive sampling. The methodology of the research includes collecting data related to visually impaired people, the production of the tactile texture media, human anatomy and Thai traditional massage from literature reviews and field studies. This information was used for analyzing and designing 14 tactile texture pictures presented to experts to evaluate and test the media.Keywords: blind, tactile texture, muscle, visual arts and design
Procedia PDF Downloads 2695693 Future-Proofing the Workforce: A Case Study of Integrated Human Capability Frameworks to Support Business Success
Authors: Penelope Paliadelis, Asheley Jones, Glenn Campbell
Abstract:
This paper discusses the development of co-designed capability frameworks for two large multinational organizations led by a university department. The aim was to create evidence-based, integrated capability frameworks that could define, identify, and measure human skill capabilities independent of specific work roles. The frameworks capture and cluster human skills required in the workplace and capture their application at various levels of mastery. Identified capability gaps inform targeted learning opportunities for workers to enhance their employability skills. The paper highlights the value of this evidence-based framework development process in capturing, defining, and assessing desired human-focused capabilities for organizational growth and success.Keywords: capability framework, human skills, work-integrated learning, credentialing, digital badging
Procedia PDF Downloads 795692 A System to Detect Inappropriate Messages in Online Social Networks
Authors: Shivani Singh, Shantanu Nakhare, Kalyani Nair, Rohan Shetty
Abstract:
As social networking is growing at a rapid pace today it is vital that we work on improving its management. Research has shown that the content present in online social networks may have significant influence on impressionable minds. If such platforms are misused, it will lead to negative consequences. Detecting insults or inappropriate messages continues to be one of the most challenging aspects of Online Social Networks (OSNs) today. We address this problem through a Machine Learning Based Soft Text Classifier approach using Support Vector Machine algorithm. The proposed system acts as a screening mechanism the alerts the user about such messages. The messages are classified according to their subject matter and each comment is labeled for the presence of profanity and insults.Keywords: machine learning, online social networks, soft text classifier, support vector machine
Procedia PDF Downloads 5085691 Musical Notation Reading versus Alphabet Reading-Comparison and Implications for Teaching Music Reading to Students with Dyslexia
Authors: Ora Geiger
Abstract:
Reading is a cognitive process of deciphering visual signs to produce meaning. During the reading process, written information of symbols and signs is received in the person’s eye and processed in the brain. This definition is relevant to both the reading of letters and the reading of musical notation. But while the letters of the alphabet are signs determined arbitrarily, notes are recorded systematically on a staff, with the location of each note on the staff indicating its relative pitch. In this paper, the researcher specifies the characteristics of alphabet reading in comparison to musical notation reading, and discusses the question whether a person diagnosed with dyslexia will necessarily have difficulty in reading musical notes. Dyslexia is a learning disorder that makes it difficult to acquire alphabet-reading skills due to difficulties expressed in the identification of letters, spelling, and other language deciphering skills. In order to read, one must be able to connect a symbol with a sound and to join the sounds into words. A person who has dyslexia finds it difficult to translate a graphic symbol into the sound that it represents. When teaching reading to children diagnosed with dyslexia, the multi-sensory approach, supporting the activation and involvement of most of the senses in the learning process, has been found to be particularly effective. According to this approach, when most senses participate in the reading learning process, it becomes more effective. During years of experience, the researcher, who is a music specialist, has been following the music reading learning process of elementary school age students, some of them diagnosed with Dyslexia, while studying to play soprano (descant) recorder. She argues that learning music reading while studying to play a musical instrument is a multi-sensory experience by its nature. The senses involved are: sight, hearing, touch, and the kinesthetic sense (motion), which provides the brain with information on the relative positions of the body. In this way, the learner experiences simultaneously visual, auditory, tactile, and kinesthetic impressions. The researcher concludes that there should be no contra-indication for teaching standard music reading to children with dyslexia if an appropriate process is offered. This conclusion is based on two main characteristics of music reading: (1) musical notation system is a systematic, logical, relative set of symbols written on a staff; and (2) music reading learning connected with playing a musical instrument is by its nature a multi-sensory activity since it combines sight, hearing, touch, and movement. This paper describes music reading teaching procedures and provides unique teaching methods that have been found to be effective for students who were diagnosed with Dyslexia. It provides theoretical explanations in addition to guidelines for music education practices.Keywords: alphabet reading, dyslexia, multisensory teaching method, music reading, recorder playing
Procedia PDF Downloads 3655690 Exploring the Determinants of Personal Finance Difficulties by Machine Learning: Focus on Socio-Economic and Behavioural Changes Brought by COVID-19
Authors: Brian Tung, Yam Wing Siu, Tsun Se Cheong
Abstract:
Purpose: This research aims to explore how personal and environmental factors, especially the socio-economic changes and behavioral changes fostered by the COVID-19 outbreak pandemic, affect the financial vulnerability of a specific segment of people in financial distress. Innovative research methodology of machine learning will be applied to data collected from over 300 local individuals in Hong Kong seeking counseling or similar services in recent years. Results: First, machine learning has found that too much exposure to digital services and information on digitized services may lead to adverse effects on respondents’ financial vulnerability. Second, the improvement in financial literacy level provides benefits to the financially vulnerable group, especially those respondents who have started with a lower level. Third, serious addiction to digital technology can lead to worsened debt servicing ability. Machine learning also has found a strong correlation between debt servicing situations and income-seeking behavior as well as spending behavior. In addition, if the vulnerable groups are able to make appropriate investments, they can reduce the probability of incurring financial distress. Finally, being too active in borrowing and repayment can result in a higher likelihood of over-indebtedness. Conclusion: Findings can be employed in formulating a better counseling strategy for professionals. Debt counseling services can be more preventive in nature. For example, according to the findings, with a low level of financial literacy, the respondents are prone to overspending and unable to react properly to the e-marketing promotion messages pop-up from digital services or even falling into financial/investment scams. In addition, people with low levels of financial knowledge will benefit from financial education. Therefore, financial education programs could include tech-savvy matters as special features.Keywords: personal finance, digitization of the economy, COVID-19 pandemic, addiction to digital technology, financial vulnerability
Procedia PDF Downloads 585689 A Novel Computer-Generated Hologram (CGH) Achieved Scheme Generated from Point Cloud by Using a Lens Array
Authors: Wei-Na Li, Mei-Lan Piao, Nam Kim
Abstract:
We proposed a novel computer-generated hologram (CGH) achieved scheme, wherein the CGH is generated from a point cloud which is transformed by a mapping relationship of a series of elemental images captured from a real three-dimensional (3D) object by using a lens array. This scheme is composed of three procedures: mapping from elemental images to point cloud, hologram generation, and hologram display. A mapping method is figured out to achieve a virtual volume date (point cloud) from a series of elemental images. This mapping method consists of two steps. Firstly, the coordinate (x, y) pairs and its appearing number are calculated from the series of sub-images, which are generated from the elemental images. Secondly, a series of corresponding coordinates (x, y, z) are calculated from the elemental images. Then a hologram is generated from the volume data that is calculated by the previous two steps. Eventually, a spatial light modulator (SLM) and a green laser beam are utilized to display this hologram and reconstruct the original 3D object. In this paper, in order to show a more auto stereoscopic display of a real 3D object, we successfully obtained the actual depth data of every discrete point of the real 3D object, and overcame the inherent drawbacks of the depth camera by obtaining point cloud from the elemental images.Keywords: elemental image, point cloud, computer-generated hologram (CGH), autostereoscopic display
Procedia PDF Downloads 5845688 A Time-Varying and Non-Stationary Convolution Spectral Mixture Kernel for Gaussian Process
Authors: Kai Chen, Shuguang Cui, Feng Yin
Abstract:
Gaussian process (GP) with spectral mixture (SM) kernel demonstrates flexible non-parametric Bayesian learning ability in modeling unknown function. In this work a novel time-varying and non-stationary convolution spectral mixture (TN-CSM) kernel with a significant enhancing of interpretability by using process convolution is introduced. A way decomposing the SM component into an auto-convolution of base SM component and parameterizing it to be input dependent is outlined. Smoothly, performing a convolution between two base SM component yields a novel structure of non-stationary SM component with much better generalized expression and interpretation. The TN-CSM perfectly allows compatibility with the stationary SM kernel in terms of kernel form and spectral base ignored and confused by previous non-stationary kernels. On synthetic and real-world datatsets, experiments show the time-varying characteristics of hyper-parameters in TN-CSM and compare the learning performance of TN-CSM with popular and representative non-stationary GP.Keywords: Gaussian process, spectral mixture, non-stationary, convolution
Procedia PDF Downloads 1965687 Enhancing Learning for Research Higher Degree Students
Authors: Jenny Hall, Alison Jaquet
Abstract:
Universities’ push toward the production of high quality research is not limited to academic staff and experienced researchers. In this environment of research rich agendas, Higher Degree Research (HDR) students are increasingly expected to engage in the publishing of good quality papers in high impact journals. IFN001: Advanced Information Research Skills (AIRS) is a credit bearing mandatory coursework requirement for Queensland University of Technology (QUT) doctorates. Since its inception in 1989, this unique blended learning program has provided the foundations for new researchers to produce original and innovative research. AIRS was redeveloped in 2012, and has now been evaluated with reference to the university’s strategic research priorities. Our research is the first comprehensive evaluation of the program from the learner perspective. We measured whether the program develops essential transferrable skills and graduate capabilities to ensure best practice in the areas of publishing and data management. In particular, we explored whether AIRS prepares students to be agile researchers with the skills to adapt to different research contexts both within and outside academia. The target group for our study consisted of HDR students and supervisors at QUT. Both quantitative and qualitative research methods were used for data collection. Gathering data was by survey and focus groups with qualitative responses analyzed using NVivo. The results of the survey show that 82% of students surveyed believe that AIRS assisted their research process and helped them learn skills they need as a researcher. The 18% of respondents who expressed reservation about the benefits of AIRS were also examined to determine the key areas of concern. These included trends related to the timing of the program early in the candidature and a belief among some students that their previous research experience was sufficient for postgraduate study. New insights have been gained into how to better support HDR learners in partnership with supervisors and how to enhance learning experiences of specific cohorts, including international students and mature learners.Keywords: data management, enhancing learning experience, publishing, research higher degree students, doctoral students
Procedia PDF Downloads 2745686 Amplifying Sine Unit-Convolutional Neural Network: An Efficient Deep Architecture for Image Classification and Feature Visualizations
Authors: Jamshaid Ul Rahman, Faiza Makhdoom, Dianchen Lu
Abstract:
Activation functions play a decisive role in determining the capacity of Deep Neural Networks (DNNs) as they enable neural networks to capture inherent nonlinearities present in data fed to them. The prior research on activation functions primarily focused on the utility of monotonic or non-oscillatory functions, until Growing Cosine Unit (GCU) broke the taboo for a number of applications. In this paper, a Convolutional Neural Network (CNN) model named as ASU-CNN is proposed which utilizes recently designed activation function ASU across its layers. The effect of this non-monotonic and oscillatory function is inspected through feature map visualizations from different convolutional layers. The optimization of proposed network is offered by Adam with a fine-tuned adjustment of learning rate. The network achieved promising results on both training and testing data for the classification of CIFAR-10. The experimental results affirm the computational feasibility and efficacy of the proposed model for performing tasks related to the field of computer vision.Keywords: amplifying sine unit, activation function, convolutional neural networks, oscillatory activation, image classification, CIFAR-10
Procedia PDF Downloads 111