Search results for: deep learning methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 21438

Search results for: deep learning methods

16638 Transforming Personal Healthcare through Patient Engagement: An In-Depth Analysis of Tools and Methods for the Digital Age

Authors: Emily Hickmann, Peggy Richter, Maren Kaehlig, Hannes Schlieter

Abstract:

Patient engagement is a cornerstone of high-quality care and essential for patients with chronic diseases to achieve improved health outcomes. Through digital transformation, possibilities to engage patients in their personal healthcare have multiplied. However, the exploitation of this potential is still lagging. To support the transmission of patient engagement theory into practice, this paper’s objective is to give a state-of-the-art overview of patient engagement tools and methods. A systematic literature review was conducted. Overall, 56 tools and methods were extracted and synthesized according to the four attributes of patient engagement, i.e., personalization, access, commitment, and therapeutic alliance. The results are discussed in terms of their potential to be implemented in digital health solutions under consideration of the “computers are social actors” (CASA) paradigm. It is concluded that digital health can catalyze patient engagement in practice, and a broad future research agenda is formulated.

Keywords: chronic diseases, digitalization, patient-centeredness, patient empowerment, patient engagement

Procedia PDF Downloads 109
16637 Design-Based Elements to Sustain Participant Activity in Massive Open Online Courses: A Case Study

Authors: C. Zimmermann, E. Lackner, M. Ebner

Abstract:

Massive Open Online Courses (MOOCs) are increasingly popular learning hubs that are boasting considerable participant numbers, innovative technical features, and a multitude of instructional resources. Still, there is a high level of evidence showing that almost all MOOCs suffer from a declining frequency of participant activity and fairly low completion rates. In this paper, we would like to share the lessons learned in implementing several design patterns that have been suggested in order to foster participant activity. Our conclusions are based on experiences with the ‘Dr. Internet’ MOOC, which was created as an xMOOC to raise awareness for a more critical approach to online health information: participants had to diagnose medical case studies. There is a growing body of recommendations (based on Learning Analytics results from earlier xMOOCs) as to how the decline in participant activity can be alleviated. One promising focus in this regard is instructional design patterns, since they have a tremendous influence on the learner’s motivation, which in turn is a crucial trigger of learning processes. Since Medieval Age storytelling, micro-learning units and specific comprehensible, narrative structures were chosen to animate the audience to follow narration. Hence, MOOC participants are not likely to abandon a course or information channel when their curiosity is kept at a continuously high level. Critical aspects that warrant consideration in this regard include shorter course duration, a narrative structure with suspense peaks (according to the ‘storytelling’ approach), and a course schedule that is diversified and stimulating, yet easy to follow. All of these criteria have been observed within the design of the Dr. Internet MOOC: 1) the standard eight week course duration was shortened down to six weeks, 2) all six case studies had a special quiz format and a corresponding resolution video which was made available in the subsequent week, 3) two out of six case studies were split up in serial video sequences to be presented over the span of two weeks, and 4) the videos were generally scheduled in a less predictable sequence. However, the statistical results from the first run of the MOOC do not indicate any strong influences on the retention rate, so we conclude with some suggestions as to why this might be and what aspects need further consideration.

Keywords: case study, Dr. internet, experience, MOOCs, design patterns

Procedia PDF Downloads 259
16636 Lessons Learned from Interlaboratory Noise Modelling in Scope of Environmental Impact Assessments in Slovenia

Authors: S. Cencek, A. Markun

Abstract:

Noise assessment methods are regularly used in scope of Environmental Impact Assessments for planned projects to assess (predict) the expected noise emissions of these projects. Different noise assessment methods could be used. In recent years, we had an opportunity to collaborate in some noise assessment procedures where noise assessments of different laboratories have been performed simultaneously. We identified some significant differences in noise assessment results between laboratories in Slovenia. We estimate that despite good input Georeferenced Data to set up acoustic model exists in Slovenia; there is no clear consensus on methods for predictive noise methods for planned projects. We analyzed input data, methods and results of predictive noise methods for two planned industrial projects, both were done independently by two laboratories. We also analyzed the data, methods and results of two interlaboratory collaborative noise models for two existing noise sources (railway and motorway). In cases of predictive noise modelling, the validations of acoustic models were performed by noise measurements of surrounding existing noise sources, but in varying durations. The acoustic characteristics of existing buildings were also not described identically. The planned noise sources were described and digitized differently. Differences in noise assessment results between different laboratories have ranged up to 10 dBA, which considerably exceeds the acceptable uncertainty ranged between 3 to 6 dBA. Contrary to predictive noise modelling, in cases of collaborative noise modelling for two existing noise sources the possibility to perform the validation noise measurements of existing noise sources greatly increased the comparability of noise modelling results. In both cases of collaborative noise modelling for existing motorway and railway, the modelling results of different laboratories were comparable. Differences in noise modeling results between different laboratories were below 5 dBA, which was acceptable uncertainty set up by interlaboratory noise modelling organizer. The lessons learned from the study were: 1) Predictive noise calculation using formulae from International standard SIST ISO 9613-2: 1997 is not an appropriate method to predict noise emissions of planned projects since due to complexity of procedure they are not used strictly, 2) The noise measurements are important tools to minimize noise assessment errors of planned projects and should be in cases of predictive noise modelling performed at least for validation of acoustic model, 3) National guidelines should be made on the appropriate data, methods, noise source digitalization, validation of acoustic model etc. in order to unify the predictive noise models and their results in scope of Environmental Impact Assessments for planned projects.

Keywords: environmental noise assessment, predictive noise modelling, spatial planning, noise measurements, national guidelines

Procedia PDF Downloads 229
16635 Systematic Evaluation of Convolutional Neural Network on Land Cover Classification from Remotely Sensed Images

Authors: Eiman Kattan, Hong Wei

Abstract:

In using Convolutional Neural Network (CNN) for classification, there is a set of hyperparameters available for the configuration purpose. This study aims to evaluate the impact of a range of parameters in CNN architecture i.e. AlexNet on land cover classification based on four remotely sensed datasets. The evaluation tests the influence of a set of hyperparameters on the classification performance. The parameters concerned are epoch values, batch size, and convolutional filter size against input image size. Thus, a set of experiments were conducted to specify the effectiveness of the selected parameters using two implementing approaches, named pertained and fine-tuned. We first explore the number of epochs under several selected batch size values (32, 64, 128 and 200). The impact of kernel size of convolutional filters (1, 3, 5, 7, 10, 15, 20, 25 and 30) was evaluated against the image size under testing (64, 96, 128, 180 and 224), which gave us insight of the relationship between the size of convolutional filters and image size. To generalise the validation, four remote sensing datasets, AID, RSD, UCMerced and RSCCN, which have different land covers and are publicly available, were used in the experiments. These datasets have a wide diversity of input data, such as number of classes, amount of labelled data, and texture patterns. A specifically designed interactive deep learning GPU training platform for image classification (Nvidia Digit) was employed in the experiments. It has shown efficiency in both training and testing. The results have shown that increasing the number of epochs leads to a higher accuracy rate, as expected. However, the convergence state is highly related to datasets. For the batch size evaluation, it has shown that a larger batch size slightly decreases the classification accuracy compared to a small batch size. For example, selecting the value 32 as the batch size on the RSCCN dataset achieves the accuracy rate of 90.34 % at the 11th epoch while decreasing the epoch value to one makes the accuracy rate drop to 74%. On the other extreme, setting an increased value of batch size to 200 decreases the accuracy rate at the 11th epoch is 86.5%, and 63% when using one epoch only. On the other hand, selecting the kernel size is loosely related to data set. From a practical point of view, the filter size 20 produces 70.4286%. The last performed image size experiment shows a dependency in the accuracy improvement. However, an expensive performance gain had been noticed. The represented conclusion opens the opportunities toward a better classification performance in various applications such as planetary remote sensing.

Keywords: CNNs, hyperparamters, remote sensing, land cover, land use

Procedia PDF Downloads 164
16634 Early Impact Prediction and Key Factors Study of Artificial Intelligence Patents: A Method Based on LightGBM and Interpretable Machine Learning

Authors: Xingyu Gao, Qiang Wu

Abstract:

Patents play a crucial role in protecting innovation and intellectual property. Early prediction of the impact of artificial intelligence (AI) patents helps researchers and companies allocate resources and make better decisions. Understanding the key factors that influence patent impact can assist researchers in gaining a better understanding of the evolution of AI technology and innovation trends. Therefore, identifying highly impactful patents early and providing support for them holds immeasurable value in accelerating technological progress, reducing research and development costs, and mitigating market positioning risks. Despite the extensive research on AI patents, accurately predicting their early impact remains a challenge. Traditional methods often consider only single factors or simple combinations, failing to comprehensively and accurately reflect the actual impact of patents. This paper utilized the artificial intelligence patent database from the United States Patent and Trademark Office and the Len.org patent retrieval platform to obtain specific information on 35,708 AI patents. Using six machine learning models, namely Multiple Linear Regression, Random Forest Regression, XGBoost Regression, LightGBM Regression, Support Vector Machine Regression, and K-Nearest Neighbors Regression, and using early indicators of patents as features, the paper comprehensively predicted the impact of patents from three aspects: technical, social, and economic. These aspects include the technical leadership of patents, the number of citations they receive, and their shared value. The SHAP (Shapley Additive exPlanations) metric was used to explain the predictions of the best model, quantifying the contribution of each feature to the model's predictions. The experimental results on the AI patent dataset indicate that, for all three target variables, LightGBM regression shows the best predictive performance. Specifically, patent novelty has the greatest impact on predicting the technical impact of patents and has a positive effect. Additionally, the number of owners, the number of backward citations, and the number of independent claims are all crucial and have a positive influence on predicting technical impact. In predicting the social impact of patents, the number of applicants is considered the most critical input variable, but it has a negative impact on social impact. At the same time, the number of independent claims, the number of owners, and the number of backward citations are also important predictive factors, and they have a positive effect on social impact. For predicting the economic impact of patents, the number of independent claims is considered the most important factor and has a positive impact on economic impact. The number of owners, the number of sibling countries or regions, and the size of the extended patent family also have a positive influence on economic impact. The study primarily relies on data from the United States Patent and Trademark Office for artificial intelligence patents. Future research could consider more comprehensive data sources, including artificial intelligence patent data, from a global perspective. While the study takes into account various factors, there may still be other important features not considered. In the future, factors such as patent implementation and market applications may be considered as they could have an impact on the influence of patents.

Keywords: patent influence, interpretable machine learning, predictive models, SHAP

Procedia PDF Downloads 39
16633 TutorBot+: Automatic Programming Assistant with Positive Feedback based on LLMs

Authors: Claudia Martínez-Araneda, Mariella Gutiérrez, Pedro Gómez, Diego Maldonado, Alejandra Segura, Christian Vidal-Castro

Abstract:

The purpose of this document is to showcase the preliminary work in developing an EduChatbot-type tool and measuring the effects of its use aimed at providing effective feedback to students in programming courses. This bot, hereinafter referred to as tutorBot+, was constructed based on chatGPT and is tasked with assisting and delivering timely positive feedback to students in the field of computer science at the Universidad Católica de Concepción. The proposed working method consists of four stages: (1) Immersion in the domain of Large Language Models (LLMs), (2) Development of the tutorBot+ prototype and integration, (3) Experiment design, and (4) Intervention. The first stage involves a literature review on the use of artificial intelligence in education and the evaluation of intelligent tutors, as well as research on types of feedback for learning and the domain of chatGPT. The second stage encompasses the development of tutorBot+, and the final stage involves a quasi-experimental study with students from the Programming and Database labs, where the learning outcome involves the development of computational thinking skills, enabling the use and measurement of the tool's effects. The preliminary results of this work are promising, as a functional chatBot prototype has been developed in both conversational and non-conversational versions integrated into an open-source online judge and programming contest platform system. There is also an exploration of the possibility of generating a custom model based on a pre-trained one tailored to the domain of programming. This includes the integration of the created tool and the design of the experiment to measure its utility.

Keywords: assessment, chatGPT, learning strategies, LLMs, timely feedback

Procedia PDF Downloads 65
16632 International Collaboration: Developing the Practice of Social Work Curriculum through Study Abroad and Participatory Research

Authors: Megan Lindsey

Abstract:

Background: Globalization presents international social work with both opportunities and challenges. Thus, the design of this international experience aligns with the three charges of the Commission on Global Social Work Education. An international collaborative effort between an American and Scottish University Social Work Program was based on an established University agreement. The presentation provides an overview of an international study abroad among American and Scottish Social Work students. Further, presenters will discuss the opportunities of international collaboration and the challenges of the project. First, we will discuss the process of a successful international collaboration. This discussion will include the planning, collaboration, execution of the experience, along with its application to the international field of social work. Second, we will discuss the development and implementation of participatory action research in which the student engage to enhance their learning experience. A collaborative qualitative research project was undertaken with three goals. First, students gained experience in Scottish social services, including agency visits and presentations. Second, a collaboration between American and Scottish MSW Students allowed the exchange of ideas and knowledge about services and social work education. Third, students collaborated on a qualitative research method to reflect on their social work education and the formation of their professional identity. Methods/Methodology: American and Scottish students engaged in participatory action research by using Photovoice methods while studying together in Scotland. The collaboration between faculty researchers framed a series of research questions. Both universities obtained IRB approval and trained students in Photovoice methods. The student teams used the research question and Photovoice method to discover images that represented their professional identity formation. Two Photovoice goals grounded the study's research question. First, the methods enabled the individual students to record and reflect on their professional strengths and concerns. Second, student teams promoted critical dialogue and knowledge about personal and professional issues through large and small group discussions of photographs. Results: The international participatory approach generated the ability for students to contextualize their common social work education and practice experiences. Team discussions between representatives of each country resulted in understanding professional identity formation and the processes of social work education that contribute to that identity. Students presented the photograph narration of their knowledge and understanding of international social work education and practice. Researchers then collaborated on finding common themes. The results found commonalities in the quality and depth of social work education. The themes found differences regarding how professional identity is formed. Students found great differences between their and American accreditation and certification. Conclusions: Faculty researchers’ collaboration themes sought to categorize the students’ experiences of their professional identity. While the social work education systems are similar, there are vast differences. The Scottish themes noted structures within American social work not found in the United Kingdom. The American researchers noted that Scotland, as does the United Kingdom, relies on programs, agencies, and the individual social worker to provide structure to identity formation. Other themes will be presented.

Keywords: higher education curriculum, international collaboration, social sciences, action research

Procedia PDF Downloads 121
16631 Higher Education Teachers' Perceptions of Core Competencies and Innovation: The Case of Mohamed V University Abu Dhabi

Authors: Khalid Soussi

Abstract:

Implementing innovative teaching and learning methods is of pivotal importance for student motivation and teaching quality. At the center of such quality are teaching competencies. The present paper investigates three teachers’ core competencies related to their innovative teaching performance: educational/pedagogical competency, teaching competency, and social competency. The paper also attempts to describe the influence of social factors on innovation in higher education. Many recent studies highlight the technological competency as an independent one, but it is believed in this study that the latter makes part of the pedagogical competency. A Likert scale questionnaire was used to measure teachers’ judgements of core competencies role in innovative teaching performance. The study also attempted to demarcate the social variables that may affect innovative teaching in higher education. The findings indicate that teachers’ educational competency and teaching competency were generally confirmed to be either important or very important for innovation in teaching performance. Regarding social competency, the study also shows that satisfaction from job, daily working hours, amount of workload, flexibility in the functioning and the quality of students are the main factors that have a large effect on teachers’ innovative teaching performance.

Keywords: higher education, innovative teaching, teaching competencies, teaching performance

Procedia PDF Downloads 142
16630 Implementation of the Science Curriculum of the Colleges of Education: Successes and Challenges

Authors: Cecilia Boakye, Joseph Ghartey Ampiah

Abstract:

In this study, we present a case study in which we explored how the 2007 science curriculum of the colleges of education in Ghana was implemented at W College of Education. Purposive sampling was used to select 13 participants, comprising 2 tutors and 6 teacher trainees from W College of Education and, 5 newly qualified Junior High School (JHS) science teachers who were products of W College. Interviews, observations and content analysis were used to collect data. Using the deductive and inductive analytic approaches, the findings showed that although upgraded laboratories have provided for teaching authentic science at W College of Education, they are rather used to accommodate large classes at the expense of practical activities. The teaching and learning methods used by the tutors do not mirror effectively the objectives of the 2007 science curriculum of the colleges of education. There are challenges such as: (a) lack/inadequate equipment and materials, (b) time constraint, and (c) an examination- oriented curriculum that influence the implementation of the curriculum. Some of the suggestions that were made are that: (a) equipment and materials should be supplied to the colleges to facilitate the proper implementation of the curriculum, and (b) class sizes should be reduced to provide enough room for practical activities.

Keywords: class size, teaching, curriculum implementation, examination-oriented curriculum, teaching and time-constraint

Procedia PDF Downloads 267
16629 People Who Live in Poverty Usually Do So Due to Circumstances Far Beyond Their Control: A Multiple Case Study on Poverty Simulation Events

Authors: Tracy Smith-Carrier

Abstract:

Burgeoning research extols the benefits of innovative experiential learning activities to increase participants’ engagement, enhance their individual learning, and bridge the gap between theory and practice. This presentation discusses findings from a multiple case study on poverty simulation events conducted with two samples: undergraduate students and community participants. After exploring the nascent research on the benefits and limitations of poverty simulation activities, the study explores whether participating in a poverty simulation resulted in changes to participants’ beliefs about the causes and effects of poverty, as well as shifts in their attitudes and actions toward people experiencing poverty. For the purposes of triangulation, quantitative and qualitative data from a variety of sources were analyzed: participant feedback surveys, qualitative responses, and pre, post, and follow-up questionnaires. Findings show statistically significant results (p<.05) from both samples on cumulative scores of the modified Attitudes Toward Poverty Scale, indicating an improvement in participants’ attitudes toward poverty. Although generally positive about their experiences, participating in the simulation did not appear to have prompted participants to take specific actions to reduce poverty. Conclusions drawn from the research study suggest that poverty simulation planners should be wary of adopting scenarios that emphasize, or fail to adequately contextualize, behaviours or responses that might perpetuate individual explanations of poverty. Moreover, organizers must carefully consider how to ensure participants in their audience currently experiencing low-income do not become emotionally distressed, triggered or further marginalized in the process. While overall participants were positive about their experiences in the simulation, the events did not appear to have prompted them to action. Moving beyond the goal of increasing participants’ understandings of poverty, interventions that foster greater engagement in poverty issues over the long-term are necessary.

Keywords: empathy, experiential learning, poverty awareness, poverty simulation

Procedia PDF Downloads 265
16628 Use of Social Networks and Mobile Technologies in Education

Authors: Václav Maněna, Roman Dostál, Štěpán Hubálovský

Abstract:

Social networks play an important role in the lives of children and young people. Along with the high penetration of mobile technologies such as smartphones and tablets among the younger generation, there is an increasing use of social networks already in elementary school. The paper presents the results of research, which was realized at schools in the Hradec Králové region. In this research, the authors focused on issues related to communications on social networks for children, teenagers and young people in the Czech Republic. This research was conducted at selected elementary, secondary and high schools using anonymous questionnaires. The results are evaluated and compared with the results of the research, which has been realized in 2008. The authors focused on the possibilities of using social networks in education. The paper presents the possibility of using the most popular social networks in education, with emphasis on increasing motivation for learning. The paper presents comparative analysis of social networks, with regard to the possibility of using in education as well.

Keywords: social networks, motivation, e-learning, mobile technology

Procedia PDF Downloads 307
16627 Basic Properties of a Fundamental Particle: Behavioral-Physical and Visual Methods for the Study of Fundamental Particle

Authors: Shukran M. Dadayev

Abstract:

To author's best knowledge, in this paper, the Basic Properties and Research methods of a Fundamental Particle is studied for the first time. That's to say, Fundamental Particle has not been discovered in the Nature yet. Because Fundamental Particle consists of specific Physical, Geometrical and Internal bases. Geometrical and Internal characteristics that are considered significant for the elementary and fundamental particles aren’t basic properties, characteristics or criteria of a Fundamental Particle. Of course, completely new Physical and Visual experimental methods of Quantum mechanics and Behavioral-Physical investigations of Particles are needed to study and discover the Fundamental Particle. These are new Physical, Visual and Behavioral-Physical experimental methods for describing and discovering the Fundamental Particle in the Nature and Microworld. Fundamental Particle consists of the same Energy-Mass-Motion system and a symmetry of Energy-Mass-Motion. Fundamental Particle supplies each of the elementary particles with the same Energy-Mass-Motion system at the same time and regulates each of the particles. Fundamental Particle gives Energy, Mass and Motion to each particles at the same time, each of the Particles consists of acquired Energy-Mass-Motion system and symmetry. Energy, Mass, Motion given by the Fundamental Particle to the particles are Symmetrical Equivalent and they remain in their primary shapes in all cases. Fundamental Particle gives Energy-Mass-Motion system and symmetry consisting of different measures and functions to each of the particles. The Motion given by the Fundamental Particle to the particles is Gravitation, Gravitational Interaction not only gives Motion, but also cause Motion by attracting. All Substances, Fields and Cosmic objects consist of Energy-Mass-Motion. The Field also includes specific Mass. They are always Energetic, Massive and Active. Fundamental Particle establishes the bases of the Nature. Supplement and Regulating of all the particles existing in the Nature belongs to Fundamental Particle.

Keywords: basic properties of a fundamental particle, behavioral-physical and visual methods, energy-mass-motion system and symmetrical equivalence, fundamental particle

Procedia PDF Downloads 3649
16626 Academic Staff Recruitment in Islamic University: A Proposed Holistic Model

Authors: Syahruddin Sumardi, Indra Fajar Alamsyah, Junaidah Hashim

Abstract:

This study attempts to explore and presents a proposed recruitment model in Islamic university which aligned with holistic role. It is a conceptual paper in nature. In turn, this study is designed to utilize exploratory approach. Literature and document review that related to this topic are used as the methods to analyse the content found. Recruitment for any organization is fundamental to achieve its goal effectively. Staffing in universities is vital due to the importance role of lecturers. Currently, Islamic universities still adopt the common process of recruitment for their academic staffs. Whereas, they have own characteristics which are embedded in their institutions. Furthermore, the FCWC (Foundation, Capability, Worldview and Commitment) model of recruitment proposes to suit the holistic character of Islamic university. Further studies are required to empirically validate the concept through systematic investigations. Additionally, measuring this model by a designed means is appreciated. The model provides the map and alternative tool of recruitment for Islamic universities to determine the process of recruitment which can appropriate their institutions. In addition, it also allows stakeholders and policy makers to consider regarding Islamic values that should inculcate in the Islamic higher learning institutions. This study initiates a foundational contribution for an early sequence of research.

Keywords: academic staff, Islamic values, recruitment model, university

Procedia PDF Downloads 163
16625 Numerical Study on the Effect of Spudcan Penetration on the Jacket Platform

Authors: Xiangming Ge, Bing Pan, Wei He, Hao Chen, Yong Zhou, Jiayao Wu, Weijiang Chu

Abstract:

How the extraction and penetration of spudcan affect the performance of the adjacent pile foundation supporting the jacket platform was studied in the program FLAC3D depending on a wind farm project in Bohai sea. The simulations were conducted at the end of the spudcan penetration, which induced a pockmark in the seabed. The effects of the distance between the pile foundation and the pockmark were studied. The displacement at the mudline arose when the pockmark was closer. The bearing capacity of this jacket platform with deep pile foundations has been less influenced by the process of spudcan penetration, which can induce severe stresses on the pile foundation. The induced rotation was also satisfied with the rotation-controlling criteria.

Keywords: offshore foundation, pile-soil interaction, spudcan penetration, FLAC3D

Procedia PDF Downloads 209
16624 Performance Evaluation of Various Segmentation Techniques on MRI of Brain Tissue

Authors: U.V. Suryawanshi, S.S. Chowhan, U.V Kulkarni

Abstract:

Accuracy of segmentation methods is of great importance in brain image analysis. Tissue classification in Magnetic Resonance brain images (MRI) is an important issue in the analysis of several brain dementias. This paper portraits performance of segmentation techniques that are used on Brain MRI. A large variety of algorithms for segmentation of Brain MRI has been developed. The objective of this paper is to perform a segmentation process on MR images of the human brain, using Fuzzy c-means (FCM), Kernel based Fuzzy c-means clustering (KFCM), Spatial Fuzzy c-means (SFCM) and Improved Fuzzy c-means (IFCM). The review covers imaging modalities, MRI and methods for noise reduction and segmentation approaches. All methods are applied on MRI brain images which are degraded by salt-pepper noise demonstrate that the IFCM algorithm performs more robust to noise than the standard FCM algorithm. We conclude with a discussion on the trend of future research in brain segmentation and changing norms in IFCM for better results.

Keywords: image segmentation, preprocessing, MRI, FCM, KFCM, SFCM, IFCM

Procedia PDF Downloads 326
16623 Machine Learning Techniques for COVID-19 Detection: A Comparative Analysis

Authors: Abeer A. Aljohani

Abstract:

COVID-19 virus spread has been one of the extreme pandemics across the globe. It is also referred to as coronavirus, which is a contagious disease that continuously mutates into numerous variants. Currently, the B.1.1.529 variant labeled as omicron is detected in South Africa. The huge spread of COVID-19 disease has affected several lives and has surged exceptional pressure on the healthcare systems worldwide. Also, everyday life and the global economy have been at stake. This research aims to predict COVID-19 disease in its initial stage to reduce the death count. Machine learning (ML) is nowadays used in almost every area. Numerous COVID-19 cases have produced a huge burden on the hospitals as well as health workers. To reduce this burden, this paper predicts COVID-19 disease is based on the symptoms and medical history of the patient. This research presents a unique architecture for COVID-19 detection using ML techniques integrated with feature dimensionality reduction. This paper uses a standard UCI dataset for predicting COVID-19 disease. This dataset comprises symptoms of 5434 patients. This paper also compares several supervised ML techniques to the presented architecture. The architecture has also utilized 10-fold cross validation process for generalization and the principal component analysis (PCA) technique for feature reduction. Standard parameters are used to evaluate the proposed architecture including F1-Score, precision, accuracy, recall, receiver operating characteristic (ROC), and area under curve (AUC). The results depict that decision tree, random forest, and neural networks outperform all other state-of-the-art ML techniques. This achieved result can help effectively in identifying COVID-19 infection cases.

Keywords: supervised machine learning, COVID-19 prediction, healthcare analytics, random forest, neural network

Procedia PDF Downloads 88
16622 An Artificial Intelligence Framework to Forecast Air Quality

Authors: Richard Ren

Abstract:

Air pollution is a serious danger to international well-being and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.Air pollution is a serious danger to international wellbeing and economies - it will kill an estimated 7 million people every year, costing world economies $2.6 trillion by 2060 due to sick days, healthcare costs, and reduced productivity. In the United States alone, 60,000 premature deaths are caused by poor air quality. For this reason, there is a crucial need to develop effective methods to forecast air quality, which can mitigate air pollution’s detrimental public health effects and associated costs by helping people plan ahead and avoid exposure. The goal of this study is to propose an artificial intelligence framework for predicting future air quality based on timing variables (i.e. season, weekday/weekend), future weather forecasts, as well as past pollutant and air quality measurements. The proposed framework utilizes multiple machine learning algorithms (logistic regression, random forest, neural network) with different specifications and averages the results of the three top-performing models to eliminate inaccuracies, weaknesses, and biases from any one individual model. Over time, the proposed framework uses new data to self-adjust model parameters and increase prediction accuracy. To demonstrate its applicability, a prototype of this framework was created to forecast air quality in Los Angeles, California using datasets from the RP4 weather data repository and EPA pollutant measurement data. The results showed good agreement between the framework’s predictions and real-life observations, with an overall 92% model accuracy. The combined model is able to predict more accurately than any of the individual models, and it is able to reliably forecast season-based variations in air quality levels. Top air quality predictor variables were identified through the measurement of mean decrease in accuracy. This study proposed and demonstrated the efficacy of a comprehensive air quality prediction framework leveraging multiple machine learning algorithms to overcome individual algorithm shortcomings. Future enhancements should focus on expanding and testing a greater variety of modeling techniques within the proposed framework, testing the framework in different locations, and developing a platform to automatically publish future predictions in the form of a web or mobile application. Accurate predictions from this artificial intelligence framework can in turn be used to save and improve lives by allowing individuals to protect their health and allowing governments to implement effective pollution control measures.

Keywords: air quality prediction, air pollution, artificial intelligence, machine learning algorithms

Procedia PDF Downloads 120
16621 Perceived Influence of Information Communication Technology on Empowerment Amongst the College of Education Physical and Health Education Students in Oyo State

Authors: I. O. Oladipo, Olusegun Adewale Ajayi, Omoniyi Oladipupo Adigun

Abstract:

Information Communication Technology (ICT) have the potential to contribute to different facets of educational development and effective learning; expanding access, promoting efficiency, improve the quality of learning, enhancing the quality of teaching and provide important mechanism for the economic crisis. Considering the prevalence of unemployment among the higher institution graduates in this nation, in which much seems not to have been achieved in this direction. In view of this, the purpose of this study is to create an awareness and enlightenment of ICT for empowerment opportunities after school. A self-developed modified 4-likert scale questionnaire was used for data collection among Colleges of Education, Physical and Health Education students in Oyo State. Inferential statistical analysis of chi-square set at 0.05 alpha levels was used to analyze the stated hypotheses. The study concludes that awareness and enlightenment of ICT significantly influence empowerment opportunities and recommended that college of education students should be encouraged on the application of ICT for job opportunity after school.

Keywords: employment, empowerment, information communication technology, physical education

Procedia PDF Downloads 385
16620 Emotional Intelligence and Age in Open Distance Learning

Authors: Naila Naseer

Abstract:

Emotional Intelligence (EI) concept is not new yet unique and interesting. EI is a person’s ability to be aware of his/her own emotions and to manage, handle and communicate emotions with others effectively. The present study was conducted to assess the relationship between emotional intelligence and age of graduate level students at Allama Iqbal Open University (AIOU). Population consisted of Allama Iqbal Open University students (B.Ed 3rd Semester, Autumn 2007) from Rawalpindi and Islamabad regions. Total number of sample consisted of 469 participants was randomly drawn out by using table of random numbers. Bar-On EQ-i was administered on the participants through personal contact. The instrument was also validated through pilot study on a random sample of 50 participants (B.Ed students Spring 2006), who had completed their B.Ed degree successfully. Data was analyzed and tabulated in percentages, frequencies, mean, standard deviation, correlation, and scatter gram in SPSS (version 16.0 for windows). The results revealed that students with higher age group had scored low on the scale (Bar-On EQ-i). Moreover, the students in low age groups exhibited higher levels of EI as compared with old age students.

Keywords: emotional intelligence, age level, learning, emotion-related feelings

Procedia PDF Downloads 325
16619 New Advanced Medical Software Technology Challenges and Evolution of the Regulatory Framework in Expert Software, Artificial Intelligence, and Machine Learning

Authors: Umamaheswari Shanmugam, Silvia Ronchi, Radu Vornicu

Abstract:

Software, artificial intelligence, and machine learning can improve healthcare through innovative and advanced technologies that are able to use the large amount and variety of data generated during healthcare services every day. As we read the news, over 500 machine learning or other artificial intelligence medical devices have now received FDA clearance or approval, the first ones even preceding the year 2000. One of the big advantages of these new technologies is the ability to get experience and knowledge from real-world use and to continuously improve their performance. Healthcare systems and institutions can have a great benefit because the use of advanced technologies improves the same time efficiency and efficacy of healthcare. Software-defined as a medical device, is stand-alone software that is intended to be used for patients for one or more of these specific medical intended uses: - diagnosis, prevention, monitoring, prediction, prognosis, treatment or alleviation of a disease, any other health conditions, replacing or modifying any part of a physiological or pathological process–manage the received information from in vitro specimens derived from the human samples (body) and without principal main action of its principal intended use by pharmacological, immunological or metabolic definition. Software qualified as medical devices must comply with the general safety and performance requirements applicable to medical devices. These requirements are necessary to ensure high performance and quality and also to protect patients’ safety. The evolution and the continuous improvement of software used in healthcare must take into consideration the increase in regulatory requirements, which are becoming more complex in each market. The gap between these advanced technologies and the new regulations is the biggest challenge for medical device manufacturers. Regulatory requirements can be considered a market barrier, as they can delay or obstacle the device approval, but they are necessary to ensure performance, quality, and safety, and at the same time, they can be a business opportunity if the manufacturer is able to define in advance the appropriate regulatory strategy. The abstract will provide an overview of the current regulatory framework, the evolution of the international requirements, and the standards applicable to medical device software in the potential market all over the world.

Keywords: artificial intelligence, machine learning, SaMD, regulatory, clinical evaluation, classification, international requirements, MDR, 510k, PMA, IMDRF, cyber security, health care systems.

Procedia PDF Downloads 87
16618 Clustering Categorical Data Using the K-Means Algorithm and the Attribute’s Relative Frequency

Authors: Semeh Ben Salem, Sami Naouali, Moetez Sallami

Abstract:

Clustering is a well known data mining technique used in pattern recognition and information retrieval. The initial dataset to be clustered can either contain categorical or numeric data. Each type of data has its own specific clustering algorithm. In this context, two algorithms are proposed: the k-means for clustering numeric datasets and the k-modes for categorical datasets. The main encountered problem in data mining applications is clustering categorical dataset so relevant in the datasets. One main issue to achieve the clustering process on categorical values is to transform the categorical attributes into numeric measures and directly apply the k-means algorithm instead the k-modes. In this paper, it is proposed to experiment an approach based on the previous issue by transforming the categorical values into numeric ones using the relative frequency of each modality in the attributes. The proposed approach is compared with a previously method based on transforming the categorical datasets into binary values. The scalability and accuracy of the two methods are experimented. The obtained results show that our proposed method outperforms the binary method in all cases.

Keywords: clustering, unsupervised learning, pattern recognition, categorical datasets, knowledge discovery, k-means

Procedia PDF Downloads 254
16617 Virtual Academy Next: Addressing Transition Challenges Through a Gamified Virtual Transition Program for Students with Disabilities

Authors: Jennifer Gallup, Joel Bocanegra, Greg Callan, Abigail Vaughn

Abstract:

Students with disabilities (SWD) engaged in a distance summer program delivered over multiple virtual mediums that used gaming principles to teach and practice self-regulated learning (SRL) through the process of exploring possible jobs. Gaming quests were developed to explore jobs and teach transition skills. Students completed specially designed quests that taught and reinforced SRL and problem-solving through individual, group, and teacher-led experiences. SRL skills learned were reinforced through guided job explorations over the context of MinecraftEDU, zoom with experts in the career, collaborations with a team over Marco Polo, and Zoom. The quests were developed and laid out on an accessible web page, with active learning opportunities and feedback conducted within multiple virtual mediums including MinecraftEDU. Gaming mediums actively engage players in role-playing, problem-solving, critical thinking, and collaboration. Gaming has been used as a medium for education since the inception of formal education. Games, and specifically board games, are pre-historic, meaning we had board games before we had written language. Today, games are widely used in education, often as a reinforcer for behavior or for rewards for work completion. Games are not often used as a direct method of instruction and assessment; however, the inclusion of games as an assessment tool and as a form of instruction increases student engagement and participation. Games naturally include collaboration, problem-solving, and communication. Therefore, our summer program was developed using gaming principles and MinecraftEDU. This manuscript describes a virtual learning summer program called Virtual Academy New and Exciting Transitions (VAN) that was redesigned from a face-to-face setting to a completely online setting with a focus on SWD aged 14-21. The focus of VAN was to address transition planning needs such as problem-solving skills, self-regulation, interviewing, job exploration, and communication for transition-aged youth diagnosed with various disabilities (e.g., learning disabilities, attention-deficit hyperactivity disorder, intellectual disability, down syndrome, autism spectrum disorder).

Keywords: autism, disabilities, transition, summer program, gaming, simulations

Procedia PDF Downloads 70
16616 Estimation and Comparison of Delay at Signalized Intersections Based on Existing Methods

Authors: Arpita Saha, Satish Chandra, Indrajit Ghosh

Abstract:

Delay implicates the time loss of a traveler while crossing an intersection. Efficiency of traffic operation at signalized intersections is assessed in terms of delay caused to an individual vehicle. Highway Capacity Manual (HCM) method and Webster’s method are the most widely used in India for delay estimation purpose. However, in India, traffic is highly heterogeneous in nature with extremely poor lane discipline. Therefore, to explore best delay estimation technique for Indian condition, a comparison was made. In this study, seven signalized intersections from three different cities where chosen. Data was collected for both during morning and evening peak hours. Only under saturated cycles were considered for this study. Delay was estimated based on the field data. With the help of Simpson’s 1/3 rd rule, delay of under saturated cycles was estimated by measuring the area under the curve of queue length and cycle time. Moreover, the field observed delay was compared with the delay estimated using HCM, Webster, Probabilistic, Taylor’s expansion and Regression methods. The drawbacks of the existing delay estimation methods to be use in Indian heterogeneous traffic conditions were figured out, and best method was proposed. It was observed that direct estimation of delay using field measured data is more accurate than existing conventional and modified methods.

Keywords: delay estimation technique, field delay, heterogeneous traffic, signalised intersection

Procedia PDF Downloads 296
16615 Enhancement of Mulberry Leaf Yield and Water Productivity in Eastern Dry Zone of Karnataka, India

Authors: Narayanappa Devakumar, Chengalappa Seenappa

Abstract:

The field experiments were conducted during Rabi 2013 and summer 2014 at College of Sericulture, Chintamani, Chickaballapur district, Karnataka, India to find out the response of mulberry to different methods, levels of irrigation and mulching. The results showed that leaf yield and water productivity of mulberry were significantly influenced by different methods, levels of irrigation and mulching. Subsurface drip with lower level of irrigation at 0.8 CPE (Cumulative Pan Evaporation) recorded higher leaf yield and water productivity (42857 kg ha-1 yr-1and 364.41 kg hacm-1) than surface drip with higher level of irrigation at 1.0 CPE (38809 kg ha-1 yr-1 and 264.10 kg hacm-1) and micro spray jet (39931 kg ha-1 yr-1 and 271.83 kg hacm-1). Further, subsurface drip recorded minimum water used to produce one kg of leaf and to earn one rupee of profit (283 L and 113 L) compared to surface drip (390 L and 156 L) and micro spray jet (379 L and 152 L) irrigation methods. Mulberry leaf yield increased and water productivity decreased with increased levels of irrigation. However, these results indicated that irrigation of mulberry with subsurface drip increased leaf yield and water productivity by saving 20% of irrigation water than surface drip and micro spray jet irrigation methods in Eastern Dry Zone (EDZ) of Karnataka.

Keywords: cumulative pan evaporation, mulaberry, subsurface drip irrigation, water productivity

Procedia PDF Downloads 274
16614 Pedagogical Effects of Using Workbooks in English Classes for the TOEIC Test: A Study on ESL Learners in Japanese Colleges

Authors: Mikako Nobuhara

Abstract:

The Test of English for International Communication (TOEIC) test, conducted by the Institute for International Business Communication (IIBC), has a huge impact on education in Japan. Almost all college students have to submit their TOEIC test scores when applying for entry-level jobs at companies. In addition, an increasing number of colleges are encouraging students to have a global vision. For this specific reason, studying for the TOEIC test is essential for English as a second language (ESL) learner to develop English communication skills. This study shows that studying by using some workbooks about the listening section of the TOEIC test clearly helps ESL learners to develop their listening skills. For this purpose, the listening test scores before and after classroom sessions were analyzed for each student. Students obtained higher scores in the listening section of the test and improved their English listening skills at the end of all the classroom sessions. In conclusion, it is important for English teachers to achieve the following objectives: (1) facilitate the learning of effective methods for correctly solving questions based on listening skills and (2) prepare listening tasks for reading aloud so as to keep up with the original speed, which is required for solving questions in the TOEIC test.

Keywords: education, ESL, listening skills, TOEIC test

Procedia PDF Downloads 252
16613 Elaboration and Validation of a Survey about Research on the Characteristics of Mentoring of University Professors’ Lifelong Learning

Authors: Nagore Guerra Bilbao, Clemente Lobato Fraile

Abstract:

This paper outlines the design and development of the MENDEPRO questionnaire, designed to analyze mentoring performance within a professional development process carried out with professors at the University of the Basque Country, Spain. The study took into account the international research carried out over the past two decades into teachers' professional development, and was also based on a thorough review of the most common instruments used to identify and analyze mentoring styles, many of which fail to provide sufficient psychometric guarantees. The present study aimed to gather empirical data in order to verify the metric quality of the questionnaire developed. To this end, the process followed to validate the theoretical construct was as follows: The formulation of the items and indicators in accordance with the study variables; the analysis of the validity and reliability of the initial questionnaire; the review of the second version of the questionnaire and the definitive measurement instrument. Content was validated through the formal agreement and consensus of 12 university professor training experts. A reduced sample of professors who had participated in a lifelong learning program was then selected for a trial evaluation of the instrument developed. After the trial, 18 items were removed from the initial questionnaire. The final version of the instrument, comprising 33 items, was then administered to a sample group of 99 participants. The results revealed a five-dimensional structure matching theoretical expectations. Also, the reliability data for both the instrument as a whole (.98) and its various dimensions (between .91 and .97) were very high. The questionnaire was thus found to have satisfactory psychometric properties and can therefore be considered apt for studying the performance of mentoring in both induction programs for young professors and lifelong learning programs for senior faculty members.

Keywords: higher education, mentoring, professional development, university teaching

Procedia PDF Downloads 175
16612 Strategies to Achieve Deep Decarbonisation in Power Generation: A Review

Authors: Abdullah Alotaiq

Abstract:

The transition to low-carbon power generation is essential for mitigating climate change and achieving sustainability. This process, however, entails considerable costs, and understanding the factors influencing these costs is critical. This is necessary to cater to the increasing demand for low-carbon electricity across the heating, industry, and transportation sectors. A crucial aspect of this transition is identifying cost-effective and feasible paths for decarbonization, which is integral to global climate mitigation efforts. It is concluded that hybrid solutions, combining different low-carbon technologies, are optimal for minimizing costs and enhancing flexibility. These solutions also address the challenges associated with phasing out existing fossil fuel-based power plants and broadening the spectrum of low-carbon power generation options.

Keywords: review, power generation, energy transition, decarbonisation

Procedia PDF Downloads 47
16611 Digital Platform of Crops for Smart Agriculture

Authors: Pascal François Faye, Baye Mor Sall, Bineta Dembele, Jeanne Ana Awa Faye

Abstract:

In agriculture, estimating crop yields is key to improving productivity and decision-making processes such as financial market forecasting and addressing food security issues. The main objective of this paper is to have tools to predict and improve the accuracy of crop yield forecasts using machine learning (ML) algorithms such as CART , KNN and SVM . We developed a mobile app and a web app that uses these algorithms for practical use by farmers. The tests show that our system (collection and deployment architecture, web application and mobile application) is operational and validates empirical knowledge on agro-climatic parameters in addition to proactive decision-making support. The experimental results obtained on the agricultural data, the performance of the ML algorithms are compared using cross-validation in order to identify the most effective ones following the agricultural data. The proposed applications demonstrate that the proposed approach is effective in predicting crop yields and provides timely and accurate responses to farmers for decision support.

Keywords: prediction, machine learning, artificial intelligence, digital agriculture

Procedia PDF Downloads 76
16610 Password Cracking on Graphics Processing Unit Based Systems

Authors: N. Gopalakrishna Kini, Ranjana Paleppady, Akshata K. Naik

Abstract:

Password authentication is one of the widely used methods to achieve authentication for legal users of computers and defense against attackers. There are many different ways to authenticate users of a system and there are many password cracking methods also developed. This paper is mainly to propose how best password cracking can be performed on a CPU-GPGPU based system. The main objective of this work is to project how quickly a password can be cracked with some knowledge about the computer security and password cracking if sufficient security is not incorporated to the system.

Keywords: GPGPU, password cracking, secret key, user authentication

Procedia PDF Downloads 284
16609 Task Based Language Learning: A Paradigm Shift in ESL/EFL Teaching and Learning: A Case Study Based Approach

Authors: Zehra Sultan

Abstract:

The study is based on the task-based language teaching approach which is found to be very effective in the EFL/ESL classroom. This approach engages learners to acquire the usage of authentic language skills by interacting with the real world through sequence of pedagogical tasks. The use of technology enhances the effectiveness of this approach. This study throws light on the historical background of TBLT and its efficacy in the EFL/ESL classroom. In addition, this study precisely talks about the implementation of this approach in the General Foundation Programme of Muscat College, Oman. It furnishes the list of the pedagogical tasks embedded in the language curriculum of General Foundation Programme (GFP) which are skillfully allied to the College Graduate Attributes. Moreover, the study also discusses the challenges pertaining to this approach from the point of view of teachers, students, and its classroom application. Additionally, the operational success of this methodology is gauged through formative assessments of the GFP, which is apparent in the students’ progress.

Keywords: task-based language teaching, authentic language, communicative approach, real world activities, ESL/EFL activities

Procedia PDF Downloads 120