Search results for: extracurricular English learning
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8146

Search results for: extracurricular English learning

3196 The Social Media, Reading Culture and Vocabulary Expansion: Three Universities from Northern Nigeria

Authors: Nasir Umar Abdullahi

Abstract:

The social media profoundly causes the reading culture to decline in Nigeria, where the English language is a second language (SL), a language of instruction (LI), as well as the target language (TL). This is because the university students have, over the years, failed to make extensive reading their closest companion, with much emphasis on reading the European novels, newspapers, magazines, etc., so as to learn language from its original or native speakers for linguistic competence. Instead, they squander the most part of their day and nocturnal hours, sending and receiving messages through social media. The end result is their vocabulary become stagnant or ebbs, and that they cannot acquire the Cox head’s 570 vocabulary, let alone the Nation’s 2000 vocabulary to use the language fluently in writing, reading, listening, and speaking and to further compete with the native speaker in varying degrees of language usages. Be that as it may, if the social media is a monster in worsening the decline in reading culture, which degenerates in the Northern part of the country in contradistinction to the Southern part, it boosts it as well, for aside the social media language, slangs, cliché, for instance, students improve their vocabulary power, and at the same time it allows the students to privately and leisurely put the language into use, by using practically some of the vocabulary they have acquired to chart, to comment, socialize to adjudge, etc. This is what this paper tries to explore in Umaru Musa Yar’adua University Al-qalam University and the Federal University Dutin-ma.

Keywords: social media, reading, vocabulary, universities

Procedia PDF Downloads 115
3195 Efficiency of Maritime Simulator Training in Oil Spill Response Competence Development

Authors: Antti Lanki, Justiina Halonen, Juuso Punnonen, Emmi Rantavuo

Abstract:

Marine oil spill response operation requires extensive vessel maneuvering and navigation skills. At-sea oil containment and recovery include both single vessel and multi-vessel operations. Towing long oil containment booms that are several hundreds of meters in length, is a challenge in itself. Boom deployment and towing in multi-vessel configurations is an added challenge that requires precise coordination and control of the vessels. Efficient communication, as a prerequisite for shared situational awareness, is needed in order to execute the response task effectively. To gain and maintain adequate maritime skills, practical training is needed. Field exercises are the most effective way of learning, but especially the related vessel operations are resource-intensive and costly. Field exercises may also be affected by environmental limitations such as high sea-state or other adverse weather conditions. In Finland, the seasonal ice-coverage also limits the training period to summer seasons only. In addition, environmental sensitiveness of the sea area restricts the use of real oil or other target substances. This paper examines, whether maritime simulator training can offer a complementary method to overcome the training challenges related to field exercises. The objective is to assess the efficiency and the learning impact of simulator training, and the specific skills that can be trained most effectively in simulators. This paper provides an overview of learning results from two oil spill response pilot courses, in which maritime navigational bridge simulators were used to train the oil spill response authorities. The simulators were equipped with an oil spill functionality module. The courses were targeted at coastal Fire and Rescue Services responsible for near shore oil spill response in Finland. The competence levels of the participants were surveyed before and after the course in order to measure potential shifts in competencies due to the simulator training. In addition to the quantitative analysis, the efficiency of the simulator training is evaluated qualitatively through feedback from the participants. The results indicate that simulator training is a valid and effective method for developing marine oil spill response competencies that complement traditional field exercises. Simulator training provides a safe environment for assessing various oil containment and recovery tactics. One of the main benefits of the simulator training was found to be the immediate feedback the spill modelling software provides on the oil spill behaviour as a reaction to response measures.

Keywords: maritime training, oil spill response, simulation, vessel manoeuvring

Procedia PDF Downloads 152
3194 An Experience of Translating an Excerpt from Sophie Adonon’s Echos de Femmes from French to English, Using Reverso.

Authors: Michael Ngongeh Mombe

Abstract:

This Paper seeks to investigate an assertion made by some colleagues that there is no need paying a human translator to translate their literary texts, that there are softwares such as Reverso that can be used to do the translation. The main objective of this study is to examine the veracity of this assertion using Reverso to translate a literary text without any post-editing by a human translator. The work is based on two theories: Skopos and Communicative theories of translation. The work is a documentary research where data were collected from published documents in libraries, on the internet and from the translation produced by Reverso. We made a comparative text analyses of both source and target texts in a bid to highlight the weaknesses and strengths of the software. Findings of this work revealed that those who advocate the use of only Machine translation do so in ignorance of the translation mistakes usually made by the software. From the review of all the 268 segments of translation, we found out that the translation produced by Reverso is fraught with errors. We therefore recommend the use of human translators to either do the translation of their literary texts or revise the translation produced by machine to conform to the skopos of the work. This paper is based on Reverso translation. Similar works in the near future will be based on the other translation softwares to determine their weaknesses and strengths.

Keywords: machine translation, human translator, Reverso, literary text

Procedia PDF Downloads 82
3193 Machine Learning Framework: Competitive Intelligence and Key Drivers Identification of Market Share Trends among Healthcare Facilities

Authors: Anudeep Appe, Bhanu Poluparthi, Lakshmi Kasivajjula, Udai Mv, Sobha Bagadi, Punya Modi, Aditya Singh, Hemanth Gunupudi, Spenser Troiano, Jeff Paul, Justin Stovall, Justin Yamamoto

Abstract:

The necessity of data-driven decisions in healthcare strategy formulation is rapidly increasing. A reliable framework which helps identify factors impacting a healthcare provider facility or a hospital (from here on termed as facility) market share is of key importance. This pilot study aims at developing a data-driven machine learning-regression framework which aids strategists in formulating key decisions to improve the facility’s market share which in turn impacts in improving the quality of healthcare services. The US (United States) healthcare business is chosen for the study, and the data spanning 60 key facilities in Washington State and about 3 years of historical data is considered. In the current analysis, market share is termed as the ratio of the facility’s encounters to the total encounters among the group of potential competitor facilities. The current study proposes a two-pronged approach of competitor identification and regression approach to evaluate and predict market share, respectively. Leveraged model agnostic technique, SHAP, to quantify the relative importance of features impacting the market share. Typical techniques in literature to quantify the degree of competitiveness among facilities use an empirical method to calculate a competitive factor to interpret the severity of competition. The proposed method identifies a pool of competitors, develops Directed Acyclic Graphs (DAGs) and feature level word vectors, and evaluates the key connected components at the facility level. This technique is robust since its data-driven, which minimizes the bias from empirical techniques. The DAGs factor in partial correlations at various segregations and key demographics of facilities along with a placeholder to factor in various business rules (for ex. quantifying the patient exchanges, provider references, and sister facilities). Identified are the multiple groups of competitors among facilities. Leveraging the competitors' identified developed and fine-tuned Random Forest Regression model to predict the market share. To identify key drivers of market share at an overall level, permutation feature importance of the attributes was calculated. For relative quantification of features at a facility level, incorporated SHAP (SHapley Additive exPlanations), a model agnostic explainer. This helped to identify and rank the attributes at each facility which impacts the market share. This approach proposes an amalgamation of the two popular and efficient modeling practices, viz., machine learning with graphs and tree-based regression techniques to reduce the bias. With these, we helped to drive strategic business decisions.

Keywords: competition, DAGs, facility, healthcare, machine learning, market share, random forest, SHAP

Procedia PDF Downloads 78
3192 Query in Grammatical Forms and Corpus Error Analysis

Authors: Katerina Florou

Abstract:

Two decades after coined the term "learner corpora" as collections of texts created by foreign or second language learners across various language contexts, and some years following suggestion to incorporate "focusing on form" within a Task-Based Learning framework, this study aims to explore how learner corpora, whether annotated with errors or not, can facilitate a focus on form in an educational setting. Argues that analyzing linguistic form serves the purpose of enabling students to delve into language and gain an understanding of different facets of the foreign language. This same objective is applicable when analyzing learner corpora marked with errors or in their raw state, but in this scenario, the emphasis lies on identifying incorrect forms. Teachers should aim to address errors or gaps in the students' second language knowledge while they engage in a task. Building on this recommendation, we compared the written output of two student groups: the first group (G1) employed the focusing on form phase by studying a specific aspect of the Italian language, namely the past participle, through examples from native speakers and grammar rules; the second group (G2) focused on form by scrutinizing their own errors and comparing them with analogous examples from a native speaker corpus. In order to test our hypothesis, we created four learner corpora. The initial two were generated during the task phase, with one representing each group of students, while the remaining two were produced as a follow-up activity at the end of the lesson. The results of the first comparison indicated that students' exposure to their own errors can enhance their grasp of a grammatical element. The study is in its second stage and more results are to be announced.

Keywords: Corpus interlanguage analysis, task based learning, Italian language as F1, learner corpora

Procedia PDF Downloads 39
3191 Design, Synthesis and Evaluation of 4-(Phenylsulfonamido)Benzamide Derivatives as Selective Butyrylcholinesterase Inhibitors

Authors: Sushil Kumar Singh, Ashok Kumar, Ankit Ganeshpurkar, Ravi Singh, Devendra Kumar

Abstract:

In spectrum of neurodegenerative diseases, Alzheimer’s disease (AD) is characterized by the presence of amyloid β plaques and neurofibrillary tangles in the brain. It results in cognitive and memory impairment due to loss of cholinergic neurons, which is considered to be one of the contributing factors. Donepezil, an acetylcholinesterase (AChE) inhibitor which also inhibits butyrylcholinesterase (BuChE) and improves the memory and brain’s cognitive functions, is the most successful and prescribed drug to treat the symptoms of AD. The present work is based on designing of the selective BuChE inhibitors using computational techniques. In this work, machine learning models were trained using classification algorithms followed by screening of diverse chemical library of compounds. The various molecular modelling and simulation techniques were used to obtain the virtual hits. The amide derivatives of 4-(phenylsulfonamido) benzoic acid were synthesized and characterized using 1H & 13C NMR, FTIR and mass spectrometry. The enzyme inhibition assays were performed on equine plasma BuChE and electric eel’s AChE by method developed by Ellman et al. Compounds 31, 34, 37, 42, 49, 52 and 54 were found to be active against equine BuChE. N-(2-chlorophenyl)-4-(phenylsulfonamido)benzamide and N-(2-bromophenyl)-4-(phenylsulfonamido)benzamide (compounds 34 and 37) displayed IC50 of 61.32 ± 7.21 and 42.64 ± 2.17 nM against equine plasma BuChE. Ortho-substituted derivatives were more active against BuChE. Further, the ortho-halogen and ortho-alkyl substituted derivatives were found to be most active among all with minimal AChE inhibition. The compounds were selective toward BuChE.

Keywords: Alzheimer disease, butyrylcholinesterase, machine learning, sulfonamides

Procedia PDF Downloads 124
3190 Promoting Open Educational Resources (OER) in Theological/Religious Education in Nigeria

Authors: Miracle Ajah

Abstract:

One of the biggest challenges facing Theological/Religious Education in Nigeria is access to quality learning materials. For instance at the Trinity (Union) Theological College, Umuahia, it was difficult for lecturers to access suitable and qualitative materials for instruction especially the ones that would suit the African context and stimulate a deep rooted interest among the students. Some textbooks written by foreign authors were readily available in the School Library, but were lacking in the College bookshops for students to own copies. Even when the College was able to order some of the books from abroad, it did not usher in the needed enthusiasm expected from the students because they were either very expensive or very difficult to understand during private studies. So it became necessary to develop contextual materials which were affordable and understandable, though with little success. The National Open University of Nigeria (NOUN)’s innovation in the development and sharing of learning resources through its Open Course ware is a welcome development and of great assistance to students. Apart from NOUN students who could easily access the materials, many others from various theological/religious institutes across the nation have benefited immensely. So, the thesis of this paper is that the promotion of open educational resources in theological/religious education in Nigeria would facilitate a better informed/equipped religious leadership, which would in turn impact its adherents for a healthier society and national development. Adopting a narrative and historical approach within the context of Nigeria’s educational system, the paper discusses: educational traditions in Nigeria; challenges facing theological/religious education in Nigeria; and benefits of open educational resources. The study goes further to making recommendations on how OER could positively influence theological/religious education in Nigeria. It is expected that theologians, religious educators, and ODL practitioners would find this work very useful.

Keywords: OER, theological education, religious education, Nigeria

Procedia PDF Downloads 327
3189 Prediction of Covid-19 Cases and Current Situation of Italy and Its Different Regions Using Machine Learning Algorithm

Authors: Shafait Hussain Ali

Abstract:

Since its outbreak in China, the Covid_19 19 disease has been caused by the corona virus SARS N coyote 2. Italy was the first Western country to be severely affected, and the first country to take drastic measures to control the disease. In start of December 2019, the sudden outbreaks of the Coronary Virus Disease was caused by a new Corona 2 virus (SARS-CO2) of acute respiratory syndrome in china city Wuhan. The World Health Organization declared the epidemic a public health emergency of international concern on January 30, 2020,. On February 14, 2020, 49,053 laboratory-confirmed deaths and 1481 deaths have been reported worldwide. The threat of the disease has forced most of the governments to implement various control measures. Therefore it becomes necessary to analyze the Italian data very carefully, in particular to investigates and to find out the present condition and the number of infected persons in the form of positive cases, death, hospitalized or some other features of infected persons will clear in simple form. So used such a model that will clearly shows the real facts and figures and also understandable to every readable person which can get some real benefit after reading it. The model used must includes(total positive cases, current positive cases, hospitalized patients, death, recovered peoples frequency rates ) all features that explains and clear the wide range facts in very simple form and helpful to administration of that country.

Keywords: machine learning tools and techniques, rapid miner tool, Naive-Bayes algorithm, predictions

Procedia PDF Downloads 92
3188 The Attitude of Students towards the Use of the Social Networks in Education

Authors: Abdulmjeid Aljerawi

Abstract:

This study aimed to investigate the students' attitudes towards the use of social networking in education. Due to the nature of the study, and on the basis of its problem, objectives, and questions, the researcher used the descriptive approach. An appropriate questionnaire was prepared and validity and reliability were ensured. The questionnaire was then applied to the study sample of 434 students from King Saud University.

Keywords: social networks, education, learning, students

Procedia PDF Downloads 264
3187 Designing an MTB-MLE for Linguistically Heterogenous Contexts: A Practitioner’s Perspective

Authors: Ajay Pinjani, Minha Khan, Ayesha Mehkeri, Anum Iftikhar

Abstract:

There is much research available on the benefits of adopting mother tongue-based multilingual education (MTB MLE) in primary school classrooms, but there is limited guidance available on how to design such programs for low-resource and linguistically diverse contexts. This paper is an effort to bridge the gap between theory and practice by offering a practitioner’s perspective on designing an MTB MLE program for linguistically heterogeneous contexts. The research compounds findings from current academic literature on MTB MLE, the study of global MTB MLE programs, interviews with practitioners, policy-makers, and academics worldwide, and a socio-linguistic survey carried out in parts of Tharparkar, Pakistan, the area selected for envisioned pilot implementation. These findings enabled the creation of ‘guiding principles’ which provide structure for the development of a contextualized and holistic MTB-MLE program. The guiding principles direct the creation of teaching and learning materials, creating effective teaching and learning environment, community engagement, and program evaluation. Additionally, the paper demonstrates the development of a context-specific language ladder framework which outlines the language journey of a child’s education, beginning with the mother tongue/ most familiar language in the early years and then gradually transitioning into other languages. Both the guiding principles and language ladder can be adapted to any multilingual context. Thus, this research provides MTB MLE practitioners with assistance in developing an MTB MLE model, which is best suited for their context.

Keywords: mother tongue based multilingual education, education design, language ladder, language issues, heterogeneous contexts

Procedia PDF Downloads 102
3186 Communication in Inclusive Education: A Qualitative Study in Poland

Authors: Klara Królewiak-Detsi, Anna Orylska, Anna Gorgolewska, Marta Boczkowska, Agata Graczykowska

Abstract:

This study investigates the communication between students and teachers in inclusive education in Poland. Specifically, we examine the communication and interaction of students with special educational needs during online learning compared to traditional face-to-face instruction. Our research questions are (1) how children with special educational needs communicate with their teachers and peers during online learning, and (2) what strategies can improve their communication skills. We conducted five focus groups with: (1) 55 children with special educational needs, (2) 65 typically developing pupils, (3) 28 professionals (psychologists and special education therapists), (4) 16 teachers, and (5) 16 parents of children with special educational needs. Our analysis focused on primary schools and used thematic analysis according to the 6-step procedure of Braun and Clarke. Our findings reveal that children with disabilities faced more difficulties communicating and interacting with others online than in face-to-face lessons. The online tools used for education were not adapted to the needs of children with disabilities, and schools lacked clear guidelines on how to pursue inclusive education online. Based on the results, we offer recommendations for online communication training and tools that are dedicated to children with special educational needs. Additionally, our results demonstrate that typically developing pupils are better in interpersonal relations and more often and effectively use social support. Children with special educational needs had similar emotional and communication challenges compared to their typically developing peers. In conclusion, our study highlights the importance of providing adequate support for the online education of children with special educational needs in inclusive classrooms.

Keywords: Inclusive education, Special educational needs, Social skills development, Online communication

Procedia PDF Downloads 111
3185 Investigating Best Strategies Towards Creating Alternative Assessment in Literature

Authors: Sandhya Rao Mehta

Abstract:

As ChatGpt and other Artificial Intelligence (AI) forms are becoming part of our regular academic world, the consequences are being gradually discussed. The extent to which an essay written by a student is itself of any value if it has been downloaded by some form of AI is perhaps central to this discourse. A larger question is whether writing should be taught as an academic skill at all. In literature classrooms, this has major consequences as writing a traditional paper is still the single most preferred form of assessment. This study suggests that it is imperative to investigate alternative forms of assessment in literature, not only because the existing forms can be written by AI, but in a larger sense, students are increasingly skeptical of the purpose of such work. The extent to which an essay actually helps the students professionally is a question that academia has not yet answered. This paper suggests that using real-world tasks like creating podcasts, video tutorials, and websites is a far better way to evaluate students' critical thinking and application of ideas, as well as to develop digital skills which are important to their future careers. Using the example of a course in literature, this study will examine the possibilities and challenges of creating digital projects as a way of confronting the complexities of student evaluation in the future. The study is based on a specific university English as a Foreign Language (EFL) context.

Keywords: assessment, literature, digital humanities, chatgpt

Procedia PDF Downloads 69
3184 A Systematic Review on The Usage of CRISPR-Cas System in The Treatment of Osteoarthritis(OA)

Authors: Atiqah Binti Ab Aziz

Abstract:

Background: It has been estimated that about 250 million people all over the world suffer from osteoarthritis (OA). Thus, OA is a major health problem in urgent need of better treatment. Problem statement: Current therapies for OA can temporarily relieve clinical symptoms and for pain management, rather than preventing or curing OA. Total knee replacement performed at the end stage of the disease is considered the only cure available. Objectives: This article aimed to explore the potential of treating osteoarthritis via the CRISPR Cas system. Methods: Articles that relate to the application of the CRISPR Cas system in osteoarthritis were extracted, categorized, and reviewed through the PRISMA method using PubMed, an engine published from November 2016 to November 2021. Results: There were 30 articles screened. Articles that fall under the categories of non-English articles, full articles that were not available, articles that were not an original articles were excluded. Ultimately, 13 articles were reviewed. Discussion: This review provides an information on the introduction of CRISPR and discussed on their mechanism of actions in extracted studies for OA treatment. Conclusions: It can be seen that not many medical research utilize the CRISPR Cas system as part of the method in the treatment of OA. Hence exploring the extent of the usage of the CRISPR Cas system in OA treatment is important to determine the research gap and point out at which of the research is needed further investigation to avoid redundancy of existing research and ensure the novelty of the research.

Keywords: osteoarthritis, treatment, CRISPR, review, therapy

Procedia PDF Downloads 158
3183 Educational Sustainability: Teaching the Next Generation of Educators in Medical Simulation

Authors: Thomas Trouton, Sebastian Tanner, Manvir Sandher

Abstract:

The use of simulation in undergraduate and postgraduate medical curricula is ever-growing, is a useful addition to the traditional apprenticeship model of learning within medical education, and better prepares graduates for the team-based approach to healthcare seen in real-life clinical practice. As a learning tool, however, undergraduate medical students often have little understanding of the theory behind the use of medical simulation and have little experience in planning and delivering their own simulated teaching sessions. We designed and implemented a student-selected component (SSC) as part of the undergraduate medical curriculum at the University of Buckingham Medical School to introduce students to the concepts behind the use of medical simulation in education and allow them to plan and deliver their own simulated medical scenario to their peers. The SSC took place over a 2-week period in the 3rd year of the undergraduate course. There was a mix of lectures, seminars and interactive group work sessions, as well as hands-on experience in the simulation suite, to introduce key concepts related to medical simulation, including technical considerations in simulation, human factors, debriefing and troubleshooting scenarios. We evaluated the success of our SSC using “Net Promotor Scores” (NPS) to assess students’ confidence in planning and facilitating a simulation-based teaching session, as well as leading a debrief session. In all three domains, we showed an increase in the confidence of the students. We also showed an increase in confidence in the management of common medical emergencies as a result of the SSC. Overall, the students who chose our SSC had the opportunity to learn new skills in medical education, with a particular focus on the use of simulation-based teaching, and feedback highlighted that a number of students would take these skills forward in their own practice. We demonstrated an increase in confidence in several domains related to the use of medical simulation in education and have hopefully inspired a new generation of medical educators.

Keywords: simulation, SSC, teaching, medical students

Procedia PDF Downloads 108
3182 Don't Just Guess and Slip: Estimating Bayesian Knowledge Tracing Parameters When Observations Are Scant

Authors: Michael Smalenberger

Abstract:

Intelligent tutoring systems (ITS) are computer-based platforms which can incorporate artificial intelligence to provide step-by-step guidance as students practice problem-solving skills. ITS can replicate and even exceed some benefits of one-on-one tutoring, foster transactivity in collaborative environments, and lead to substantial learning gains when used to supplement the instruction of a teacher or when used as the sole method of instruction. A common facet of many ITS is their use of Bayesian Knowledge Tracing (BKT) to estimate parameters necessary for the implementation of the artificial intelligence component, and for the probability of mastery of a knowledge component relevant to the ITS. While various techniques exist to estimate these parameters and probability of mastery, none directly and reliably ask the user to self-assess these. In this study, 111 undergraduate students used an ITS in a college-level introductory statistics course for which detailed transaction-level observations were recorded, and users were also routinely asked direct questions that would lead to such a self-assessment. Comparisons were made between these self-assessed values and those obtained using commonly used estimation techniques. Our findings show that such self-assessments are particularly relevant at the early stages of ITS usage while transaction level data are scant. Once a user’s transaction level data become available after sufficient ITS usage, these can replace the self-assessments in order to eliminate the identifiability problem in BKT. We discuss how these findings are relevant to the number of exercises necessary to lead to mastery of a knowledge component, the associated implications on learning curves, and its relevance to instruction time.

Keywords: Bayesian Knowledge Tracing, Intelligent Tutoring System, in vivo study, parameter estimation

Procedia PDF Downloads 160
3181 Mathematics Professional Development: Uptake and Impacts on Classroom Practice

Authors: Karen Koellner, Nanette Seago, Jennifer Jacobs, Helen Garnier

Abstract:

Although studies of teacher professional development (PD) are prevalent, surprisingly most have only produced incremental shifts in teachers’ learning and their impact on students. There is a critical need to understand what teachers take up and use in their classroom practice after attending PD and why we often do not see greater changes in learning and practice. This paper is based on a mixed methods efficacy study of the Learning and Teaching Geometry (LTG) video-based mathematics professional development materials. The extent to which the materials produce a beneficial impact on teachers’ mathematics knowledge, classroom practices, and their students’ knowledge in the domain of geometry through a group-randomized experimental design are considered. Included is a close-up examination of a small group of teachers to better understand their interpretations of the workshops and their classroom uptake. The participants included 103 secondary mathematics teachers serving grades 6-12 from two US states in different regions. Randomization was conducted at the school level, with 23 schools and 49 teachers assigned to the treatment group and 18 schools and 54 teachers assigned to the comparison group. The case study examination included twelve treatment teachers. PD workshops for treatment teachers began in Summer 2016. Nine full days of professional development were offered to teachers, beginning with the one-week institute (Summer 2016) and four days of PD throughout the academic year. The same facilitator-led all of the workshops, after completing a facilitator preparation process that included a multi-faceted assessment of fidelity. The overall impact of the LTG PD program was assessed from multiple sources: two teacher content assessments, two PD embedded assessments, pre-post-post videotaped classroom observations, and student assessments. Additional data were collected from the case study teachers including additional videotaped classroom observations and interviews. Repeated measures ANOVA analyses were used to detect patterns of change in the treatment teachers’ content knowledge before and after completion of the LTG PD, relative to the comparison group. No significant effects were found across the two groups of teachers on the two teacher content assessments. Teachers were rated on the quality of their mathematics instruction captured in videotaped classroom observations using the Math in Common Observation Protocol. On average, teachers who attended the LTG PD intervention improved their ability to engage students in mathematical reasoning and to provide accurate, coherent, and well-justified mathematical content. In addition, the LTG PD intervention and instruction that engaged students in mathematical practices both positively and significantly predicted greater student knowledge gains. Teacher knowledge was not a significant predictor. Twelve treatment teachers self-selected to serve as case study teachers to provide additional videotapes in which they felt they were using something from the PD they learned and experienced. Project staff analyzed the videos, compared them to previous videos and interviewed the teachers regarding their uptake of the PD related to content knowledge, pedagogical knowledge and resources used. The full paper will include the case study of Ana to illustrate the factors involved in what teachers take up and use from participating in the LTG PD.

Keywords: geometry, mathematics professional development, pedagogical content knowledge, teacher learning

Procedia PDF Downloads 110
3180 Artificial Intelligent Tax Simulator to Minimize Tax Liability for Multinational Corporations

Authors: Sean Goltz, Michael Mayo

Abstract:

The purpose of this research is to use Global-Regulation.com database of the world laws, focusing on tax treaties between countries, in order to create an AI-driven tax simulator that will run an AI agent through potential tax scenarios across countries. The AI agent goal is to identify the scenario that will result in minimum tax liability based on tax treaties between countries. The results will be visualized by a three dimensional matrix. This will be an online web application. Multinational corporations are running their business through multiple countries. These countries, in turn, have a tax treaty with many other countries to regulate the payment of taxes on income that is transferred between these countries. As a result, planning the best tax scenario across multiple countries and numerous tax treaties is almost impossible. This research propose to use Global-Regulation.com database of word laws in English (machine translated by Google and Microsoft API’s) in order to create a simulator that will include the information in the tax treaties. Once ready, an AI agent will be sent through the simulator to identify the scenario that will result in minimum tax liability. Identifying the best tax scenario across countries may save multinational corporations, like Google, billions of dollars annually. Given the nature of the raw data and the domain of taxes (i.e., numbers), this is a promising ground to employ artificial intelligence towards a practical and beneficial purpose.

Keywords: taxation, law, multinational, corporation

Procedia PDF Downloads 183
3179 Stuck Spaces as Moments of Learning: Uncovering Threshold Concepts in Teacher Candidate Experiences of Teaching in Inclusive Classrooms

Authors: Joy Chadwick

Abstract:

There is no doubt that classrooms of today are more complex and diverse than ever before. Preparing teacher candidates to meet these challenges is essential to ensure the retention of teachers within the profession and to ensure that graduates begin their teaching careers with the knowledge and understanding of how to effectively meet the diversity of students they will encounter. Creating inclusive classrooms requires teachers to have a repertoire of effective instructional skills and strategies. Teachers must also have the mindset to embrace diversity and value the uniqueness of individual students in their care. This qualitative study analyzed teacher candidates' experiences as they completed a fourteen-week teaching practicum while simultaneously completing a university course focused on inclusive pedagogy. The research investigated the challenges and successes teacher candidates had in navigating the translation of theory related to inclusive pedagogy into their teaching practice. Applying threshold concept theory as a framework, the research explored the troublesome concepts, liminal spaces, and transformative experiences as connected to inclusive practices. Threshold concept theory suggests that within all disciplinary fields, there exists particular threshold concepts that serve as gateways or portals into previously inaccessible ways of thinking and practicing. It is in these liminal spaces that conceptual shifts in thinking and understanding and deep learning can occur. The threshold concept framework provided a lens to examine teacher candidate struggles and successes with the inclusive education course content and the application of this content to their practicum experiences. A qualitative research approach was used, which included analyzing twenty-nine course reflective journals and six follow up one-to-one semi structured interviews. The journals and interview transcripts were coded and themed using NVivo software. Threshold concept theory was then applied to the data to uncover the liminal or stuck spaces of learning and the ways in which the teacher candidates navigated those challenging places of teaching. The research also sought to uncover potential transformative shifts in teacher candidate understanding as connected to teaching in an inclusive classroom. The findings suggested that teacher candidates experienced difficulties when they did not feel they had the knowledge, skill, or time to meet the needs of the students in the way they envisioned they should. To navigate the frustration of this thwarted vision, they relied on present and previous course content and experiences, collaborative work with other teacher candidates and their mentor teachers, and a proactive approach to planning for students. Transformational shifts were most evident in their ability to reframe their perceptions of children from a deficit or disability lens to a strength-based belief in the potential of students. It was evident that through their course work and practicum experiences, their beliefs regarding struggling students shifted as they saw the value of embracing neurodiversity, the importance of relationships, and planning for and teaching through a strength-based approach. Research findings have implications for teacher education programs and for understanding threshold concepts theory as connected to practice-based learning experiences.

Keywords: inclusion, inclusive education, liminal space, teacher education, threshold concepts, troublesome knowledge

Procedia PDF Downloads 61
3178 Enhanced Multi-Scale Feature Extraction Using a DCNN by Proposing Dynamic Soft Margin SoftMax for Face Emotion Detection

Authors: Armin Nabaei, M. Omair Ahmad, M. N. S. Swamy

Abstract:

Many facial expression and emotion recognition methods in the traditional approaches of using LDA, PCA, and EBGM have been proposed. In recent years deep learning models have provided a unique platform addressing by automatically extracting the features for the detection of facial expression and emotions. However, deep networks require large training datasets to extract automatic features effectively. In this work, we propose an efficient emotion detection algorithm using face images when only small datasets are available for training. We design a deep network whose feature extraction capability is enhanced by utilizing several parallel modules between the input and output of the network, each focusing on the extraction of different types of coarse features with fined grained details to break the symmetry of produced information. In fact, we leverage long range dependencies, which is one of the main drawback of CNNs. We develop this work by introducing a Dynamic Soft-Margin SoftMax.The conventional SoftMax suffers from reaching to gold labels very soon, which take the model to over-fitting. Because it’s not able to determine adequately discriminant feature vectors for some variant class labels. We reduced the risk of over-fitting by using a dynamic shape of input tensor instead of static in SoftMax layer with specifying a desired Soft- Margin. In fact, it acts as a controller to how hard the model should work to push dissimilar embedding vectors apart. For the proposed Categorical Loss, by the objective of compacting the same class labels and separating different class labels in the normalized log domain.We select penalty for those predictions with high divergence from ground-truth labels.So, we shorten correct feature vectors and enlarge false prediction tensors, it means we assign more weights for those classes with conjunction to each other (namely, “hard labels to learn”). By doing this work, we constrain the model to generate more discriminate feature vectors for variant class labels. Finally, for the proposed optimizer, our focus is on solving weak convergence of Adam optimizer for a non-convex problem. Our noteworthy optimizer is working by an alternative updating gradient procedure with an exponential weighted moving average function for faster convergence and exploiting a weight decay method to help drastically reducing the learning rate near optima to reach the dominant local minimum. We demonstrate the superiority of our proposed work by surpassing the first rank of three widely used Facial Expression Recognition datasets with 93.30% on FER-2013, and 16% improvement compare to the first rank after 10 years, reaching to 90.73% on RAF-DB, and 100% k-fold average accuracy for CK+ dataset, and shown to provide a top performance to that provided by other networks, which require much larger training datasets.

Keywords: computer vision, facial expression recognition, machine learning, algorithms, depp learning, neural networks

Procedia PDF Downloads 60
3177 Predictive Maintenance: Machine Condition Real-Time Monitoring and Failure Prediction

Authors: Yan Zhang

Abstract:

Predictive maintenance is a technique to predict when an in-service machine will fail so that maintenance can be planned in advance. Analytics-driven predictive maintenance is gaining increasing attention in many industries such as manufacturing, utilities, aerospace, etc., along with the emerging demand of Internet of Things (IoT) applications and the maturity of technologies that support Big Data storage and processing. This study aims to build an end-to-end analytics solution that includes both real-time machine condition monitoring and machine learning based predictive analytics capabilities. The goal is to showcase a general predictive maintenance solution architecture, which suggests how the data generated from field machines can be collected, transmitted, stored, and analyzed. We use a publicly available aircraft engine run-to-failure dataset to illustrate the streaming analytics component and the batch failure prediction component. We outline the contributions of this study from four aspects. First, we compare the predictive maintenance problems from the view of the traditional reliability centered maintenance field, and from the view of the IoT applications. When evolving to the IoT era, predictive maintenance has shifted its focus from ensuring reliable machine operations to improve production/maintenance efficiency via any maintenance related tasks. It covers a variety of topics, including but not limited to: failure prediction, fault forecasting, failure detection and diagnosis, and recommendation of maintenance actions after failure. Second, we review the state-of-art technologies that enable a machine/device to transmit data all the way through the Cloud for storage and advanced analytics. These technologies vary drastically mainly based on the power source and functionality of the devices. For example, a consumer machine such as an elevator uses completely different data transmission protocols comparing to the sensor units in an environmental sensor network. The former may transfer data into the Cloud via WiFi directly. The latter usually uses radio communication inherent the network, and the data is stored in a staging data node before it can be transmitted into the Cloud when necessary. Third, we illustrate show to formulate a machine learning problem to predict machine fault/failures. By showing a step-by-step process of data labeling, feature engineering, model construction and evaluation, we share following experiences: (1) what are the specific data quality issues that have crucial impact on predictive maintenance use cases; (2) how to train and evaluate a model when training data contains inter-dependent records. Four, we review the tools available to build such a data pipeline that digests the data and produce insights. We show the tools we use including data injection, streaming data processing, machine learning model training, and the tool that coordinates/schedules different jobs. In addition, we show the visualization tool that creates rich data visualizations for both real-time insights and prediction results. To conclude, there are two key takeaways from this study. (1) It summarizes the landscape and challenges of predictive maintenance applications. (2) It takes an example in aerospace with publicly available data to illustrate each component in the proposed data pipeline and showcases how the solution can be deployed as a live demo.

Keywords: Internet of Things, machine learning, predictive maintenance, streaming data

Procedia PDF Downloads 374
3176 Data Analysis Tool for Predicting Water Scarcity in Industry

Authors: Tassadit Issaadi Hamitouche, Nicolas Gillard, Jean Petit, Valerie Lavaste, Celine Mayousse

Abstract:

Water is a fundamental resource for the industry. It is taken from the environment either from municipal distribution networks or from various natural water sources such as the sea, ocean, rivers, aquifers, etc. Once used, water is discharged into the environment, reprocessed at the plant or treatment plants. These withdrawals and discharges have a direct impact on natural water resources. These impacts can apply to the quantity of water available, the quality of the water used, or to impacts that are more complex to measure and less direct, such as the health of the population downstream from the watercourse, for example. Based on the analysis of data (meteorological, river characteristics, physicochemical substances), we wish to predict water stress episodes and anticipate prefectoral decrees, which can impact the performance of plants and propose improvement solutions, help industrialists in their choice of location for a new plant, visualize possible interactions between companies to optimize exchanges and encourage the pooling of water treatment solutions, and set up circular economies around the issue of water. The development of a system for the collection, processing, and use of data related to water resources requires the functional constraints specific to the latter to be made explicit. Thus the system will have to be able to store a large amount of data from sensors (which is the main type of data in plants and their environment). In addition, manufacturers need to have 'near-real-time' processing of information in order to be able to make the best decisions (to be rapidly notified of an event that would have a significant impact on water resources). Finally, the visualization of data must be adapted to its temporal and geographical dimensions. In this study, we set up an infrastructure centered on the TICK application stack (for Telegraf, InfluxDB, Chronograf, and Kapacitor), which is a set of loosely coupled but tightly integrated open source projects designed to manage huge amounts of time-stamped information. The software architecture is coupled with the cross-industry standard process for data mining (CRISP-DM) data mining methodology. The robust architecture and the methodology used have demonstrated their effectiveness on the study case of learning the level of a river with a 7-day horizon. The management of water and the activities within the plants -which depend on this resource- should be considerably improved thanks, on the one hand, to the learning that allows the anticipation of periods of water stress, and on the other hand, to the information system that is able to warn decision-makers with alerts created from the formalization of prefectoral decrees.

Keywords: data mining, industry, machine Learning, shortage, water resources

Procedia PDF Downloads 110
3175 Combining Diffusion Maps and Diffusion Models for Enhanced Data Analysis

Authors: Meng Su

Abstract:

High-dimensional data analysis often presents challenges in capturing the complex, nonlinear relationships and manifold structures inherent to the data. This article presents a novel approach that leverages the strengths of two powerful techniques, Diffusion Maps and Diffusion Probabilistic Models (DPMs), to address these challenges. By integrating the dimensionality reduction capability of Diffusion Maps with the data modeling ability of DPMs, the proposed method aims to provide a comprehensive solution for analyzing and generating high-dimensional data. The Diffusion Map technique preserves the nonlinear relationships and manifold structure of the data by mapping it to a lower-dimensional space using the eigenvectors of the graph Laplacian matrix. Meanwhile, DPMs capture the dependencies within the data, enabling effective modeling and generation of new data points in the low-dimensional space. The generated data points can then be mapped back to the original high-dimensional space, ensuring consistency with the underlying manifold structure. Through a detailed example implementation, the article demonstrates the potential of the proposed hybrid approach to achieve more accurate and effective modeling and generation of complex, high-dimensional data. Furthermore, it discusses possible applications in various domains, such as image synthesis, time-series forecasting, and anomaly detection, and outlines future research directions for enhancing the scalability, performance, and integration with other machine learning techniques. By combining the strengths of Diffusion Maps and DPMs, this work paves the way for more advanced and robust data analysis methods.

Keywords: diffusion maps, diffusion probabilistic models (DPMs), manifold learning, high-dimensional data analysis

Procedia PDF Downloads 87
3174 Automated Adaptions of Semantic User- and Service Profile Representations by Learning the User Context

Authors: Nicole Merkle, Stefan Zander

Abstract:

Ambient Assisted Living (AAL) describes a technological and methodological stack of (e.g. formal model-theoretic semantics, rule-based reasoning and machine learning), different aspects regarding the behavior, activities and characteristics of humans. Hence, a semantic representation of the user environment and its relevant elements are required in order to allow assistive agents to recognize situations and deduce appropriate actions. Furthermore, the user and his/her characteristics (e.g. physical, cognitive, preferences) need to be represented with a high degree of expressiveness in order to allow software agents a precise evaluation of the users’ context models. The correct interpretation of these context models highly depends on temporal, spatial circumstances as well as individual user preferences. In most AAL approaches, model representations of real world situations represent the current state of a universe of discourse at a given point in time by neglecting transitions between a set of states. However, the AAL domain currently lacks sufficient approaches that contemplate on the dynamic adaptions of context-related representations. Semantic representations of relevant real-world excerpts (e.g. user activities) help cognitive, rule-based agents to reason and make decisions in order to help users in appropriate tasks and situations. Furthermore, rules and reasoning on semantic models are not sufficient for handling uncertainty and fuzzy situations. A certain situation can require different (re-)actions in order to achieve the best results with respect to the user and his/her needs. But what is the best result? To answer this question, we need to consider that every smart agent requires to achieve an objective, but this objective is mostly defined by domain experts who can also fail in their estimation of what is desired by the user and what not. Hence, a smart agent has to be able to learn from context history data and estimate or predict what is most likely in certain contexts. Furthermore, different agents with contrary objectives can cause collisions as their actions influence the user’s context and constituting conditions in unintended or uncontrolled ways. We present an approach for dynamically updating a semantic model with respect to the current user context that allows flexibility of the software agents and enhances their conformance in order to improve the user experience. The presented approach adapts rules by learning sensor evidence and user actions using probabilistic reasoning approaches, based on given expert knowledge. The semantic domain model consists basically of device-, service- and user profile representations. In this paper, we present how this semantic domain model can be used in order to compute the probability of matching rules and actions. We apply this probability estimation to compare the current domain model representation with the computed one in order to adapt the formal semantic representation. Our approach aims at minimizing the likelihood of unintended interferences in order to eliminate conflicts and unpredictable side-effects by updating pre-defined expert knowledge according to the most probable context representation. This enables agents to adapt to dynamic changes in the environment which enhances the provision of adequate assistance and affects positively the user satisfaction.

Keywords: ambient intelligence, machine learning, semantic web, software agents

Procedia PDF Downloads 266
3173 Ontology-Driven Knowledge Discovery and Validation from Admission Databases: A Structural Causal Model Approach for Polytechnic Education in Nigeria

Authors: Bernard Igoche Igoche, Olumuyiwa Matthew, Peter Bednar, Alexander Gegov

Abstract:

This study presents an ontology-driven approach for knowledge discovery and validation from admission databases in Nigerian polytechnic institutions. The research aims to address the challenges of extracting meaningful insights from vast amounts of admission data and utilizing them for decision-making and process improvement. The proposed methodology combines the knowledge discovery in databases (KDD) process with a structural causal model (SCM) ontological framework. The admission database of Benue State Polytechnic Ugbokolo (Benpoly) is used as a case study. The KDD process is employed to mine and distill knowledge from the database, while the SCM ontology is designed to identify and validate the important features of the admission process. The SCM validation is performed using the conditional independence test (CIT) criteria, and an algorithm is developed to implement the validation process. The identified features are then used for machine learning (ML) modeling and prediction of admission status. The results demonstrate the adequacy of the SCM ontological framework in representing the admission process and the high predictive accuracies achieved by the ML models, with k-nearest neighbors (KNN) and support vector machine (SVM) achieving 92% accuracy. The study concludes that the proposed ontology-driven approach contributes to the advancement of educational data mining and provides a foundation for future research in this domain.

Keywords: admission databases, educational data mining, machine learning, ontology-driven knowledge discovery, polytechnic education, structural causal model

Procedia PDF Downloads 46
3172 Evaluating the Satisfaction of Chinese Consumers toward Influencers at TikTok

Authors: Noriyuki Suyama

Abstract:

The progress and spread of digitalization have led to the provision of a variety of new services. The recent progress in digitization can be attributed to rapid developments in science and technology. First, the research and diffusion of artificial intelligence (AI) has made dramatic progress. Around 2000, the third wave of AI research, which had been underway for about 50 years, arrived. Specifically, machine learning and deep learning were made possible in AI, and the ability of AI to acquire knowledge, define the knowledge, and update its own knowledge in a quantitative manner made the use of big data practical even for commercial PCs. On the other hand, with the spread of social media, information exchange has become more common in our daily lives, and the lending and borrowing of goods and services, in other words, the sharing economy, has become widespread. The scope of this trend is not limited to any industry, and its momentum is growing as the SDGs take root. In addition, the Social Network Service (SNS), a part of social media, has brought about the evolution of the retail business. In the past few years, social network services (SNS) involving users or companies have especially flourished. The People's Republic of China (hereinafter referred to as "China") is a country that is stimulating enormous consumption through its own unique SNS, which is different from the SNS used in developed countries around the world. This paper focuses on the effectiveness and challenges of influencer marketing by focusing on the influence of influencers on users' behavior and satisfaction with Chinese SNSs. Specifically, Conducted was the quantitative survey of Tik Tok users living in China, with the aim of gaining new insights from the analysis and discussions. As a result, we found several important findings and knowledge.

Keywords: customer satisfaction, social networking services, influencer marketing, Chinese consumers’ behavior

Procedia PDF Downloads 82
3171 Training AI to Be Empathetic and Determining the Psychotype of a Person During a Conversation with a Chatbot

Authors: Aliya Grig, Konstantin Sokolov, Igor Shatalin

Abstract:

The report describes the methodology for collecting data and building an ML model for determining the personality psychotype using profiling and personality traits methods based on several short messages of a user communicating on an arbitrary topic with a chitchat bot. In the course of the experiments, the minimum amount of text was revealed to confidently determine aspects of personality. Model accuracy - 85%. Users' language of communication is English. AI for a personalized communication with a user based on his mood, personality, and current emotional state. Features investigated during the research: personalized communication; providing empathy; adaptation to a user; predictive analytics. In the report, we describe the processes that captures both structured and unstructured data pertaining to a user in large quantities and diverse forms. This data is then effectively processed through ML tools to construct a knowledge graph and draw inferences regarding users of text messages in a comprehensive manner. Specifically, the system analyzes users' behavioral patterns and predicts future scenarios based on this analysis. As a result of the experiments, we provide for further research on training AI models to be empathetic, creating personalized communication for a user

Keywords: AI, empathetic, chatbot, AI models

Procedia PDF Downloads 75
3170 Contextualizing the Translation and Reconstruction of “Zhengqueyiliguan” in Promoting China-Africa Diplomatic Discourse

Authors: Queen Modestus

Abstract:

China’s diplomatic discourse is a complicated system underpinned by divergent translations. As such, the concept of “Zhengque Yi Li Guan” (or zhèngquè Yì Lì Guān in Pinyin), like numerous other Chinese concepts, is not well appreciated due, mainly, to the dearth of a uniform official translation and poor interpretation. So, the paper seeks to draw awareness to this vital issue that has been neglected in the China-Africa diplomatic discourse. As a way of filling this gap in the literature, the author argues that literal translation is not enough in diplomatic discourse. To identify the English equivalent of “Zhenque,” “Yi,” “Li,” and “Guan,” searches of Chinese classical works were carried out. Based on the scope of this paper, emphasis will be placed on “Yi” because it plays a critical role in the “zhèngquèyìlìguān” concept and has been the term with the most controversial interpretation. From a diplomatic-linguistic standpoint coupled with the help of the Critical Discourse Analysis (CDA) developed by Norman Fairclough and Teun A. van Dijk as this paper’s theoretical framework, this paper seeks to address the above issue by exploring the philosophical translations and interpretations of ZqYLG in China-Africa diplomatic discourse, with emphasis place on the discourse reconstruction of “Yi.” As a way of reconstructing “Yi” to fit the present reality of China-Africa diplomatic discourse, the author argues that “justice” (meaning impartiality) is a better translation for “Yi.”

Keywords: Africa, diplomatic discourse, reconstruction, translation, 正确义利观, interests

Procedia PDF Downloads 73
3169 Analyzing the Perceptions of Accounting Practitioners regarding Communication Skills of Distance-Learning Graduates

Authors: Carol S. Binnekade, Deon Scott, Christina C. Shuttleworth, Annelien A. Van Rooyen

Abstract:

Higher education institutions are constantly challenged to deliver skilled graduates into the workplace. Employers expect graduates to have the required technical knowledge as well as various pervasive skills. This also applies to accountants who need to know the technical requirements of financial reporting and be able to communicate with individuals, teams and clients at a high level. Accountants need to develop effective business conversational skills and use these skills to communicate up, down and across organizations, taking into consideration cultural and gender diversity. In addition, they need to master business writing and presentation skills. However, providing students with these skills in a distance-learning environment where interaction between students and instructors is limited, is a challenge for academics. The study on which this paper reports, forms part of a larger body of research, which explored the perceptions of accounting practitioners of the communication skills (or lack thereof) of recently qualified accounting students. Feedback (qualitative and quantitative) was obtained from various accounting practitioners in South Africa. Taking into consideration that distance learners communicate mainly with their instructors via email communication and their assignments are submitted using various word processor software, the researchers were of the opinion that the accounting graduates would be capable of communicating effectively once they entered the workplace. However, the research findings, inter alia, suggested that the accounting graduates lacked communication skills and that training was needed to differentiate between business and social communication once they entered the workplace. Recommendations on how these communication challenges may be addressed by higher education institutions are provided.

Keywords: accounting practitioners, communication skills, distance education, pervasive skills

Procedia PDF Downloads 192
3168 Beyond Rhetoric and Buzzword, Policies and Politics: Towards Practical Institutional Involvement in Science and Technology Teacher Education Programmes for Sustainable Development

Authors: Alvin Uchenna Ugwu

Abstract:

The United Nation’s 2030 agenda and Global Action Programme (GAP) for implementation of the Sustainable Development Goals (SDGs), has mandated all sectors in the societies, including education, to develop strategies towards actualizing sustainability in all facets of the society, by the year 2030. Education is no doubt a key tool for social change. However, educational institutions in most African nations need a paradigmatic shift to strike a balance between policies (curricular) and practices, with regards to Education for Sustainable Development (ESD). The paradigm shift in this regard is described as whole-institution/school approach. The whole institution approaches advocate action-focused ESD. In other words, ESD policy and curriculum makers, formal and non-formal education institutions, need to ‘practice what they preach’. This paper is developed from an ongoing study carried out by the author and guided by two research questions: -What are the views of intermediate phase science and technology preservice teachers on the ESD content included in the science and technology modules? -What challenges or enable intermediate phase science and technology pre-service teachers to learn about ESD in science and technology modules? The study drew from the views and experiences of preservice science teachers, learning about ESD in a university’s college of education in South Africa. Using qualitative case study research design, the research data were generated via questionnaires and focus group discussions. Analysis of generated data indicates that universities and institutions of higher learning need to demonstrate practical involvement while implementing ESD in societies, rather than just standing as knowledge media. Findings of the study further suggest that natural sciences and technology courses in teacher education programmes and other institutions of higher learning, should be perceived as key transformative tools in shaping the consciousness of students towards integrating and fostering ESD in developing countries such as South Africa. Thus, this paper seeks to promote ‘Whole Institution Involvement’ in teacher education colleges in South Africa, as a measure of improving ESD in higher education settings. The paper suggests that in order to achieve ESD in higher education settings and beyond, policies and practices should be reexamined beyond rhetoric and buzzwords. The paper further argues that implementation of ESD is largely influenced by context, hence two different contexts should be examined empirically.

Keywords: education for sustainable development, higher education institutions, pre-service science teachers, qualitative case study research, whole institution involvement

Procedia PDF Downloads 156
3167 The Interplay of Locus of Control, Academic Achievement and Biological Variables among Iranian Online EFL Learners

Authors: Azizeh Chalak, Niloufar Nasri

Abstract:

Students' academic achievement, along with the effects of different variables, has been a serious concern of educators since long ago. This study was an attempt to investigate the interplay of Locus of Control (LOC), academic achievement and biological variables among Iranian online EFL Learners. The participants of the study included 100 students of different age groups and genders studying English online at Iran Language Institute (ILI), Isfahan, Iran. The instrument used was Trice Academic LOC questionnaire which identifies orientations of internality or externality. The participants' Grade Point Averages (GPAs) were used as the measure of their academic achievement. A series of independent samples t-tests were performed on the data. The results of the study showed that (a) there were no significant differences between male and female participants in LOC orientation, (b) there was no relationship between LOC and academic achievement among internal males and females, (c) external females were better achievers than external males, (d) and the age had no significant relationship with LOC and academic achievement. It can be concluded that the social, cultural patterns of genders have changed. This study might help sociologists and psychologists as well as applied linguists in that they reflect the recent social changes and their effects on the LOC and their consequent implications in teaching languages.

Keywords: academic achievement, biological variables, Iranian online EFL learners, locus of control

Procedia PDF Downloads 408