Search results for: document processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4269

Search results for: document processing

3249 Time and Cost Prediction Models for Language Classification Over a Large Corpus on Spark

Authors: Jairson Barbosa Rodrigues, Paulo Romero Martins Maciel, Germano Crispim Vasconcelos

Abstract:

This paper presents an investigation of the performance impacts regarding the variation of five factors (input data size, node number, cores, memory, and disks) when applying a distributed implementation of Naïve Bayes for text classification of a large Corpus on the Spark big data processing framework. Problem: The algorithm's performance depends on multiple factors, and knowing before-hand the effects of each factor becomes especially critical as hardware is priced by time slice in cloud environments. Objectives: To explain the functional relationship between factors and performance and to develop linear predictor models for time and cost. Methods: the solid statistical principles of Design of Experiments (DoE), particularly the randomized two-level fractional factorial design with replications. This research involved 48 real clusters with different hardware arrangements. The metrics were analyzed using linear models for screening, ranking, and measurement of each factor's impact. Results: Our findings include prediction models and show some non-intuitive results about the small influence of cores and the neutrality of memory and disks on total execution time, and the non-significant impact of data input scale on costs, although notably impacts the execution time.

Keywords: big data, design of experiments, distributed machine learning, natural language processing, spark

Procedia PDF Downloads 96
3248 Voice Liveness Detection Using Kolmogorov Arnold Networks

Authors: Arth J. Shah, Madhu R. Kamble

Abstract:

Voice biometric liveness detection is customized to certify an authentication process of the voice data presented is genuine and not a recording or synthetic voice. With the rise of deepfakes and other equivalently sophisticated spoofing generation techniques, it’s becoming challenging to ensure that the person on the other end is a live speaker or not. Voice Liveness Detection (VLD) system is a group of security measures which detect and prevent voice spoofing attacks. Motivated by the recent development of the Kolmogorov-Arnold Network (KAN) based on the Kolmogorov-Arnold theorem, we proposed KAN for the VLD task. To date, multilayer perceptron (MLP) based classifiers have been used for the classification tasks. We aim to capture not only the compositional structure of the model but also to optimize the values of univariate functions. This study explains the mathematical as well as experimental analysis of KAN for VLD tasks, thereby opening a new perspective for scientists to work on speech and signal processing-based tasks. This study emerges as a combination of traditional signal processing tasks and new deep learning models, which further proved to be a better combination for VLD tasks. The experiments are performed on the POCO and ASVSpoof 2017 V2 database. We used Constant Q-transform, Mel, and short-time Fourier transform (STFT) based front-end features and used CNN, BiLSTM, and KAN as back-end classifiers. The best accuracy is 91.26 % on the POCO database using STFT features with the KAN classifier. In the ASVSpoof 2017 V2 database, the lowest EER we obtained was 26.42 %, using CQT features and KAN as a classifier.

Keywords: Kolmogorov Arnold networks, multilayer perceptron, pop noise, voice liveness detection

Procedia PDF Downloads 12
3247 Documentation of Traditional Knowledge on Wild Medicinal Plants of Egypt

Authors: Nahla S. Abdel-Azim, Khaled A. Shams, Elsayed A. Omer, Mahmoud M. Sakr

Abstract:

Medicinal plants play a significant role in the health care system in Egypt. Knowledge developed over the years by people is mostly unrecorded and orally passes on from one generation to the next. This knowledge is facing the danger of becoming extinct. Therefore there is an urgent need to document the medicinal and aromatic plants associated with traditional knowledge. The Egyptian Encyclopedia of wild medicinal plants (EEWMP) is the first attempt to collect most of the basic elements of the medicinal plant resources of Egypt and their traditional uses. It includes scientific data on about 500 medicinal plants in the form of monographs. Each monograph contains all available information and scientific data on the selected species including the following: names, description, distribution, parts used, habitat, conservational status, active or major chemical constituents, folk medicinal uses and heritage resources, pharmacological and biological activities, authentication, pharmaceutical products, and cultivation. The DNA bar-coding is also included (when available). A brief Arabic summary is given for every monograph. This work revealed the diversity in plant parts used in the treatment of different ailments. In addition, the traditional knowledge gathered can be considered a good starting point for effective in situ and ex-situ conservation of endangered plant species.

Keywords: encyclopedia, medicinal plant, traditional medicine, wild flora

Procedia PDF Downloads 185
3246 Model Canvas and Process for Educational Game Design in Outcome-Based Education

Authors: Ratima Damkham, Natasha Dejdumrong, Priyakorn Pusawiro

Abstract:

This paper explored the solution in game design to help game designers in the educational game designing using digital educational game model canvas (DEGMC) and digital educational game form (DEGF) based on Outcome-based Education program. DEGMC and DEGF can help designers develop an overview of the game while designing and planning their own game. The way to clearly assess players’ ability from learning outcomes and support their game learning design is by using the tools. Designers can balance educational content and entertainment in designing a game by using the strategies of the Business Model Canvas and design the gameplay and players’ ability assessment from learning outcomes they need by referring to the Constructive Alignment. Furthermore, they can use their design plan in this research to write their Game Design Document (GDD). The success of the research was evaluated by four experts’ perspectives in the education and computer field. From the experiments, the canvas and form helped the game designers model their game according to the learning outcomes and analysis of their own game elements. This method can be a path to research an educational game design in the future.

Keywords: constructive alignment, constructivist theory, educational game, outcome-based education

Procedia PDF Downloads 334
3245 Child Mortality in Portuguese Speaking Africa Countries: Levels and Trends, 1975-2021

Authors: Alcino Panguana

Abstract:

All Portuguese-speaking African countries are in Sub-Saharan Africa, a region that has high infant mortality rates, being responsible for 49.6% of deaths in Portuguese-speaking African countries, Angola has levels of infant mortality among children, where 2017, 156 children who died before reaching 1 year of life in 1000 live births. Although there is an increase in studies that document trends and specific causes of infant mortality in each country, historical-comparative studies of infant mortality among these countries remain rare. Understanding the trend of this indicator is important for policymakers and planners in order to improve access to successful child survival operations. Lusophone Africa continues with high infant mortality rates in the order of 64 deaths per thousand births. Assuming heterogeneities that can characterize these countries, raise an analysis investigated indicator at the country level to understand the pattern and historical trend of infant mortality within Lusophone Africa from the year 2021. The result is to understand the levels and evolution of infant mortality in Portuguese-speaking African countries.

Keywords: child mortality, levels, trends, lusophone African countries

Procedia PDF Downloads 115
3244 The Development of Competency with a Training Curriculum via Electronic Media for Condominium Managers

Authors: Chisakan Papapankiad

Abstract:

The purposes of this research were 1) to study the competency of condominium managers, 2) to create the training curriculum via electronic media for condominium managers, and 3) to evaluate the training curriculum for condominium managers. The research methods included document analysis, interview, questionnaire, and a try-out. A total of 20 experts were selected to collect data by using Delphi technique. The designed curriculum was tried out with 30 condominium managers. The important steps of conducting this research included analyzing and synthesizing, creating interview questions, conducting factor analysis and developing the training curriculum, editing by experts, and trying out with sample groups. The findings revealed that there were five core competencies: leadership, human resources management, management, communication, and self-development. The training curriculum was designed and all the learning materials were put into a CD. The evaluation of the training curriculum was performed by five experts and the training curriculum was found to be cohesive and suitable for use in the real world. Moreover, the findings also revealed three important issues: 1) the competencies of the respondents after the experiment were higher than before the experiment and this had a level of significance of 0.01, 2) the competencies remained with the respondents at least 12 weeks and this also had a level of significance of 0.01, and 3) the overall level of satisfaction from the respondents were 'the highest level'.

Keywords: competency training curriculum, condominium managers, electronic media

Procedia PDF Downloads 274
3243 Document Analysis for Modelling iTV Advertising towards Impulse Purchase

Authors: Azizah Che Omar

Abstract:

The study provides a systematic literature review which analyzed the literature for the purpose of looking for concepts, theories, approaches and guidelines in order to propose a conceptual design model of interactive television advertising toward impulse purchase (iTVAdIP). An extensive review of literature was purposely carried out to understand the concepts of interactive television (iTV). Therefore, some elements; iTV guidelines, advertising theories, persuasive approaches, and the impulse purchase elements were analyzed to reach the scope of this work. The extensive review was also a necessity to achieve the objective of this study, which was to determine the concept of iTVAdIP design model. Through systematic review analysis, this study discovered that all the previous models did not emphasize the conceptual design model of interactive television advertising. As a result, the finding showed that the concept of the proposed model should contain the iTV guidelines, advertising theory, persuasive approach and impulse purchase elements. In addition, a summary diagram for the development of the proposed model is depicted to provide clearer understanding towards the concepts of conceptual design model of iTVAdIP.

Keywords: impulse purchase, interactive television advertising, human computer interaction, advertising theories

Procedia PDF Downloads 351
3242 Identifying Reforms Required in Construction Contracts from Resolved Disputed Cases

Authors: K. C. Iyer, Yogita Manan Bindal, Sumit Kumar Bakshi

Abstract:

The construction industry is plagued with disputes and litigation in India with many stalled projects seeking dispute resolution. This has an adverse effect on the performance and overall project delivery and impacts future investments within the industry. While construction industry is the major driver of growth, there has not been major reforms in the government construction contracts. The study is aimed at identifying the proactive means of dispute avoidance, focusing on reforms required within the construction contracts, by studying 49 arbitration awards of construction disputes. The claims presented in the awards are aggregated to study the causes linked to the contract document and are referred against the prospective recommendation and practices as surveyed from literature review of research papers. Within contract administration, record keeping has been a major concern as they are required by the parties to substantiate the claims or the counterclaims and therefore are essential in any dispute redressal process. The study also observes that the right judgment is inhibited when the record keeping is improper and due to lack of coherence between documents, the dispute resolution period is also prolonged. The finding of the research will be relevant to industry practitioners in contract drafting with a view to avoid disputes.

Keywords: construction contract, contract administration, contract management, dispute avoidance

Procedia PDF Downloads 249
3241 Clinical Validation of an Automated Natural Language Processing Algorithm for Finding COVID-19 Symptoms and Complications in Patient Notes

Authors: Karolina Wieczorek, Sophie Wiliams

Abstract:

Introduction: Patient data is often collected in Electronic Health Record Systems (EHR) for purposes such as providing care as well as reporting data. This information can be re-used to validate data models in clinical trials or in epidemiological studies. Manual validation of automated tools is vital to pick up errors in processing and to provide confidence in the output. Mentioning a disease in a discharge letter does not necessarily mean that a patient suffers from this disease. Many of them discuss a diagnostic process, different tests, or discuss whether a patient has a certain disease. The COVID-19 dataset in this study used natural language processing (NLP), an automated algorithm which extracts information related to COVID-19 symptoms, complications, and medications prescribed within the hospital. Free-text patient clinical patient notes are rich sources of information which contain patient data not captured in a structured form, hence the use of named entity recognition (NER) to capture additional information. Methods: Patient data (discharge summary letters) were exported and screened by an algorithm to pick up relevant terms related to COVID-19. Manual validation of automated tools is vital to pick up errors in processing and to provide confidence in the output. A list of 124 Systematized Nomenclature of Medicine (SNOMED) Clinical Terms has been provided in Excel with corresponding IDs. Two independent medical student researchers were provided with a dictionary of SNOMED list of terms to refer to when screening the notes. They worked on two separate datasets called "A” and "B”, respectively. Notes were screened to check if the correct term had been picked-up by the algorithm to ensure that negated terms were not picked up. Results: Its implementation in the hospital began on March 31, 2020, and the first EHR-derived extract was generated for use in an audit study on June 04, 2020. The dataset has contributed to large, priority clinical trials (including International Severe Acute Respiratory and Emerging Infection Consortium (ISARIC) by bulk upload to REDcap research databases) and local research and audit studies. Successful sharing of EHR-extracted datasets requires communicating the provenance and quality, including completeness and accuracy of this data. The results of the validation of the algorithm were the following: precision (0.907), recall (0.416), and F-score test (0.570). Percentage enhancement with NLP extracted terms compared to regular data extraction alone was low (0.3%) for relatively well-documented data such as previous medical history but higher (16.6%, 29.53%, 30.3%, 45.1%) for complications, presenting illness, chronic procedures, acute procedures respectively. Conclusions: This automated NLP algorithm is shown to be useful in facilitating patient data analysis and has the potential to be used in more large-scale clinical trials to assess potential study exclusion criteria for participants in the development of vaccines.

Keywords: automated, algorithm, NLP, COVID-19

Procedia PDF Downloads 83
3240 Combined Synchrotron Radiography and Diffraction for in Situ Study of Reactive Infiltration of Aluminum into Iron Porous Preform

Authors: S. Djaziri, F. Sket, A. Hynowska, S. Milenkovic

Abstract:

The use of Fe-Al based intermetallics as an alternative to Cr/Ni based stainless steels is very promising for industrial applications that use critical raw materials parts under extreme conditions. However, the development of advanced Fe-Al based intermetallics with appropriate mechanical properties presents several challenges that involve appropriate processing and microstructure control. A processing strategy is being developed which aims at producing a net-shape porous Fe-based preform that is infiltrated with molten Al or Al-alloy. In the present work, porous Fe-based preforms produced by two different methods (selective laser melting (SLM) and Kochanek-process (KE)) are studied during infiltration with molten aluminum. In the objective to elucidate the mechanisms underlying the formation of Fe-Al intermetallic phases during infiltration, an in-house furnace has been designed for in situ observation of infiltration at synchrotron facilities combining x-ray radiography (XR) and x-ray diffraction (XRD) techniques. The feasibility of this approach has been demonstrated, and information about the melt flow front propagation has been obtained. In addition, reactive infiltration has been achieved where a bi-phased intermetallic layer has been identified to be formed between the solid Fe and liquid Al. In particular, a tongue-like Fe₂Al₅ phase adhering to the Fe and a needle-like Fe₄Al₁₃ phase adhering to the Al were observed. The growth of the intermetallic compound was found to be dependent on the temperature gradient present along the preform as well as on the reaction time which will be discussed in view of the different obtained results.

Keywords: combined synchrotron radiography and diffraction, Fe-Al intermetallic compounds, in-situ molten Al infiltration, porous solid Fe preforms

Procedia PDF Downloads 210
3239 Reverse Logistics Network Optimization for E-Commerce

Authors: Albert W. K. Tan

Abstract:

This research consolidates a comprehensive array of publications from peer-reviewed journals, case studies, and seminar reports focused on reverse logistics and network design. By synthesizing this secondary knowledge, our objective is to identify and articulate key decision factors crucial to reverse logistics network design for e-commerce. Through this exploration, we aim to present a refined mathematical model that offers valuable insights for companies seeking to optimize their reverse logistics operations. The primary goal of this research endeavor is to develop a comprehensive framework tailored to advising organizations and companies on crafting effective networks for their reverse logistics operations, thereby facilitating the achievement of their organizational goals. This involves a thorough examination of various network configurations, weighing their advantages and disadvantages to ensure alignment with specific business objectives. The key objectives of this research include: (i) Identifying pivotal factors pertinent to network design decisions within the realm of reverse logistics across diverse supply chains. (ii) Formulating a structured framework designed to offer informed recommendations for sound network design decisions applicable to relevant industries and scenarios. (iii) Propose a mathematical model to optimize its reverse logistics network. A conceptual framework for designing a reverse logistics network has been developed through a combination of insights from the literature review and information gathered from company websites. This framework encompasses four key stages in the selection of reverse logistics operations modes: (1) Collection, (2) Sorting and testing, (3) Processing, and (4) Storage. Key factors to consider in reverse logistics network design: I) Centralized vs. decentralized processing: Centralized processing, a long-standing practice in reverse logistics, has recently gained greater attention from manufacturing companies. In this system, all products within the reverse logistics pipeline are brought to a central facility for sorting, processing, and subsequent shipment to their next destinations. Centralization offers the advantage of efficiently managing the reverse logistics flow, potentially leading to increased revenues from returned items. Moreover, it aids in determining the most appropriate reverse channel for handling returns. On the contrary, a decentralized system is more suitable when products are returned directly from consumers to retailers. In this scenario, individual sales outlets serve as gatekeepers for processing returns. Considerations encompass the product lifecycle, product value and cost, return volume, and the geographic distribution of returns. II) In-house vs. third-party logistics providers: The decision between insourcing and outsourcing in reverse logistics network design is pivotal. In insourcing, a company handles the entire reverse logistics process, including material reuse. In contrast, outsourcing involves third-party providers taking on various aspects of reverse logistics. Companies may choose outsourcing due to resource constraints or lack of expertise, with the extent of outsourcing varying based on factors such as personnel skills and cost considerations. Based on the conceptual framework, the authors have constructed a mathematical model that optimizes reverse logistics network design decisions. The model will consider key factors identified in the framework, such as transportation costs, facility capacities, and lead times. The authors have employed mixed LP to find the optimal solutions that minimize costs while meeting organizational objectives.

Keywords: reverse logistics, supply chain management, optimization, e-commerce

Procedia PDF Downloads 18
3238 Context Detection in Spreadsheets Based on Automatically Inferred Table Schema

Authors: Alexander Wachtel, Michael T. Franzen, Walter F. Tichy

Abstract:

Programming requires years of training. With natural language and end user development methods, programming could become available to everyone. It enables end users to program their own devices and extend the functionality of the existing system without any knowledge of programming languages. In this paper, we describe an Interactive Spreadsheet Processing Module (ISPM), a natural language interface to spreadsheets that allows users to address ranges within the spreadsheet based on inferred table schema. Using the ISPM, end users are able to search for values in the schema of the table and to address the data in spreadsheets implicitly. Furthermore, it enables them to select and sort the spreadsheet data by using natural language. ISPM uses a machine learning technique to automatically infer areas within a spreadsheet, including different kinds of headers and data ranges. Since ranges can be identified from natural language queries, the end users can query the data using natural language. During the evaluation 12 undergraduate students were asked to perform operations (sum, sort, group and select) using the system and also Excel without ISPM interface, and the time taken for task completion was compared across the two systems. Only for the selection task did users take less time in Excel (since they directly selected the cells using the mouse) than in ISPM, by using natural language for end user software engineering, to overcome the present bottleneck of professional developers.

Keywords: natural language processing, natural language interfaces, human computer interaction, end user development, dialog systems, data recognition, spreadsheet

Procedia PDF Downloads 288
3237 Studying the Effect of Reducing Thermal Processing over the Bioactive Composition of Non-Centrifugal Cane Sugar: Towards Natural Products with High Therapeutic Value

Authors: Laura Rueda-Gensini, Jader Rodríguez, Juan C. Cruz, Carolina Munoz-Camargo

Abstract:

There is an emerging interest in botanicals and plant extracts for medicinal practices due to their widely reported health benefits. A large variety of phytochemicals found in plants have been correlated with antioxidant, immunomodulatory, and analgesic properties, which makes plant-derived products promising candidates for modulating the progression and treatment of numerous diseases. Non-centrifugal cane sugar (NCS), in particular, has been known for its high antioxidant and nutritional value, but composition-wise variability due to changing environmental and processing conditions have considerably limited its use in the nutraceutical and biomedical fields. This work is therefore aimed at assessing the effect of thermal exposure during NCS production over its bioactive composition and, in turn, its therapeutic value. Accordingly, two modified dehydration methods are proposed that employ: (i) vacuum-aided evaporation, which reduces the necessary temperatures to dehydrate the sample, and (ii) window refractance evaporation, which reduces thermal exposure time. The biochemical composition of NCS produced under these two methods was compared to traditionally-produced NCS by estimating their total polyphenolic and protein content with Folin-Ciocalteu and Bradford assays, as well as identifying the major phenolic compounds in each sample via HPLC-coupled mass spectrometry. Their antioxidant activities were also compared as measured by their scavenging potential of ABTS and DPPH radicals. Results show that the two modified production methods enhance polyphenolic and protein yield in resulting NCS samples when compared to traditional production methods. In particular, reducing employed temperatures with vacuum-aided evaporation demonstrated to be superior at preserving polyphenolic compounds, as evidenced both in the total and individual polyphenol concentrations. However, antioxidant activities were not significantly different between these. Although additional studies should be performed to determine if the observed compositional differences affect other therapeutic activities (e.g., anti-inflammatory, analgesic, and immunoprotective), these results suggest that reducing thermal exposure holds great promise for the production of natural products with enhanced nutritional value.

Keywords: non-centrifugal cane sugar, polyphenolic compounds, thermal processing, antioxidant activity

Procedia PDF Downloads 74
3236 Enhancing Temporal Extrapolation of Wind Speed Using a Hybrid Technique: A Case Study in West Coast of Denmark

Authors: B. Elshafei, X. Mao

Abstract:

The demand for renewable energy is significantly increasing, major investments are being supplied to the wind power generation industry as a leading source of clean energy. The wind energy sector is entirely dependable and driven by the prediction of wind speed, which by the nature of wind is very stochastic and widely random. This s0tudy employs deep multi-fidelity Gaussian process regression, used to predict wind speeds for medium term time horizons. Data of the RUNE experiment in the west coast of Denmark were provided by the Technical University of Denmark, which represent the wind speed across the study area from the period between December 2015 and March 2016. The study aims to investigate the effect of pre-processing the data by denoising the signal using empirical wavelet transform (EWT) and engaging the vector components of wind speed to increase the number of input data layers for data fusion using deep multi-fidelity Gaussian process regression (GPR). The outcomes were compared using root mean square error (RMSE) and the results demonstrated a significant increase in the accuracy of predictions which demonstrated that using vector components of the wind speed as additional predictors exhibits more accurate predictions than strategies that ignore them, reflecting the importance of the inclusion of all sub data and pre-processing signals for wind speed forecasting models.

Keywords: data fusion, Gaussian process regression, signal denoise, temporal extrapolation

Procedia PDF Downloads 125
3235 Proposals for the Thermal Regulation of Buildings in Algeria: A New Energy Label for Social Housing

Authors: Marco Morini, Nicolandrea Calabrese, Dario Chello

Abstract:

Despite the international commitment of Algeria towards the development of energy efficiency and renewable energy in the country, the internal energy demand has been continuously growing during the last decade due to the substantial increase of population and of living conditions, which in turn has led to an unprecedented expansion of the residential building sector. The thermal building regulation is the technical document that establishes the calculation framework for the thermal performance of buildings in Algeria, setting up minimum obligatory targets for the thermal performance of new buildings. An update of this regulation is due in the coming years, and this paper discusses some proposals in this regard, with the aim to improve the energy efficiency of the building sector, particularly with regard to social housing. In particular, it proposes a methodology for drafting an energy performance label of new Algerian residential buildings, moving from the results of the thermal compliance verification and sizing of technical systems as defined in the RTB. Such an energy performance label – whose calculation method is briefly described in the paper – aims to raise citizens' awareness of the benefits of energy efficiency. It can represent the first step in a process of integrating technical installations into the calculation of the energy performance of buildings in Algeria.

Keywords: building, energy certification, energy efficiency, social housing, international cooperation, Mediterranean region

Procedia PDF Downloads 126
3234 Distance Training Packages on Providing for Learner with Special Needs

Authors: Jareeluk Ratanaphan

Abstract:

The purposed of this research were; 1.To survey the teacher’s needs on knowledge about special education management for special needs learner 2.To development of distance training packages on providing for learner with special needs. 3. To study the effects of using the packages on trainee’s achievement. 4. To study the effects of using the packages on trainee’s opinion on the distance training packages. The design of the experiment was research and development. The research sample for survey were 86 teachers, and 22 teachers for study the effects of using the packages on achievement and opinion. The research instrument comprised: 1) training packages on special education management for special needs learner 2) achievement test 3) questionnaire. Mean, percentage, standard deviation, t-test and content analysis were used for data analysis. The findings of the research were as follows: 1. The teacher’s needs on knowledge about teaching for learner with learning disability, mental retardation, autism, physical and health impairment and research in special education. 2. The package composed of special education management for special needs student document and manual of distance training packages. The efficiency of packages was established at 79.50/81.35. 3. The results of using the packages were the posttest average scores of trainee’s achievement were higher than pretest. 4. The trainee’s opinion on the package was at the highest level.

Keywords: distance training, training package, teacher, learner with special needs

Procedia PDF Downloads 320
3233 The Relation between Cognitive Fluency and Utterance Fluency in Second Language Spoken Fluency: Studying Fluency through a Psycholinguistic Lens

Authors: Tannistha Dasgupta

Abstract:

This study explores the aspects of second language (L2) spoken fluency that are related to L2 linguistic knowledge and processing skill. It draws on Levelt’s ‘blueprint’ of the L2 speaker which discusses the cognitive issues underlying the act of speaking. However, L2 speaking assessments have largely neglected the underlying mechanism involved in language production; emphasis is given on the relationship between subjective ratings of L2 speech sample and objectively measured aspects of fluency. Hence, in this study, the relation between L2 linguistic knowledge and processing skill i.e. Cognitive Fluency (CF), and objectively measurable aspects of L2 spoken fluency i.e. Utterance Fluency (UF) is examined. The participants of the study are L2 learners of English, studying at high school level in Hyderabad, India. 50 participants with intermediate level of proficiency in English performed several lexical retrieval tasks and attention-shifting tasks to measure CF, and 8 oral tasks to measure UF. Each aspect of UF (speed, pause, and repair) were measured against the scores of CF to find out those aspects of UF which are reliable indicators of CF. Quantitative analysis of the data shows that among the three aspects of UF; speed is the best predictor of CF, and pause is weakly related to CF. The study suggests that including the speed aspect of UF could make L2 fluency assessment more reliable, valid, and objective. Thus, incorporating the assessment of psycholinguistic mechanisms into L2 spoken fluency testing, could result in fairer evaluation.

Keywords: attention-shifting, cognitive fluency, lexical retrieval, utterance fluency

Procedia PDF Downloads 695
3232 Digitalisation of the Railway Industry: Recent Advances in the Field of Dialogue Systems: Systematic Review

Authors: Andrei Nosov

Abstract:

This paper discusses the development directions of dialogue systems within the digitalisation of the railway industry, where technologies based on conversational AI are already potentially applied or will be applied. Conversational AI is one of the popular natural language processing (NLP) tasks, as it has great prospects for real-world applications today. At the same time, it is a challenging task as it involves many areas of NLP based on complex computations and deep insights from linguistics and psychology. In this review, we focus on dialogue systems and their implementation in the railway domain. We comprehensively review the state-of-the-art research results on dialogue systems and analyse them from three perspectives: type of problem to be solved, type of model, and type of system. In particular, from the perspective of the type of tasks to be solved, we discuss characteristics and applications. This will help to understand how to prioritise tasks. In terms of the type of models, we give an overview that will allow researchers to become familiar with how to apply them in dialogue systems. By analysing the types of dialogue systems, we propose an unconventional approach in contrast to colleagues who traditionally contrast goal-oriented dialogue systems with open-domain systems. Our view focuses on considering retrieval and generative approaches. Furthermore, the work comprehensively presents evaluation methods and datasets for dialogue systems in the railway domain to pave the way for future research. Finally, some possible directions for future research are identified based on recent research results.

Keywords: digitalisation, railway, dialogue systems, conversational AI, natural language processing, natural language understanding, natural language generation

Procedia PDF Downloads 47
3231 The Characteristics of the Graduates Based on Thailand Qualification Framework (TQF) of the Faculty of Industrial Technology, Suan Sunandha Rajabhat University

Authors: Apinya Mungaomklang, Natakamol Lookkham

Abstract:

The purpose of this research is to study the characteristics of the graduates based on Thailand Qualification Framework (TQF) of the Faculty of Industrial Technology, Suan Sunandha Rajabhat University. The population of the research was employers/entrepreneurs/supervisors of students who were doing Professional Experiences course in their respective organizations during semester 1/2012. Data were collected during the month of September 2012 from the total number of 100 people. The tool used in this research was a questionnaire developed by the research team. Data were analyzed using percentage, mean and standard deviation using a computer program. The results showed that most of the surveyed organizations were private companies. The program with most students doing Professional Experiences course was Safety Technology and Occupational Health. The nature of work that most students did was associated with the document. Employers/ entrepreneurs/employers’ opinions on the characteristics of the graduates based on TQF received high scores. Cognitive skills received the highest score, followed by interpersonal relationships and responsibilities, ethics and moral, numerical analysis skills, communication and information technology skills, and knowledge, respectively.

Keywords: graduates characteristics, Thailand Qualification Framework, employers, entrepreneurs

Procedia PDF Downloads 302
3230 Predictive Analysis of Chest X-rays Using NLP and Large Language Models with the Indiana University Dataset and Random Forest Classifier

Authors: Azita Ramezani, Ghazal Mashhadiagha, Bahareh Sanabakhsh

Abstract:

This study researches the combination of Random. Forest classifiers with large language models (LLMs) and natural language processing (NLP) to improve diagnostic accuracy in chest X-ray analysis using the Indiana University dataset. Utilizing advanced NLP techniques, the research preprocesses textual data from radiological reports to extract key features, which are then merged with image-derived data. This improved dataset is analyzed with Random Forest classifiers to predict specific clinical results, focusing on the identification of health issues and the estimation of case urgency. The findings reveal that the combination of NLP, LLMs, and machine learning not only increases diagnostic precision but also reliability, especially in quickly identifying critical conditions. Achieving an accuracy of 99.35%, the model shows significant advancements over conventional diagnostic techniques. The results emphasize the large potential of machine learning in medical imaging, suggesting that these technologies could greatly enhance clinician judgment and patient outcomes by offering quicker and more precise diagnostic approximations.

Keywords: natural language processing (NLP), large language models (LLMs), random forest classifier, chest x-ray analysis, medical imaging, diagnostic accuracy, indiana university dataset, machine learning in healthcare, predictive modeling, clinical decision support systems

Procedia PDF Downloads 19
3229 A Contested Territory in a Sacralized Landscape: The Fight of the Gich Community over Semien Mountains National Park

Authors: Marshet Girmay

Abstract:

Local community involvement is widely considered vital for the sustainability of heritage management. Yet, it is often the case that heritage-related projects lag behind in community involvement. In the Semien Mountains the creation, first, and expansion, later, of the National Park has led to several conflicts with the local communities that for centuries have inhabited the area. Local communities have only been passive actors in the plans to expand the Park set up by UNESCO and by local decision makers. This paper investigates the causes that led the Gich community, one of the communities affected by the Park’s expansion, to refuse the resettlement plan offered by the authorities. Qualitative research methods were employed, including document analysis, community conference and interview of informants. The paper shows that although the local community of Gich was highly attached to the Park’s heritage assets, their level of involvement in the heritage management was very low due to shortcomings in the design and implementation of official policies. Therefore, their attitude towards the Park’s managers has been until the present one of mistrust and opposition. The paper recommends to policy-makers a series of measures more sensitive towards local communities, such as that the development agencies act as true communication facilitators and regional authorities nurture sincere relationships with the locals.

Keywords: Gich, heritage management, local communities, Semen Mountains, sustainability, UNESCO, world heritage site

Procedia PDF Downloads 320
3228 Profiling Risky Code Using Machine Learning

Authors: Zunaira Zaman, David Bohannon

Abstract:

This study explores the application of machine learning (ML) for detecting security vulnerabilities in source code. The research aims to assist organizations with large application portfolios and limited security testing capabilities in prioritizing security activities. ML-based approaches offer benefits such as increased confidence scores, false positives and negatives tuning, and automated feedback. The initial approach using natural language processing techniques to extract features achieved 86% accuracy during the training phase but suffered from overfitting and performed poorly on unseen datasets during testing. To address these issues, the study proposes using the abstract syntax tree (AST) for Java and C++ codebases to capture code semantics and structure and generate path-context representations for each function. The Code2Vec model architecture is used to learn distributed representations of source code snippets for training a machine-learning classifier for vulnerability prediction. The study evaluates the performance of the proposed methodology using two datasets and compares the results with existing approaches. The Devign dataset yielded 60% accuracy in predicting vulnerable code snippets and helped resist overfitting, while the Juliet Test Suite predicted specific vulnerabilities such as OS-Command Injection, Cryptographic, and Cross-Site Scripting vulnerabilities. The Code2Vec model achieved 75% accuracy and a 98% recall rate in predicting OS-Command Injection vulnerabilities. The study concludes that even partial AST representations of source code can be useful for vulnerability prediction. The approach has the potential for automated intelligent analysis of source code, including vulnerability prediction on unseen source code. State-of-the-art models using natural language processing techniques and CNN models with ensemble modelling techniques did not generalize well on unseen data and faced overfitting issues. However, predicting vulnerabilities in source code using machine learning poses challenges such as high dimensionality and complexity of source code, imbalanced datasets, and identifying specific types of vulnerabilities. Future work will address these challenges and expand the scope of the research.

Keywords: code embeddings, neural networks, natural language processing, OS command injection, software security, code properties

Procedia PDF Downloads 90
3227 Extracting the Coupled Dynamics in Thin-Walled Beams from Numerical Data Bases

Authors: Mohammad A. Bani-Khaled

Abstract:

In this work we use the Discrete Proper Orthogonal Decomposition transform to characterize the properties of coupled dynamics in thin-walled beams by exploiting numerical simulations obtained from finite element simulations. The outcomes of the will improve our understanding of the linear and nonlinear coupled behavior of thin-walled beams structures. Thin-walled beams have widespread usage in modern engineering application in both large scale structures (aeronautical structures), as well as in nano-structures (nano-tubes). Therefore, detailed knowledge in regard to the properties of coupled vibrations and buckling in these structures are of great interest in the research community. Due to the geometric complexity in the overall structure and in particular in the cross-sections it is necessary to involve computational mechanics to numerically simulate the dynamics. In using numerical computational techniques, it is not necessary to over simplify a model in order to solve the equations of motions. Computational dynamics methods produce databases of controlled resolution in time and space. These numerical databases contain information on the properties of the coupled dynamics. In order to extract the system dynamic properties and strength of coupling among the various fields of the motion, processing techniques are required. Time- Proper Orthogonal Decomposition transform is a powerful tool for processing databases for the dynamics. It will be used to study the coupled dynamics of thin-walled basic structures. These structures are ideal to form a basis for a systematic study of coupled dynamics in structures of complex geometry.

Keywords: coupled dynamics, geometric complexity, proper orthogonal decomposition (POD), thin walled beams

Procedia PDF Downloads 406
3226 Community and School Partnerships: Raising Student Outcomes through Shared Goals and Values Using Integrated Learning as a Change Model

Authors: Sheila Santharamohana, Susan Bennett

Abstract:

Historically, the attrition rates in secondary schools of Indigenous people or Orang Asli of Malaysia have been a cause for nationwide concern. Efforts to increase student engagement focusing on curriculum re-design and aid have not had the targeted impact. The scope of the research explored a change model incorporating project-based learning and wrap-around support through school-community partnerships to increase Orang Asli engagement, student outcomes and improve cultural connectedness. The evaluation methodology was mixed-method comprising a student questionnaire, interviews, and document analysis. Data and evidence were gathered from school staff, community, the Orang Asli governmental authority (JAKOA) and external agencies. Findings from the year-long research suggests shared values and goals in school-community partnerships foster responsive leadership and is key to safeguarding vulnerable Orang Asli, resulting in improved student outcomes. The research highlighted the barriers to the recognition and distinct needs and unique values of the Orang Asli that impact their educational equity and outcomes.

Keywords: Indigenous Education, Cultural Connectedness, School-Community Partnership, Student Outcomes

Procedia PDF Downloads 118
3225 Automatic Classification of Lung Diseases from CT Images

Authors: Abobaker Mohammed Qasem Farhan, Shangming Yang, Mohammed Al-Nehari

Abstract:

Pneumonia is a kind of lung disease that creates congestion in the chest. Such pneumonic conditions lead to loss of life of the severity of high congestion. Pneumonic lung disease is caused by viral pneumonia, bacterial pneumonia, or Covidi-19 induced pneumonia. The early prediction and classification of such lung diseases help to reduce the mortality rate. We propose the automatic Computer-Aided Diagnosis (CAD) system in this paper using the deep learning approach. The proposed CAD system takes input from raw computerized tomography (CT) scans of the patient's chest and automatically predicts disease classification. We designed the Hybrid Deep Learning Algorithm (HDLA) to improve accuracy and reduce processing requirements. The raw CT scans have pre-processed first to enhance their quality for further analysis. We then applied a hybrid model that consists of automatic feature extraction and classification. We propose the robust 2D Convolutional Neural Network (CNN) model to extract the automatic features from the pre-processed CT image. This CNN model assures feature learning with extremely effective 1D feature extraction for each input CT image. The outcome of the 2D CNN model is then normalized using the Min-Max technique. The second step of the proposed hybrid model is related to training and classification using different classifiers. The simulation outcomes using the publically available dataset prove the robustness and efficiency of the proposed model compared to state-of-art algorithms.

Keywords: CT scan, Covid-19, deep learning, image processing, lung disease classification

Procedia PDF Downloads 130
3224 MXene-Based Self-Sensing of Damage in Fiber Composites

Authors: Latha Nataraj, Todd Henry, Micheal Wallock, Asha Hall, Christine Hatter, Babak Anasori, Yury Gogotsi

Abstract:

Multifunctional composites with enhanced strength and toughness for superior damage tolerance are essential for advanced aerospace and military applications. Detection of structural changes prior to visible damage may be achieved by incorporating fillers with tunable properties such as two-dimensional (2D) nanomaterials with high aspect ratios and more surface-active sites. While 2D graphene with large surface areas, good mechanical properties, and high electrical conductivity seems ideal as a filler, the single-atomic thickness can lead to bending and rolling during processing, requiring post-processing to bond to polymer matrices. Lately, an emerging family of 2D transition metal carbides and nitrides, MXenes, has attracted much attention since their discovery in 2011. Metallic electronic conductivity and good mechanical properties, even with increased polymer content, coupled with hydrophilicity make MXenes a good candidate as a filler material in polymer composites and exceptional as multifunctional damage indicators in composites. Here, we systematically study MXene-based (Ti₃C₂) coated on glass fibers for fiber reinforced polymer composite for self-sensing using microscopy and micromechanical testing. Further testing is in progress through the investigation of local variations in optical, acoustic, and thermal properties within the damage sites in response to strain caused by mechanical loading.

Keywords: damage sensing, fiber composites, MXene, self-sensing

Procedia PDF Downloads 108
3223 A Multidimensional Analysis of English as a Medium of Instruction in Algerian Higher Education: Policy, Practices and Attitudes

Authors: Imene Medfouni

Abstract:

In the context of postcolonial Algeria, language policy, language planning as well as language attitudes have recently stirred up contested debates in higher education system. This linguistic and politically-oriented conflict have constantly created a complex environment for learning. In the light of this observation, English language situates itself at the core of this debate with respects to its international status and potential influences. This presentation is based on ongoing research that aims to gain a better understanding of the introduction of English as a medium of instruction (EMI) in a postcolonial context, marked by multilingualism and language conflict. This research offers interesting insights to critically explore EMI from different perspectives: policy, practices, and attitudes. By means of methodological triangulation, this research integrates a mixed approach, whereby the sources of data triangulation will be elicited from the following methods: classroom observations, document analysis, focus groups, questionnaires and interviews. Preliminary findings suggest that English language might not replace French status in Algerian universities because of the latter strong presence and diffusion within Algerian linguistic landscape.

Keywords: English as a lingua franca, English as a medium of instruction, language policy and planning, multilingualism, postcolonial contexts, World Englishes

Procedia PDF Downloads 231
3222 Development of an Energy Independant DC Building Demonstrator for Insulated Island Site

Authors: Olivia Bory Devisme, Denis Genon-Catalot, Frederic Alicalapa, Pierre-Olivier Lucas De Peslouan, Jean-Pierre Chabriat

Abstract:

In the context of climate change, it is essential that island territories gain energy autonomy. Currently mostly dependent on fossil fuels, the island of Reunion lo- cated in the Indian Ocean nevertheless has a high potential for solar energy. As the market for photovoltaic panels has been growing in recent years, the issues of energy losses linked to the multiple conversions from direct current to alternating current are emerging. In order to quantify these advantages and disadvantages by a comparative study, this document present the measurements carried out on a direct current test bench, particularly for lighting, ventilation, air condi- tioning and office equipment for the tertiary sector. All equipment is supplied with DC power from energy produced by photovoltaic panels. A weather sta- tion, environmental indoor sensors, and drivers are also used to control energy. Self-consumption is encouraged in order to manage different priorities between user consumption and energy storage in a lithium iron phosphate battery. The measurements are compared to a conventional electrical architecture (DC-AC- DC) for energy consumption, equipment overheating, cost, and life cycle analysis.

Keywords: DC microgrids, solar energy, smart buildings, storage

Procedia PDF Downloads 146
3221 Mobile Augmented Reality for Collaboration in Operation

Authors: Chong-Yang Qiao

Abstract:

Mobile augmented reality (MAR) tracking targets from the surroundings and aids operators for interactive data and procedures visualization, potential equipment and system understandably. Operators remotely communicate and coordinate with each other for the continuous tasks, information and data exchange between control room and work-site. In the routine work, distributed control system (DCS) monitoring and work-site manipulation require operators interact in real-time manners. The critical question is the improvement of user experience in cooperative works through applying Augmented Reality in the traditional industrial field. The purpose of this exploratory study is to find the cognitive model for the multiple task performance by MAR. In particular, the focus will be on the comparison between different tasks and environment factors which influence information processing. Three experiments use interface and interaction design, the content of start-up, maintenance and stop embedded in the mobile application. With the evaluation criteria of time demands and human errors, and analysis of the mental process and the behavior action during the multiple tasks, heuristic evaluation was used to find the operators performance with different situation factors, and record the information processing in recognition, interpretation, judgment and reasoning. The research will find the functional properties of MAR and constrain the development of the cognitive model. Conclusions can be drawn that suggest MAR is easy to use and useful for operators in the remote collaborative works.

Keywords: mobile augmented reality, remote collaboration, user experience, cognition model

Procedia PDF Downloads 181
3220 Automatic Segmentation of 3D Tomographic Images Contours at Radiotherapy Planning in Low Cost Solution

Authors: D. F. Carvalho, A. O. Uscamayta, J. C. Guerrero, H. F. Oliveira, P. M. Azevedo-Marques

Abstract:

The creation of vector contours slices (ROIs) on body silhouettes in oncologic patients is an important step during the radiotherapy planning in clinic and hospitals to ensure the accuracy of oncologic treatment. The radiotherapy planning of patients is performed by complex softwares focused on analysis of tumor regions, protection of organs at risk (OARs) and calculation of radiation doses for anomalies (tumors). These softwares are supplied for a few manufacturers and run over sophisticated workstations with vector processing presenting a cost of approximately twenty thousand dollars. The Brazilian project SIPRAD (Radiotherapy Planning System) presents a proposal adapted to the emerging countries reality that generally does not have the monetary conditions to acquire some radiotherapy planning workstations, resulting in waiting queues for new patients treatment. The SIPRAD project is composed by a set of integrated and interoperabilities softwares that are able to execute all stages of radiotherapy planning on simple personal computers (PCs) in replace to the workstations. The goal of this work is to present an image processing technique, computationally feasible, that is able to perform an automatic contour delineation in patient body silhouettes (SIPRAD-Body). The SIPRAD-Body technique is performed in tomography slices under grayscale images, extending their use with a greedy algorithm in three dimensions. SIPRAD-Body creates an irregular polyhedron with the Canny Edge adapted algorithm without the use of preprocessing filters, as contrast and brightness. In addition, comparing the technique SIPRAD-Body with existing current solutions is reached a contours similarity at least 78%. For this comparison is used four criteria: contour area, contour length, difference between the mass centers and Jaccard index technique. SIPRAD-Body was tested in a set of oncologic exams provided by the Clinical Hospital of the University of Sao Paulo (HCRP-USP). The exams were applied in patients with different conditions of ethnology, ages, tumor severities and body regions. Even in case of services that have already workstations, it is possible to have SIPRAD working together PCs because of the interoperability of communication between both systems through the DICOM protocol that provides an increase of workflow. Therefore, the conclusion is that SIPRAD-Body technique is feasible because of its degree of similarity in both new radiotherapy planning services and existing services.

Keywords: radiotherapy, image processing, DICOM RT, Treatment Planning System (TPS)

Procedia PDF Downloads 281