Search results for: machine translation (MT)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3320

Search results for: machine translation (MT)

2360 Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling

Authors: Florin Leon, Silvia Curteanu

Abstract:

Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.

Keywords: batch bulk methyl methacrylate polymerization, adaptive sampling, machine learning, large margin nearest neighbor regression

Procedia PDF Downloads 301
2359 Communicative and Artistic Machines: A Survey of Models and Experiments on Artificial Agents

Authors: Artur Matuck, Guilherme F. Nobre

Abstract:

Machines can be either tool, media, or social agents. Advances in technology have been delivering machines capable of autonomous expression, both through communication and art. This paper deals with models (theoretical approach) and experiments (applied approach) related to artificial agents. On one hand it traces how social sciences' scholars have worked with topics such as text automatization, man-machine writing cooperation, and communication. On the other hand it covers how computer sciences' scholars have built communicative and artistic machines, including the programming of creativity. The aim is to present a brief survey on artificially intelligent communicators and artificially creative writers, and provide the basis to understand the meta-authorship and also to new and further man-machine co-authorship.

Keywords: artificial communication, artificial creativity, artificial writers, meta-authorship, robotic art

Procedia PDF Downloads 289
2358 Analysis of the Significance of Multimedia Channels Using Sparse PCA and Regularized SVD

Authors: Kourosh Modarresi

Abstract:

The abundance of media channels and devices has given users a variety of options to extract, discover, and explore information in the digital world. Since, often, there is a long and complicated path that a typical user may venture before taking any (significant) action (such as purchasing goods and services), it is critical to know how each node (media channel) in the path of user has contributed to the final action. In this work, the significance of each media channel is computed using statistical analysis and machine learning techniques. More specifically, “Regularized Singular Value Decomposition”, and “Sparse Principal Component” has been used to compute the significance of each channel toward the final action. The results of this work are a considerable improvement compared to the present approaches.

Keywords: multimedia attribution, sparse principal component, regularization, singular value decomposition, feature significance, machine learning, linear systems, variable shrinkage

Procedia PDF Downloads 306
2357 Neural Network and Support Vector Machine for Prediction of Foot Disorders Based on Foot Analysis

Authors: Monireh Ahmadi Bani, Adel Khorramrouz, Lalenoor Morvarid, Bagheri Mahtab

Abstract:

Background:- Foot disorders are common in musculoskeletal problems. Plantar pressure distribution measurement is one the most important part of foot disorders diagnosis for quantitative analysis. However, the association of plantar pressure and foot disorders is not clear. With the growth of dataset and machine learning methods, the relationship between foot disorders and plantar pressures can be detected. Significance of the study:- The purpose of this study was to predict the probability of common foot disorders based on peak plantar pressure distribution and center of pressure during walking. Methodologies:- 2323 participants were assessed in a foot therapy clinic between 2015 and 2021. Foot disorders were diagnosed by an experienced physician and then they were asked to walk on a force plate scanner. After the data preprocessing, due to the difference in walking time and foot size, we normalized the samples based on time and foot size. Some of force plate variables were selected as input to a deep neural network (DNN), and the probability of any each foot disorder was measured. In next step, we used support vector machine (SVM) and run dataset for each foot disorder (classification of yes or no). We compared DNN and SVM for foot disorders prediction based on plantar pressure distributions and center of pressure. Findings:- The results demonstrated that the accuracy of deep learning architecture is sufficient for most clinical and research applications in the study population. In addition, the SVM approach has more accuracy for predictions, enabling applications for foot disorders diagnosis. The detection accuracy was 71% by the deep learning algorithm and 78% by the SVM algorithm. Moreover, when we worked with peak plantar pressure distribution, it was more accurate than center of pressure dataset. Conclusion:- Both algorithms- deep learning and SVM will help therapist and patients to improve the data pool and enhance foot disorders prediction with less expense and error after removing some restrictions properly.

Keywords: deep neural network, foot disorder, plantar pressure, support vector machine

Procedia PDF Downloads 343
2356 EEG-Based Screening Tool for School Student’s Brain Disorders Using Machine Learning Algorithms

Authors: Abdelrahman A. Ramzy, Bassel S. Abdallah, Mohamed E. Bahgat, Sarah M. Abdelkader, Sherif H. ElGohary

Abstract:

Attention-Deficit/Hyperactivity Disorder (ADHD), epilepsy, and autism affect millions of children worldwide, many of which are undiagnosed despite the fact that all of these disorders are detectable in early childhood. Late diagnosis can cause severe problems due to the late treatment and to the misconceptions and lack of awareness as a whole towards these disorders. Moreover, electroencephalography (EEG) has played a vital role in the assessment of neural function in children. Therefore, quantitative EEG measurement will be utilized as a tool for use in the evaluation of patients who may have ADHD, epilepsy, and autism. We propose a screening tool that uses EEG signals and machine learning algorithms to detect these disorders at an early age in an automated manner. The proposed classifiers used with epilepsy as a step taken for the work done so far, provided an accuracy of approximately 97% using SVM, Naïve Bayes and Decision tree, while 98% using KNN, which gives hope for the work yet to be conducted.

Keywords: ADHD, autism, epilepsy, EEG, SVM

Procedia PDF Downloads 186
2355 Machine Learning Models for the Prediction of Heating and Cooling Loads of a Residential Building

Authors: Aaditya U. Jhamb

Abstract:

Due to the current energy crisis that many countries are battling, energy-efficient buildings are the subject of extensive research in the modern technological era because of growing worries about energy consumption and its effects on the environment. The paper explores 8 factors that help determine energy efficiency for a building: (relative compactness, surface area, wall area, roof area, overall height, orientation, glazing area, and glazing area distribution), with Tsanas and Xifara providing a dataset. The data set employed 768 different residential building models to anticipate heating and cooling loads with a low mean squared error. By optimizing these characteristics, machine learning algorithms may assess and properly forecast a building's heating and cooling loads, lowering energy usage while increasing the quality of people's lives. As a result, the paper studied the magnitude of the correlation between these input factors and the two output variables using various statistical methods of analysis after determining which input variable was most closely associated with the output loads. The most conclusive model was the Decision Tree Regressor, which had a mean squared error of 0.258, whilst the least definitive model was the Isotonic Regressor, which had a mean squared error of 21.68. This paper also investigated the KNN Regressor and the Linear Regression, which had to mean squared errors of 3.349 and 18.141, respectively. In conclusion, the model, given the 8 input variables, was able to predict the heating and cooling loads of a residential building accurately and precisely.

Keywords: energy efficient buildings, heating load, cooling load, machine learning models

Procedia PDF Downloads 91
2354 Designing and Prototyping Permanent Magnet Generators for Wind Energy

Authors: T. Asefi, J. Faiz, M. A. Khan

Abstract:

This paper introduces dual rotor axial flux machines with surface mounted and spoke type ferrite permanent magnets with concentrated windings; they are introduced as alternatives to a generator with surface mounted Nd-Fe-B magnets. The output power, voltage, speed and air gap clearance for all the generators are identical. The machine designs are optimized for minimum mass using a population-based algorithm, assuming the same efficiency as the Nd-Fe-B machine. A finite element analysis (FEA) is applied to predict the performance, emf, developed torque, cogging torque, no load losses, leakage flux and efficiency of both ferrite generators and that of the Nd-Fe-B generator. To minimize cogging torque, different rotor pole topologies and different pole arc to pole pitch ratios are investigated by means of 3D FEA. It was found that the surface mounted ferrite generator topology is unable to develop the nominal electromagnetic torque, and has higher torque ripple and is heavier than the spoke type machine. Furthermore, it was shown that the spoke type ferrite permanent magnet generator has favorable performance and could be an alternative to rare-earth permanent magnet generators, particularly in wind energy applications. Finally, the analytical and numerical results are verified using experimental results.

Keywords: axial flux, permanent magnet generator, dual rotor, ferrite permanent magnet generator, finite element analysis, wind turbines, cogging torque, population-based algorithms

Procedia PDF Downloads 146
2353 Predicting Data Center Resource Usage Using Quantile Regression to Conserve Energy While Fulfilling the Service Level Agreement

Authors: Ahmed I. Alutabi, Naghmeh Dezhabad, Sudhakar Ganti

Abstract:

Data centers have been growing in size and dema nd continuously in the last two decades. Planning for the deployment of resources has been shallow and always resorted to over-provisioning. Data center operators try to maximize the availability of their services by allocating multiple of the needed resources. One resource that has been wasted, with little thought, has been energy. In recent years, programmable resource allocation has paved the way to allow for more efficient and robust data centers. In this work, we examine the predictability of resource usage in a data center environment. We use a number of models that cover a wide spectrum of machine learning categories. Then we establish a framework to guarantee the client service level agreement (SLA). Our results show that using prediction can cut energy loss by up to 55%.

Keywords: machine learning, artificial intelligence, prediction, data center, resource allocation, green computing

Procedia PDF Downloads 104
2352 Developing an Out-of-Distribution Generalization Model Selection Framework through Impurity and Randomness Measurements and a Bias Index

Authors: Todd Zhou, Mikhail Yurochkin

Abstract:

Out-of-distribution (OOD) detection is receiving increasing amounts of attention in the machine learning research community, boosted by recent technologies, such as autonomous driving and image processing. This newly-burgeoning field has called for the need for more effective and efficient methods for out-of-distribution generalization methods. Without accessing the label information, deploying machine learning models to out-of-distribution domains becomes extremely challenging since it is impossible to evaluate model performance on unseen domains. To tackle this out-of-distribution detection difficulty, we designed a model selection pipeline algorithm and developed a model selection framework with different impurity and randomness measurements to evaluate and choose the best-performing models for out-of-distribution data. By exploring different randomness scores based on predicted probabilities, we adopted the out-of-distribution entropy and developed a custom-designed score, ”CombinedScore,” as the evaluation criterion. This proposed score was created by adding labeled source information into the judging space of the uncertainty entropy score using harmonic mean. Furthermore, the prediction bias was explored through the equality of opportunity violation measurement. We also improved machine learning model performance through model calibration. The effectiveness of the framework with the proposed evaluation criteria was validated on the Folktables American Community Survey (ACS) datasets.

Keywords: model selection, domain generalization, model fairness, randomness measurements, bias index

Procedia PDF Downloads 120
2351 Work in the Industry of the Future-Investigations of Human-Machine Interactions

Authors: S. Schröder, P. Ennen, T. Langer, S. Müller, M. Shehadeh, M. Haberstroh, F. Hees

Abstract:

Since a bit over a year ago, Festo AG and Co. KG, Festo Didactic SE, robomotion GmbH, the researchers of the Cybernetics-Lab IMA/ZLW and IfU, as well as the Human-Computer Interaction Center at the RWTH Aachen University, have been working together in the focal point of assembly competences to realize different scenarios in the field of human-machine interaction (HMI). In the framework of project ARIZ, questions concerning the future of production within the fourth industrial revolution are dealt with. There are many perspectives of human-robot collaboration that consist Industry 4.0 on an individual, organization and enterprise level, and these will be addressed in ARIZ. The aim of the ARIZ projects is to link AI-Approaches to assembly problems and to implement them as prototypes in demonstrators. To do so, island and flow based production scenarios will be simulated and realized as prototypes. These prototypes will serve as applications of flexible robotics as well as AI-based planning and control of production process. Using the demonstrators, human interaction strategies will be examined with an information system on one hand, and a robotic system on the other. During the tests, prototypes of workspaces that illustrate prospective production work forms will be represented. The human being will remain a central element in future productions and will increasingly be in charge of managerial tasks. Questions thus arise within the overall perspective, primarily concerning the role of humans within these technological revolutions, as well as their ability to act and design respectively to the acceptance of such systems. Roles, such as the 'Trainer' of intelligent systems may become a possibility in such assembly scenarios.

Keywords: human-machine interaction, information technology, island based production, assembly competences

Procedia PDF Downloads 199
2350 Effect of Personality Traits on Classification of Political Orientation

Authors: Vesile Evrim, Aliyu Awwal

Abstract:

Today as in the other domains, there are an enormous number of political transcripts available in the Web which is waiting to be mined and used for various purposes such as statistics and recommendations. Therefore, automatically determining the political orientation on these transcripts becomes crucial. The methodologies used by machine learning algorithms to do the automatic classification are based on different features such as Linguistic. Considering the ideology differences between Liberals and Conservatives, in this paper, the effect of Personality Traits on political orientation classification is studied. This is done by considering the correlation between LIWC features and the BIG Five Personality Traits. Several experiments are conducted on Convote U.S. Congressional-Speech dataset with seven benchmark classification algorithms. The different methodologies are applied on selecting different feature sets that constituted by 8 to 64 varying number of features. While Neuroticism is obtained to be the most differentiating personality trait on classification of political polarity, when its top 10 representative features are combined with several classification algorithms, it outperformed the results presented in previous research.

Keywords: politics, personality traits, LIWC, machine learning

Procedia PDF Downloads 491
2349 Optimal Design of Multi-Machine Power System Stabilizers Using Interactive Honey Bee Mating Optimization

Authors: Hossein Ghadimi, Alireza Alizadeh, Oveis Abedinia, Noradin Ghadimi

Abstract:

This paper presents an enhanced Honey Bee Mating Optimization (HBMO) to solve the optimal design of multi machine power system stabilizer (PSSs) parameters, which is called the Interactive Honey Bee Mating Optimization (IHBMO). Power System Stabilizers (PSSs) are now routinely used in the industry to damp out power system oscillations. The design problem of the proposed controller is formulated as an optimization problem and IHBMO algorithm is employed to search for optimal controller parameters. The proposed method is applied to multi-machine power system (MPS). The method suggested in this paper can be used for designing robust power system stabilizers for guaranteeing the required closed loop performance over a prespecified range of operating and system conditions. The simplicity in design and implementation of the proposed stabilizers makes them better suited for practical applications in real plants. The non-linear simulation results are presented under wide range of operating conditions in comparison with the PSO and CPSS base tuned stabilizer one through FD and ITAE performance indices. The results evaluation shows that the proposed control strategy achieves good robust performance for a wide range of system parameters and load changes in the presence of system nonlinearities and is superior to the other controllers.

Keywords: power system stabilizer, IHBMO, multimachine, nonlinearities

Procedia PDF Downloads 502
2348 Automatic Speech Recognition Systems Performance Evaluation Using Word Error Rate Method

Authors: João Rato, Nuno Costa

Abstract:

The human verbal communication is a two-way process which requires a mutual understanding that will result in some considerations. This kind of communication, also called dialogue, besides the supposed human agents it can also be performed between human agents and machines. The interaction between Men and Machines, by means of a natural language, has an important role concerning the improvement of the communication between each other. Aiming at knowing the performance of some speech recognition systems, this document shows the results of the accomplished tests according to the Word Error Rate evaluation method. Besides that, it is also given a set of information linked to the systems of Man-Machine communication. After this work has been made, conclusions were drawn regarding the Speech Recognition Systems, among which it can be mentioned their poor performance concerning the voice interpretation in noisy environments.

Keywords: automatic speech recognition, man-machine conversation, speech recognition, spoken dialogue systems, word error rate

Procedia PDF Downloads 319
2347 Exploring the Determinants of Personal Finance Difficulties by Machine Learning: Focus on Socio-Economic and Behavioural Changes Brought by COVID-19

Authors: Brian Tung, Yam Wing Siu, Tsun Se Cheong

Abstract:

Purpose: This research aims to explore how personal and environmental factors, especially the socio-economic changes and behavioral changes fostered by the COVID-19 outbreak pandemic, affect the financial vulnerability of a specific segment of people in financial distress. Innovative research methodology of machine learning will be applied to data collected from over 300 local individuals in Hong Kong seeking counseling or similar services in recent years. Results: First, machine learning has found that too much exposure to digital services and information on digitized services may lead to adverse effects on respondents’ financial vulnerability. Second, the improvement in financial literacy level provides benefits to the financially vulnerable group, especially those respondents who have started with a lower level. Third, serious addiction to digital technology can lead to worsened debt servicing ability. Machine learning also has found a strong correlation between debt servicing situations and income-seeking behavior as well as spending behavior. In addition, if the vulnerable groups are able to make appropriate investments, they can reduce the probability of incurring financial distress. Finally, being too active in borrowing and repayment can result in a higher likelihood of over-indebtedness. Conclusion: Findings can be employed in formulating a better counseling strategy for professionals. Debt counseling services can be more preventive in nature. For example, according to the findings, with a low level of financial literacy, the respondents are prone to overspending and unable to react properly to the e-marketing promotion messages pop-up from digital services or even falling into financial/investment scams. In addition, people with low levels of financial knowledge will benefit from financial education. Therefore, financial education programs could include tech-savvy matters as special features.

Keywords: personal finance, digitization of the economy, COVID-19 pandemic, addiction to digital technology, financial vulnerability

Procedia PDF Downloads 54
2346 A Study on the Application of Accelerated Life Test to Electric Motor for Machine Tools

Authors: Youn-Hwan Kim, Jae-Won Moon, Hae-Joong Kim

Abstract:

This paper introduces the results of the study on the development of accelerated life test methods for the motor used in machine tools. In recent years, as well as efficiency for motors, there is a growing need for research on life expectancy of motors. It is considered impossible to calculate the acceleration coefficient by increasing the rotational load or temperature load as the acceleration stress in the motor system because the temperature of the copper exceeds the wire thermal class rating. This paper describes the equipment development procedure for the highly accelerated life test (HALT) of the 12kW three-phase squirrel-cage induction motors (SCIMs). After the test, the lifetime analysis was carried out, and it is compared with the life expectancy by finite element method (FEM) and bearing theory.

Keywords: acceleration coefficient, bearing, HALT, life expectancy, motor

Procedia PDF Downloads 276
2345 Application of Model Tree in the Prediction of TBM Rate of Penetration with Synthetic Minority Oversampling Technique

Authors: Ehsan Mehryaar

Abstract:

The rate of penetration is (RoP) one of the vital factors in the cost and time of tunnel boring projects; therefore, predicting it can lead to a substantial increase in the efficiency of the project. RoP is heavily dependent geological properties of the project site and TBM properties. In this study, 151-point data from Queen’s water tunnel is collected, which includes unconfined compression strength, peak slope index, angle with weak planes, and distance between planes of weaknesses. Since the size of the data is small, it was observed that it is imbalanced. To solve that problem synthetic minority oversampling technique is utilized. The model based on the model tree is proposed, where each leaf consists of a support vector machine model. Proposed model performance is then compared to existing empirical equations in the literature.

Keywords: Model tree, SMOTE, rate of penetration, TBM(tunnel boring machine), SVM

Procedia PDF Downloads 173
2344 Acquisition of French (L3) Direct Object by Persian (L1) Speakers of English (L2) as EFL Learners

Authors: Ali Akbar Jabbari

Abstract:

The present study assessed the acquisition of L3 French direct objects by Persian speakers who had already learned English as their L2. The ultimate goal of this paper is to extend the current knowledge about the CLI phenomenon in the realm of third language acquisition by examining the role of Persian and English as background languages and learners’ English level of proficiency in their performance on French direct object. To fulfill this, the assumptions of three L3 hypotheses, namely L1 Transfer, L2 Status Factor, and Cumulative Enhancement Model, were examined. The research sample was comprised of 40 undergraduate students in the fields of English language and literature and translation studies at Birjand University in Iran. According to the English proficiency level of learners revealed by the Quick Oxford English Placement test, the participants were grouped as upper intermediate and lower intermediate. A grammaticality judgment and a translation test were administered to gather the required data on learners' comprehension and production of the desired structure in French. It was demonstrated that the rate of positive transfer from previously learned languages was more potent than the rate of negative transfer. A Comparison of groups' performances revealed a significant difference between upper and lower intermediate groups in positing French direct objects correctly. However, the upper intermediate group did not significantly differ from the lower intermediate group in negative transfer. It can be said that by increasing the L2 proficiency of the learners, they could use their previous linguistic knowledge more efficiently. Although further examinations are needed, the current study contributed to a better characterization of cross-linguistic influence in third language acquisition. The findings help French teachers and learners to positively exploit the prior knowledge of Persian and English and apply it in in the multilingual context of French direct object's teaching and learning process.

Keywords: Cross-Linguistic Influence, Persian, French & English Direct Object, Third Language Acquisition, Language Transfer

Procedia PDF Downloads 66
2343 Big Data in Telecom Industry: Effective Predictive Techniques on Call Detail Records

Authors: Sara ElElimy, Samir Moustafa

Abstract:

Mobile network operators start to face many challenges in the digital era, especially with high demands from customers. Since mobile network operators are considered a source of big data, traditional techniques are not effective with new era of big data, Internet of things (IoT) and 5G; as a result, handling effectively different big datasets becomes a vital task for operators with the continuous growth of data and moving from long term evolution (LTE) to 5G. So, there is an urgent need for effective Big data analytics to predict future demands, traffic, and network performance to full fill the requirements of the fifth generation of mobile network technology. In this paper, we introduce data science techniques using machine learning and deep learning algorithms: the autoregressive integrated moving average (ARIMA), Bayesian-based curve fitting, and recurrent neural network (RNN) are employed for a data-driven application to mobile network operators. The main framework included in models are identification parameters of each model, estimation, prediction, and final data-driven application of this prediction from business and network performance applications. These models are applied to Telecom Italia Big Data challenge call detail records (CDRs) datasets. The performance of these models is found out using a specific well-known evaluation criteria shows that ARIMA (machine learning-based model) is more accurate as a predictive model in such a dataset than the RNN (deep learning model).

Keywords: big data analytics, machine learning, CDRs, 5G

Procedia PDF Downloads 137
2342 Feature Analysis of Predictive Maintenance Models

Authors: Zhaoan Wang

Abstract:

Research in predictive maintenance modeling has improved in the recent years to predict failures and needed maintenance with high accuracy, saving cost and improving manufacturing efficiency. However, classic prediction models provide little valuable insight towards the most important features contributing to the failure. By analyzing and quantifying feature importance in predictive maintenance models, cost saving can be optimized based on business goals. First, multiple classifiers are evaluated with cross-validation to predict the multi-class of failures. Second, predictive performance with features provided by different feature selection algorithms are further analyzed. Third, features selected by different algorithms are ranked and combined based on their predictive power. Finally, linear explainer SHAP (SHapley Additive exPlanations) is applied to interpret classifier behavior and provide further insight towards the specific roles of features in both local predictions and global model behavior. The results of the experiments suggest that certain features play dominant roles in predictive models while others have significantly less impact on the overall performance. Moreover, for multi-class prediction of machine failures, the most important features vary with type of machine failures. The results may lead to improved productivity and cost saving by prioritizing sensor deployment, data collection, and data processing of more important features over less importance features.

Keywords: automated supply chain, intelligent manufacturing, predictive maintenance machine learning, feature engineering, model interpretation

Procedia PDF Downloads 126
2341 Image Processing-Based Maize Disease Detection Using Mobile Application

Authors: Nathenal Thomas

Abstract:

In the food chain and in many other agricultural products, corn, also known as maize, which goes by the scientific name Zea mays subsp, is a widely produced agricultural product. Corn has the highest adaptability. It comes in many different types, is employed in many different industrial processes, and is more adaptable to different agro-climatic situations. In Ethiopia, maize is among the most widely grown crop. Small-scale corn farming may be a household's only source of food in developing nations like Ethiopia. The aforementioned data demonstrates that the country's requirement for this crop is excessively high, and conversely, the crop's productivity is very low for a variety of reasons. The most damaging disease that greatly contributes to this imbalance between the crop's supply and demand is the corn disease. The failure to diagnose diseases in maize plant until they are too late is one of the most important factors influencing crop output in Ethiopia. This study will aid in the early detection of such diseases and support farmers during the cultivation process, directly affecting the amount of maize produced. The diseases in maize plants, such as northern leaf blight and cercospora leaf spot, have distinct symptoms that are visible. This study aims to detect the most frequent and degrading maize diseases using the most efficiently used subset of machine learning technology, deep learning so, called Image Processing. Deep learning uses networks that can be trained from unlabeled data without supervision (unsupervised). It is a feature that simulates the exercises the human brain goes through when digesting data. Its applications include speech recognition, language translation, object classification, and decision-making. Convolutional Neural Network (CNN) for Image Processing, also known as convent, is a deep learning class that is widely used for image classification, image detection, face recognition, and other problems. it will also use this algorithm as the state-of-the-art for my research to detect maize diseases by photographing maize leaves using a mobile phone.

Keywords: CNN, zea mays subsp, leaf blight, cercospora leaf spot

Procedia PDF Downloads 72
2340 Statistical Wavelet Features, PCA, and SVM-Based Approach for EEG Signals Classification

Authors: R. K. Chaurasiya, N. D. Londhe, S. Ghosh

Abstract:

The study of the electrical signals produced by neural activities of human brain is called Electroencephalography. In this paper, we propose an automatic and efficient EEG signal classification approach. The proposed approach is used to classify the EEG signal into two classes: epileptic seizure or not. In the proposed approach, we start with extracting the features by applying Discrete Wavelet Transform (DWT) in order to decompose the EEG signals into sub-bands. These features, extracted from details and approximation coefficients of DWT sub-bands, are used as input to Principal Component Analysis (PCA). The classification is based on reducing the feature dimension using PCA and deriving the support-vectors using Support Vector Machine (SVM). The experimental are performed on real and standard dataset. A very high level of classification accuracy is obtained in the result of classification.

Keywords: discrete wavelet transform, electroencephalogram, pattern recognition, principal component analysis, support vector machine

Procedia PDF Downloads 631
2339 Predictive Analysis of Chest X-rays Using NLP and Large Language Models with the Indiana University Dataset and Random Forest Classifier

Authors: Azita Ramezani, Ghazal Mashhadiagha, Bahareh Sanabakhsh

Abstract:

This study researches the combination of Random. Forest classifiers with large language models (LLMs) and natural language processing (NLP) to improve diagnostic accuracy in chest X-ray analysis using the Indiana University dataset. Utilizing advanced NLP techniques, the research preprocesses textual data from radiological reports to extract key features, which are then merged with image-derived data. This improved dataset is analyzed with Random Forest classifiers to predict specific clinical results, focusing on the identification of health issues and the estimation of case urgency. The findings reveal that the combination of NLP, LLMs, and machine learning not only increases diagnostic precision but also reliability, especially in quickly identifying critical conditions. Achieving an accuracy of 99.35%, the model shows significant advancements over conventional diagnostic techniques. The results emphasize the large potential of machine learning in medical imaging, suggesting that these technologies could greatly enhance clinician judgment and patient outcomes by offering quicker and more precise diagnostic approximations.

Keywords: natural language processing (NLP), large language models (LLMs), random forest classifier, chest x-ray analysis, medical imaging, diagnostic accuracy, indiana university dataset, machine learning in healthcare, predictive modeling, clinical decision support systems

Procedia PDF Downloads 37
2338 Contextual SenSe Model: Word Sense Disambiguation using Sense and Sense Value of Context Surrounding the Target

Authors: Vishal Raj, Noorhan Abbas

Abstract:

Ambiguity in NLP (Natural language processing) refers to the ability of a word, phrase, sentence, or text to have multiple meanings. This results in various kinds of ambiguities such as lexical, syntactic, semantic, anaphoric and referential am-biguities. This study is focused mainly on solving the issue of Lexical ambiguity. Word Sense Disambiguation (WSD) is an NLP technique that aims to resolve lexical ambiguity by determining the correct meaning of a word within a given context. Most WSD solutions rely on words for training and testing, but we have used lemma and Part of Speech (POS) tokens of words for training and testing. Lemma adds generality and POS adds properties of word into token. We have designed a novel method to create an affinity matrix to calculate the affinity be-tween any pair of lemma_POS (a token where lemma and POS of word are joined by underscore) of given training set. Additionally, we have devised an al-gorithm to create the sense clusters of tokens using affinity matrix under hierar-chy of POS of lemma. Furthermore, three different mechanisms to predict the sense of target word using the affinity/similarity value are devised. Each contex-tual token contributes to the sense of target word with some value and whichever sense gets higher value becomes the sense of target word. So, contextual tokens play a key role in creating sense clusters and predicting the sense of target word, hence, the model is named Contextual SenSe Model (CSM). CSM exhibits a noteworthy simplicity and explication lucidity in contrast to contemporary deep learning models characterized by intricacy, time-intensive processes, and chal-lenging explication. CSM is trained on SemCor training data and evaluated on SemEval test dataset. The results indicate that despite the naivety of the method, it achieves promising results when compared to the Most Frequent Sense (MFS) model.

Keywords: word sense disambiguation (wsd), contextual sense model (csm), most frequent sense (mfs), part of speech (pos), natural language processing (nlp), oov (out of vocabulary), lemma_pos (a token where lemma and pos of word are joined by underscore), information retrieval (ir), machine translation (mt)

Procedia PDF Downloads 98
2337 Translation and Validation of the Thai Version of the Japanese Sleep Questionnaire for Preschoolers

Authors: Natcha Lueangapapong, Chariya Chuthapisith, Lunliya Thampratankul

Abstract:

Background: There is a need to find an appropriate tool to help healthcare providers determine sleep problems in children for early diagnosis and management. The Japanese Sleep Questionnaire for Preschoolers (JSQ-P) is a parent-reported sleep questionnaire that has good psychometric properties and can be used in the context of Asian culture, which is likely suitable for Thai children. Objectives: This study aimed to translate and validate the Japanese Sleep Questionnaire for Preschoolers (JSQ-P) into a Thai version and to evaluate factors associated with sleep disorders in preschoolers. Methods: After approval by the original developer, the cross-cultural adaptation process of JSQ-P was performed, including forward translation, reconciliation, backward translation, and final approval of the Thai version of JSQ-P (TH-JSQ-P) by the original creator. This study was conducted between March 2021 and February 2022. The TH-JSQ-P was completed by 2,613 guardians whose children were aged 2-6 years twice in 10-14 days to assess its reliability and validity. Content validity was measured by an index of item-objective congruence (IOC) and a content validity index (CVI). Face validity, content validity, structural validity, construct validity (discriminant validity), criterion validity and predictive validity were assessed. The sensitivity and specificity of the TH-JSQ-P were also measured by using a total JSQ-P score cutoff point 84, recommended by the original JSQ-P and each subscale score among the clinical samples of obstructive sleep apnea syndrome. Results: Internal consistency reliability, evaluated by Cronbach’s α coefficient, showed acceptable reliability in all subscales of JSQ-P. It also had good test-retest reliability, as the intraclass correlation coefficient (ICC) for all items ranged between 0.42-0.84. The content validity was acceptable. For structural validity, our results indicated that the final factor solution for the Th-JSQ-P was comparable to the original JSQ-P. For construct validity, age group was one of the clinical parameters associated with some sleep problems. In detail, parasomnias, insomnia, daytime excessive sleepiness and sleep habits significantly decreased when the children got older; on the other hand, insufficient sleep was significantly increased with age. For criterion validity, all subscales showed a correlation with the Epworth Sleepiness Scale (r = -0.049-0.349). In predictive validity, the Epworth Sleepiness Scale was significantly a strong factor that correlated to sleep problems in all subscales of JSQ-P except in the subscale of sleep habit. The sensitivity and specificity of the total JSQ-P score were 0.72 and 0.66, respectively. Conclusion: The Thai version of JSQ-P has good internal consistency reliability and test-retest reliability. It passed 6 validity tests, and this can be used to evaluate sleep problems in preschool children in Thailand. Furthermore, it has satisfactory general psychometric properties and good reliability and validity. The data collected in examining the sensitivity of the Thai version revealed that the JSQ-P could detect differences in sleep problems among children with obstructive sleep apnea syndrome. This confirmed that the measure is sensitive and can be used to discriminate sleep problems among different children.

Keywords: preschooler, questionnaire, validation, Thai version

Procedia PDF Downloads 100
2336 CRISPR-DT: Designing gRNAs for the CRISPR-Cpf1 System with Improved Target Efficiency and Specificity

Authors: Houxiang Zhu, Chun Liang

Abstract:

The CRISPR-Cpf1 system has been successfully applied in genome editing. However, target efficiency of the CRISPR-Cpf1 system varies among different gRNA sequences. The published CRISPR-Cpf1 gRNA data was reanalyzed. Many sequences and structural features of gRNAs (e.g., the position-specific nucleotide composition, position-nonspecific nucleotide composition, GC content, minimum free energy, and melting temperature) correlated with target efficiency were found. Using machine learning technology, a support vector machine (SVM) model was created to predict target efficiency for any given gRNAs. The first web service application, CRISPR-DT (CRISPR DNA Targeting), has been developed to help users design optimal gRNAs for the CRISPR-Cpf1 system by considering both target efficiency and specificity. CRISPR-DT will empower researchers in genome editing.

Keywords: CRISPR-Cpf1, genome editing, target efficiency, target specificity

Procedia PDF Downloads 257
2335 Experimental Investigation on Flexural Properties of Bamboo Fibres Polypropylene Composites

Authors: Tigist Girma Kidane, Yalew Dessalegn Asfaw

Abstract:

Abstract: The current investigation aims to measure the longitudinal and transversal three-point bending tests of bamboo fibres polypropylene composites (BFPPCs) for the application of the automobile industry. Research has not been done on the properties of Ethiopian bamboo fibres for the utilization of composite development. The samples of bamboo plants have been harvested in 3–groups of age, 2–harvesting seasons, and 3–regions of bamboo species. Roll milling machine used for the extraction of bamboo fibres which has been developed by the authors. Chemical constituents measured using gravimetric methods. Unidirectional bamboo fibres prepreg has been produced using PP and hot press machine, then BFPPCs were produced using 6 layers of prepregs at automatic hot press machine. Age, harvesting month, and bamboo species have a statistically significant effect on the longitudinal and transverse flexural strength (FS), modulus of elasticity (MOE), and failure strain at α = 0.05 as evaluated by one-way ANOVA. 2–yrs old of BFPPCs have the highest FS and MOE, whereas November has the highest value of flexural properties. The highest to the lowest FS and MOE of BFPPCs has measured in Injibara, Mekaneselam, and Kombolcha, respectively. The transverse 3-point bending test has a lower FS and MOE compared to the longitudinal direction. The chemical constituents of Injibara, Mekaneselam, and Kombolcha have the highest to the lowest, respectively. 2-years old of bamboo fibres has the highest chemical constituent. The chemical constituents improved the flexural properties. Bamboo fibres in Ethiopia can be relevant for composite development, which has been applied in the area of requiring higher flexural properties.

Keywords: age, bamboo species, flexural properties, harvesting season, polypropylene

Procedia PDF Downloads 49
2334 Design and Simulation of a Double-Stator Linear Induction Machine with Short Squirrel-Cage Mover

Authors: David Rafetseder, Walter Bauer, Florian Poltschak, Wolfgang Amrhein

Abstract:

A flat double-stator linear induction machine (DSLIM) with a short squirrel-cage mover is designed for high thrust force at moderate speed < 5m/s. The performance and motor parameters are determined on the basis of a 2D time-transient simulation with the finite element (FE) software Maxwell 2015. Design guidelines and transformation rules for space vector theory of the LIM are presented. Resulting thrust calculated by flux and current vectors is compared with the FE results showing good coherence and reduced noise. The parameters of the equivalent circuit model are obtained.

Keywords: equivalent circuit model, finite element model, linear induction motor, space vector theory

Procedia PDF Downloads 563
2333 Comparative Study Using WEKA for Red Blood Cells Classification

Authors: Jameela Ali, Hamid A. Jalab, Loay E. George, Abdul Rahim Ahmad, Azizah Suliman, Karim Al-Jashamy

Abstract:

Red blood cells (RBC) are the most common types of blood cells and are the most intensively studied in cell biology. The lack of RBCs is a condition in which the amount of hemoglobin level is lower than normal and is referred to as “anemia”. Abnormalities in RBCs will affect the exchange of oxygen. This paper presents a comparative study for various techniques for classifying the RBCs as normal, or abnormal (anemic) using WEKA. WEKA is an open source consists of different machine learning algorithms for data mining applications. The algorithm tested are Radial Basis Function neural network, Support vector machine, and K-Nearest Neighbors algorithm. Two sets of combined features were utilized for classification of blood cells images. The first set, exclusively consist of geometrical features, was used to identify whether the tested blood cell has a spherical shape or non-spherical cells. While the second set, consist mainly of textural features was used to recognize the types of the spherical cells. We have provided an evaluation based on applying these classification methods to our RBCs image dataset which were obtained from Serdang Hospital-alaysia, and measuring the accuracy of test results. The best achieved classification rates are 97%, 98%, and 79% for Support vector machines, Radial Basis Function neural network, and K-Nearest Neighbors algorithm respectively.

Keywords: K-nearest neighbors algorithm, radial basis function neural network, red blood cells, support vector machine

Procedia PDF Downloads 406
2332 Enabling Oral Communication and Accelerating Recovery: The Creation of a Novel Low-Cost Electroencephalography-Based Brain-Computer Interface for the Differently Abled

Authors: Rishabh Ambavanekar

Abstract:

Expressive Aphasia (EA) is an oral disability, common among stroke victims, in which the Broca’s area of the brain is damaged, interfering with verbal communication abilities. EA currently has no technological solutions and its only current viable solutions are inefficient or only available to the affluent. This prompts the need for an affordable, innovative solution to facilitate recovery and assist in speech generation. This project proposes a novel concept: using a wearable low-cost electroencephalography (EEG) device-based brain-computer interface (BCI) to translate a user’s inner dialogue into words. A low-cost EEG device was developed and found to be 10 to 100 times less expensive than any current EEG device on the market. As part of the BCI, a machine learning (ML) model was developed and trained using the EEG data. Two stages of testing were conducted to analyze the effectiveness of the device: a proof-of-concept and a final solution test. The proof-of-concept test demonstrated an average accuracy of above 90% and the final solution test demonstrated an average accuracy of above 75%. These two successful tests were used as a basis to demonstrate the viability of BCI research in developing lower-cost verbal communication devices. Additionally, the device proved to not only enable users to verbally communicate but has the potential to also assist in accelerated recovery from the disorder.

Keywords: neurotechnology, brain-computer interface, neuroscience, human-machine interface, BCI, HMI, aphasia, verbal disability, stroke, low-cost, machine learning, ML, image recognition, EEG, signal analysis

Procedia PDF Downloads 116
2331 Classifier for Liver Ultrasound Images

Authors: Soumya Sajjan

Abstract:

Liver cancer is the most common cancer disease worldwide in men and women, and is one of the few cancers still on the rise. Liver disease is the 4th leading cause of death. According to new NHS (National Health Service) figures, deaths from liver diseases have reached record levels, rising by 25% in less than a decade; heavy drinking, obesity, and hepatitis are believed to be behind the rise. In this study, we focus on Development of Diagnostic Classifier for Ultrasound liver lesion. Ultrasound (US) Sonography is an easy-to-use and widely popular imaging modality because of its ability to visualize many human soft tissues/organs without any harmful effect. This paper will provide an overview of underlying concepts, along with algorithms for processing of liver ultrasound images Naturaly, Ultrasound liver lesion images are having more spackle noise. Developing classifier for ultrasound liver lesion image is a challenging task. We approach fully automatic machine learning system for developing this classifier. First, we segment the liver image by calculating the textural features from co-occurrence matrix and run length method. For classification, Support Vector Machine is used based on the risk bounds of statistical learning theory. The textural features for different features methods are given as input to the SVM individually. Performance analysis train and test datasets carried out separately using SVM Model. Whenever an ultrasonic liver lesion image is given to the SVM classifier system, the features are calculated, classified, as normal and diseased liver lesion. We hope the result will be helpful to the physician to identify the liver cancer in non-invasive method.

Keywords: segmentation, Support Vector Machine, ultrasound liver lesion, co-occurance Matrix

Procedia PDF Downloads 405