Search results for: features engineering methods for forecasting
19512 Recurrent Neural Networks for Classifying Outliers in Electronic Health Record Clinical Text
Authors: Duncan Wallace, M-Tahar Kechadi
Abstract:
In recent years, Machine Learning (ML) approaches have been successfully applied to an analysis of patient symptom data in the context of disease diagnosis, at least where such data is well codified. However, much of the data present in Electronic Health Records (EHR) are unlikely to prove suitable for classic ML approaches. Furthermore, as scores of data are widely spread across both hospitals and individuals, a decentralized, computationally scalable methodology is a priority. The focus of this paper is to develop a method to predict outliers in an out-of-hours healthcare provision center (OOHC). In particular, our research is based upon the early identification of patients who have underlying conditions which will cause them to repeatedly require medical attention. OOHC act as an ad-hoc delivery of triage and treatment, where interactions occur without recourse to a full medical history of the patient in question. Medical histories, relating to patients contacting an OOHC, may reside in several distinct EHR systems in multiple hospitals or surgeries, which are unavailable to the OOHC in question. As such, although a local solution is optimal for this problem, it follows that the data under investigation is incomplete, heterogeneous, and comprised mostly of noisy textual notes compiled during routine OOHC activities. Through the use of Deep Learning methodologies, the aim of this paper is to provide the means to identify patient cases, upon initial contact, which are likely to relate to such outliers. To this end, we compare the performance of Long Short-Term Memory, Gated Recurrent Units, and combinations of both with Convolutional Neural Networks. A further aim of this paper is to elucidate the discovery of such outliers by examining the exact terms which provide a strong indication of positive and negative case entries. While free-text is the principal data extracted from EHRs for classification, EHRs also contain normalized features. Although the specific demographical features treated within our corpus are relatively limited in scope, we examine whether it is beneficial to include such features among the inputs to our neural network, or whether these features are more successfully exploited in conjunction with a different form of a classifier. In this section, we compare the performance of randomly generated regression trees and support vector machines and determine the extent to which our classification program can be improved upon by using either of these machine learning approaches in conjunction with the output of our Recurrent Neural Network application. The output of our neural network is also used to help determine the most significant lexemes present within the corpus for determining high-risk patients. By combining the confidence of our classification program in relation to lexemes within true positive and true negative cases, with an inverse document frequency of the lexemes related to these cases, we can determine what features act as the primary indicators of frequent-attender and non-frequent-attender cases, providing a human interpretable appreciation of how our program classifies cases.Keywords: artificial neural networks, data-mining, machine learning, medical informatics
Procedia PDF Downloads 13119511 Forecasting Etching Behavior Silica Sand Using the Design of Experiments Method
Authors: Kefaifi Aissa, Sahraoui Tahar, Kheloufi Abdelkrim, Anas Sabiha, Hannane Farouk
Abstract:
The aim of this study is to show how the Design of Experiments Method (DOE) can be put into use as a practical approach for silica sand etching behavior modeling during its primary step of leaching. In the present work, we have studied etching effect on particle size during a primary step of leaching process on Algerian silica sand with florid acid (HF) at 20% and 30 % during 4 and 8 hours. Therefore, a new purity of the sand is noted depending on the time of leaching. This study was expanded by a numerical approach using a method of experiment design, which shows the influence of each parameter and the interaction between them in the process and approved the obtained experimental results. This model is a predictive approach using hide software. Based on the measured parameters experimentally in the interior of the model, the use of DOE method can make it possible to predict the outside parameters of the model in question and can give us the optimize response without making the experimental measurement.Keywords: acid leaching, design of experiments method(DOE), purity silica, silica etching
Procedia PDF Downloads 28619510 A Review of Travel Data Collection Methods
Authors: Muhammad Awais Shafique, Eiji Hato
Abstract:
Household trip data is of crucial importance for managing present transportation infrastructure as well as to plan and design future facilities. It also provides basis for new policies implemented under Transportation Demand Management. The methods used for household trip data collection have changed with passage of time, starting with the conventional face-to-face interviews or paper-and-pencil interviews and reaching to the recent approach of employing smartphones. This study summarizes the step-wise evolution in the travel data collection methods. It provides a comprehensive review of the topic, for readers interested to know the changing trends in the data collection field.Keywords: computer, smartphone, telephone, travel survey
Procedia PDF Downloads 31319509 Unsupervised Part-of-Speech Tagging for Amharic Using K-Means Clustering
Authors: Zelalem Fantahun
Abstract:
Part-of-speech tagging is the process of assigning a part-of-speech or other lexical class marker to each word into naturally occurring text. Part-of-speech tagging is the most fundamental and basic task almost in all natural language processing. In natural language processing, the problem of providing large amount of manually annotated data is a knowledge acquisition bottleneck. Since, Amharic is one of under-resourced language, the availability of tagged corpus is the bottleneck problem for natural language processing especially for POS tagging. A promising direction to tackle this problem is to provide a system that does not require manually tagged data. In unsupervised learning, the learner is not provided with classifications. Unsupervised algorithms seek out similarity between pieces of data in order to determine whether they can be characterized as forming a group. This paper explicates the development of unsupervised part-of-speech tagger using K-Means clustering for Amharic language since large amount of data is produced in day-to-day activities. In the development of the tagger, the following procedures are followed. First, the unlabeled data (raw text) is divided into 10 folds and tokenization phase takes place; at this level, the raw text is chunked at sentence level and then into words. The second phase is feature extraction which includes word frequency, syntactic and morphological features of a word. The third phase is clustering. Among different clustering algorithms, K-means is selected and implemented in this study that brings group of similar words together. The fourth phase is mapping, which deals with looking at each cluster carefully and the most common tag is assigned to a group. This study finds out two features that are capable of distinguishing one part-of-speech from others these are morphological feature and positional information and show that it is possible to use unsupervised learning for Amharic POS tagging. In order to increase performance of the unsupervised part-of-speech tagger, there is a need to incorporate other features that are not included in this study, such as semantic related information. Finally, based on experimental result, the performance of the system achieves a maximum of 81% accuracy.Keywords: POS tagging, Amharic, unsupervised learning, k-means
Procedia PDF Downloads 45119508 Visualization of Quantitative Thresholds in Stocks
Authors: Siddhant Sahu, P. James Daniel Paul
Abstract:
Technical analysis comprised by various technical indicators is a holistic way of representing price movement of stocks in the market. Various forms of indicators have evolved from the primitive ones in the past decades. There have been many attempts to introduce volume as a major determinant to determine strong patterns in market forecasting. The law of demand defines the relationship between the volume and price. Most of the traders are familiar with the volume game. Including the time dimension to the law of demand provides a different visualization to the theory. While attempting the same, it was found that there are different thresholds in the market for different companies. These thresholds have a significant influence on the price. This article is an attempt in determining the thresholds for companies using the three dimensional graphs for optimizing the portfolios. It also emphasizes on the magnitude of importance of volumes as a key factor for determining of predicting strong price movements, bullish and bearish markets. It uses a comprehensive data set of major companies which form a major chunk of the Indian automotive sector and are thus used as an illustration.Keywords: technical analysis, expert system, law of demand, stocks, portfolio analysis, Indian automotive sector
Procedia PDF Downloads 31719507 3D Human Reconstruction over Cloud Based Image Data via AI and Machine Learning
Authors: Kaushik Sathupadi, Sandesh Achar
Abstract:
Human action recognition modeling is a critical task in machine learning. These systems require better techniques for recognizing body parts and selecting optimal features based on vision sensors to identify complex action patterns efficiently. Still, there is a considerable gap and challenges between images and videos, such as brightness, motion variation, and random clutters. This paper proposes a robust approach for classifying human actions over cloud-based image data. First, we apply pre-processing and detection, human and outer shape detection techniques. Next, we extract valuable information in terms of cues. We extract two distinct features: fuzzy local binary patterns and sequence representation. Then, we applied a greedy, randomized adaptive search procedure for data optimization and dimension reduction, and for classification, we used a random forest. We tested our model on two benchmark datasets, AAMAZ and the KTH Multi-view football datasets. Our HMR framework significantly outperforms the other state-of-the-art approaches and achieves a better recognition rate of 91% and 89.6% over the AAMAZ and KTH multi-view football datasets, respectively.Keywords: computer vision, human motion analysis, random forest, machine learning
Procedia PDF Downloads 3719506 Physical and Chemical Alternative Methods of Fresh Produce Disinfection
Authors: Tuji Jemal Ahmed
Abstract:
Fresh produce is an essential component of a healthy diet. However, it can also be a potential source of pathogenic microorganisms that can cause foodborne illnesses. Traditional disinfection methods, such as washing with water and chlorine, have limitations and may not effectively remove or inactivate all microorganisms. This has led to the development of alternative/new methods of fresh produce disinfection, including physical and chemical methods. In this paper, we explore the physical and chemical new methods of fresh produce disinfection, their advantages and disadvantages, and their suitability for different types of produce. Physical methods of disinfection, such as ultraviolet (UV) radiation and high-pressure processing (HPP), are crucial in ensuring the microbiological safety of fresh produce. UV radiation uses short-wavelength UV-C light to damage the DNA and RNA of microorganisms, and HPP applies high levels of pressure to fresh produce to reduce the microbial load. These physical methods are highly effective in killing a wide range of microorganisms, including bacteria, viruses, and fungi. However, they may not penetrate deep enough into the product to kill all microorganisms and can alter the sensory characteristics of the product. Chemical methods of disinfection, such as acidic electrolyzed water (AEW), ozone, and peroxyacetic acid (PAA), are also important in ensuring the microbiological safety of fresh produce. AEW uses a low concentration of hypochlorous acid and a high concentration of hydrogen ions to inactivate microorganisms, ozone uses ozone gas to damage the cell membranes and DNA of microorganisms, and PAA uses a combination of hydrogen peroxide and acetic acid to inactivate microorganisms. These chemical methods are highly effective in killing a wide range of microorganisms, but they may cause discoloration or changes in the texture and flavor of some products and may require specialized equipment and trained personnel to produce and apply. In conclusion, the selection of the most suitable method of fresh produce disinfection should take into consideration the type of product, the level of microbial contamination, the effectiveness of the method in reducing the microbial load, and any potential negative impacts on the sensory characteristics, nutritional composition, and safety of the produce.Keywords: fresh produce, pathogenic microorganisms, foodborne illnesses, disinfection methods
Procedia PDF Downloads 7419505 Optimal Framework of Policy Systems with Innovation: Use of Strategic Design for Evolution of Decisions
Authors: Yuna Lee
Abstract:
In the current policy process, there has been a growing interest in more open approaches that incorporate creativity and innovation based on the forecasting groups composed by the public and experts together into scientific data-driven foresight methods to implement more effective policymaking. Especially, citizen participation as collective intelligence in policymaking with design and deep scale of innovation at the global level has been developed and human-centred design thinking is considered as one of the most promising methods for strategic foresight. Yet, there is a lack of a common theoretical foundation for a comprehensive approach for the current situation of and post-COVID-19 era, and substantial changes in policymaking practice are insignificant and ongoing with trial and error. This project hypothesized that rigorously developed policy systems and tools that support strategic foresight by considering the public understanding could maximize ways to create new possibilities for a preferable future, however, it must involve a better understating of Behavioural Insights, including individual and cultural values, profit motives and needs, and psychological motivations, for implementing holistic and multilateral foresight and creating more positive possibilities. To what extent is the policymaking system theoretically possible that incorporates the holistic and comprehensive foresight and policy process implementation, assuming that theory and practice, in reality, are different and not connected? What components and environmental conditions should be included in the strategic foresight system to enhance the capacity of decision from policymakers to predict alternative futures, or detect uncertainties of the future more accurately? And, compared to the required environmental condition, what are the environmental vulnerabilities of the current policymaking system? In this light, this research contemplates the question of how effectively policymaking practices have been implemented through the synthesis of scientific, technology-oriented innovation with the strategic design for tackling complex societal challenges and devising more significant insights to make society greener and more liveable. Here, this study conceptualizes the notions of a new collaborative way of strategic foresight that aims to maximize mutual benefits between policy actors and citizens through the cooperation stemming from evolutionary game theory. This study applies mixed methodology, including interviews of policy experts, with the case in which digital transformation and strategic design provided future-oriented solutions or directions to cities’ sustainable development goals and society-wide urgent challenges such as COVID-19. As a result, artistic and sensual interpreting capabilities through strategic design promote a concrete form of ideas toward a stable connection from the present to the future and enhance the understanding and active cooperation among decision-makers, stakeholders, and citizens. Ultimately, an improved theoretical foundation proposed in this study is expected to help strategically respond to the highly interconnected future changes of the post-COVID-19 world.Keywords: policymaking, strategic design, sustainable innovation, evolution of cooperation
Procedia PDF Downloads 19419504 Identification of Landslide Features Using Back-Propagation Neural Network on LiDAR Digital Elevation Model
Authors: Chia-Hao Chang, Geng-Gui Wang, Jee-Cheng Wu
Abstract:
The prediction of a landslide is a difficult task because it requires a detailed study of past activities using a complete range of investigative methods to determine the changing condition. In this research, first step, LiDAR 1-meter by 1-meter resolution of digital elevation model (DEM) was used to generate six environmental factors of landslide. Then, back-propagation neural networks (BPNN) was adopted to identify scarp, landslide areas and non-landslide areas. The BPNN uses 6 environmental factors in input layer and 1 output layer. Moreover, 6 landslide areas are used as training areas and 4 landslide areas as test areas in the BPNN. The hidden layer is set to be 1 and 2; the hidden layer neurons are set to be 4, 5, 6, 7 and 8; the learning rates are set to be 0.01, 0.1 and 0.5. When using 1 hidden layer with 7 neurons and the learning rate sets to be 0.5, the result of Network training root mean square error is 0.001388. Finally, evaluation of BPNN classification accuracy by the confusion matrix shows that the overall accuracy can reach 94.4%, and the Kappa value is 0.7464.Keywords: digital elevation model, DEM, environmental factors, back-propagation neural network, BPNN, LiDAR
Procedia PDF Downloads 14419503 Use of Machine Learning Algorithms to Pediatric MR Images for Tumor Classification
Authors: I. Stathopoulos, V. Syrgiamiotis, E. Karavasilis, A. Ploussi, I. Nikas, C. Hatzigiorgi, K. Platoni, E. P. Efstathopoulos
Abstract:
Introduction: Brain and central nervous system (CNS) tumors form the second most common group of cancer in children, accounting for 30% of all childhood cancers. MRI is the key imaging technique used for the visualization and management of pediatric brain tumors. Initial characterization of tumors from MRI scans is usually performed via a radiologist’s visual assessment. However, different brain tumor types do not always demonstrate clear differences in visual appearance. Using only conventional MRI to provide a definite diagnosis could potentially lead to inaccurate results, and so histopathological examination of biopsy samples is currently considered to be the gold standard for obtaining definite diagnoses. Machine learning is defined as the study of computational algorithms that can use, complex or not, mathematical relationships and patterns from empirical and scientific data to make reliable decisions. Concerning the above, machine learning techniques could provide effective and accurate ways to automate and speed up the analysis and diagnosis for medical images. Machine learning applications in radiology are or could potentially be useful in practice for medical image segmentation and registration, computer-aided detection and diagnosis systems for CT, MR or radiography images and functional MR (fMRI) images for brain activity analysis and neurological disease diagnosis. Purpose: The objective of this study is to provide an automated tool, which may assist in the imaging evaluation and classification of brain neoplasms in pediatric patients by determining the glioma type, grade and differentiating between different brain tissue types. Moreover, a future purpose is to present an alternative way of quick and accurate diagnosis in order to save time and resources in the daily medical workflow. Materials and Methods: A cohort, of 80 pediatric patients with a diagnosis of posterior fossa tumor, was used: 20 ependymomas, 20 astrocytomas, 20 medulloblastomas and 20 healthy children. The MR sequences used, for every single patient, were the following: axial T1-weighted (T1), axial T2-weighted (T2), FluidAttenuated Inversion Recovery (FLAIR), axial diffusion weighted images (DWI), axial contrast-enhanced T1-weighted (T1ce). From every sequence only a principal slice was used that manually traced by two expert radiologists. Image acquisition was carried out on a GE HDxt 1.5-T scanner. The images were preprocessed following a number of steps including noise reduction, bias-field correction, thresholding, coregistration of all sequences (T1, T2, T1ce, FLAIR, DWI), skull stripping, and histogram matching. A large number of features for investigation were chosen, which included age, tumor shape characteristics, image intensity characteristics and texture features. After selecting the features for achieving the highest accuracy using the least number of variables, four machine learning classification algorithms were used: k-Nearest Neighbour, Support-Vector Machines, C4.5 Decision Tree and Convolutional Neural Network. The machine learning schemes and the image analysis are implemented in the WEKA platform and MatLab platform respectively. Results-Conclusions: The results and the accuracy of images classification for each type of glioma by the four different algorithms are still on process.Keywords: image classification, machine learning algorithms, pediatric MRI, pediatric oncology
Procedia PDF Downloads 14919502 Electro-Mechanical Response and Engineering Properties of Piezocomposite with Imperfect Interface
Authors: Rattanan Tippayaphalapholgul, Yasothorn Sapsathiarn
Abstract:
Composites of piezoelectric materials are widely use in practical applications such as nondestructive testing devices, smart adaptive structures and medical devices. A thorough understanding of coupled electro-elastic response and properties of piezocomposite are crucial for the development and design of piezoelectric composite materials used in advanced applications. The micromechanics analysis is employed in this paper to determine the response and engineering properties of the piezocomposite. A mechanical imperfect interface bonding between piezoelectric inclusion and polymer matrix is taken into consideration in the analysis. The micromechanics analysis is based on the Boundary Element Method (BEM) together with the periodic micro-field micromechanics theory. A selected set of numerical results is presented to investigate the influence of volume ratio and interface bonding condition on effective piezocomposite material coefficients and portray basic features of coupled electroelastic response within the domain of piezocomposite unit cell.Keywords: effective engineering properties, electroelastic response, imperfect interface, piezocomposite
Procedia PDF Downloads 23119501 Energy Communities from Municipality Level to Province Level: A Comparison Using Autoregressive Integrated Moving Average Model
Authors: Amro Issam Hamed Attia Ramadan, Marco Zappatore, Pasquale Balena, Antonella Longo
Abstract:
Considering the energetic crisis that is hitting Europe, it becomes more and more necessary to change the energy policies to depend less on fossil fuels and replace them with energy from renewable sources. This has triggered the urge to use clean energy not only to satisfy energy needs and fulfill the required consumption but also to decrease the danger of climatic changes due to harmful emissions. Many countries have already started creating energetic communities based on renewable energy sources. The first step to understanding energy needs in any place is to perfectly know the consumption. In this work, we aim to estimate electricity consumption for a municipality that makes up part of a rural area located in southern Italy using forecast models that allow for the estimation of electricity consumption for the next ten years, and we then apply the same model to the province where the municipality is located and estimate the future consumption for the same period to examine whether it is possible to start from the municipality level to reach the province level when creating energy communities.Keywords: ARIMA, electricity consumption, forecasting models, time series
Procedia PDF Downloads 17419500 Rendering Cognition Based Learning in Coherence with Development within the Context of PostgreSQL
Authors: Manuela Nayantara Jeyaraj, Senuri Sucharitharathna, Chathurika Senarath, Yasanthy Kanagaraj, Indraka Udayakumara
Abstract:
PostgreSQL is an Object Relational Database Management System (ORDBMS) that has been in existence for a while. Despite the superior features that it wraps and packages to manage database and data, the database community has not fully realized the importance and advantages of PostgreSQL. Hence, this research tends to focus on provisioning a better environment of development for PostgreSQL in order to induce the utilization and elucidate the importance of PostgreSQL. PostgreSQL is also known to be the world’s most elementary SQL-compliant open source ORDBMS. But, users have not yet resolved to PostgreSQL due to the facts that it is still under the layers and the complexity of its persistent textual environment for an introductory user. Simply stating this, there is a dire need to explicate an easy way of making the users comprehend the procedure and standards with which databases are created, tables and the relationships among them, manipulating queries and their flow based on conditions in PostgreSQL to help the community resolve to PostgreSQL at an augmented rate. Hence, this research under development within the context tends to initially identify the dominant features provided by PostgreSQL over its competitors. Following the identified merits, an analysis on why the database community holds a hesitance in migrating to PostgreSQL’s environment will be carried out. These will be modulated and tailored based on the scope and the constraints discovered. The resultant of the research proposes a system that will serve as a designing platform as well as a learning tool that will provide an interactive method of learning via a visual editor mode and incorporate a textual editor for well-versed users. The study is based on conjuring viable solutions that analyze a user’s cognitive perception in comprehending human computer interfaces and the behavioural processing of design elements. By providing a visually draggable and manipulative environment to work with Postgresql databases and table queries, it is expected to highlight the elementary features displayed by Postgresql over any other existent systems in order to grasp and disseminate the importance and simplicity offered by this to a hesitant user.Keywords: cognition, database, PostgreSQL, text-editor, visual-editor
Procedia PDF Downloads 28319499 Transforming Water-Energy-Gas Industry through Smart Metering and Blockchain Technology
Authors: Khoi A. Nguyen, Rodney A. Stewart, Hong Zhang
Abstract:
Advanced metering technologies coupled with informatics creates an opportunity to form digital multi-utility service providers. These providers will be able to concurrently collect a customers’ medium-high resolution water, electricity and gas demand data and provide user-friendly platforms to feed this information back to customers and supply/distribution utility organisations. With the emergence of blockchain technology, a new research area has been explored which helps bring this multi-utility service provider concept to a much higher level. This study aims at introducing a breakthrough system architecture where smart metering technology in water, energy, and gas (WEG) are combined with blockchain technology to provide customer a novel real-time consumption report and decentralized resource trading platform. A pilot study on 4 properties in Australia has been undertaken to demonstrate this system, where benefits for customers and utilities are undeniable.Keywords: blockchain, digital multi-utility, end use, demand forecasting
Procedia PDF Downloads 17119498 Improvement of Bone Scintography Image Using Image Texture Analysis
Authors: Yousif Mohamed Y. Abdallah, Eltayeb Wagallah
Abstract:
Image enhancement allows the observer to see details in images that may not be immediately observable in the original image. Image enhancement is the transformation or mapping of one image to another. The enhancement of certain features in images is accompanied by undesirable effects. To achieve maximum image quality after denoising, a new, low order, local adaptive Gaussian scale mixture model and median filter were presented, which accomplishes nonlinearities from scattering a new nonlinear approach for contrast enhancement of bones in bone scan images using both gamma correction and negative transform methods. The usual assumption of a distribution of gamma and Poisson statistics only lead to overestimation of the noise variance in regions of low intensity but to underestimation in regions of high intensity and therefore to non-optional results. The contrast enhancement results were obtained and evaluated using MatLab program in nuclear medicine images of the bones. The optimal number of bins, in particular the number of gray-levels, is chosen automatically using entropy and average distance between the histogram of the original gray-level distribution and the contrast enhancement function’s curve.Keywords: bone scan, nuclear medicine, Matlab, image processing technique
Procedia PDF Downloads 50919497 The Observable Method for the Regularization of Shock-Interface Interactions
Authors: Teng Li, Kamran Mohseni
Abstract:
This paper presents an inviscid regularization technique that is capable of regularizing the shocks and sharp interfaces simultaneously in the shock-interface interaction simulations. The direct numerical simulation of flows involving shocks has been investigated for many years and a lot of numerical methods were developed to capture the shocks. However, most of these methods rely on the numerical dissipation to regularize the shocks. Moreover, in high Reynolds number flows, the nonlinear terms in hyperbolic Partial Differential Equations (PDE) dominates, constantly generating small scale features. This makes direct numerical simulation of shocks even harder. The same difficulty happens in two-phase flow with sharp interfaces where the nonlinear terms in the governing equations keep sharpening the interfaces to discontinuities. The main idea of the proposed technique is to average out the small scales that is below the resolution (observable scale) of the computational grid by filtering the convective velocity in the nonlinear terms in the governing PDE. This technique is named “observable method” and it results in a set of hyperbolic equations called observable equations, namely, observable Navier-Stokes or Euler equations. The observable method has been applied to the flow simulations involving shocks, turbulence, and two-phase flows, and the results are promising. In the current paper, the observable method is examined on the performance of regularizing shocks and interfaces at the same time in shock-interface interaction problems. Bubble-shock interactions and Richtmyer-Meshkov instability are particularly chosen to be studied. Observable Euler equations will be numerically solved with pseudo-spectral discretization in space and third order Total Variation Diminishing (TVD) Runge Kutta method in time. Results are presented and compared with existing publications. The interface acceleration and deformation and shock reflection are particularly examined.Keywords: compressible flow simulation, inviscid regularization, Richtmyer-Meshkov instability, shock-bubble interactions.
Procedia PDF Downloads 34919496 Linguistic Features for Sentence Difficulty Prediction in Aspect-Based Sentiment Analysis
Authors: Adrian-Gabriel Chifu, Sebastien Fournier
Abstract:
One of the challenges of natural language understanding is to deal with the subjectivity of sentences, which may express opinions and emotions that add layers of complexity and nuance. Sentiment analysis is a field that aims to extract and analyze these subjective elements from text, and it can be applied at different levels of granularity, such as document, paragraph, sentence, or aspect. Aspect-based sentiment analysis is a well-studied topic with many available data sets and models. However, there is no clear definition of what makes a sentence difficult for aspect-based sentiment analysis. In this paper, we explore this question by conducting an experiment with three data sets: ”Laptops”, ”Restaurants”, and ”MTSC” (Multi-Target-dependent Sentiment Classification), and a merged version of these three datasets. We study the impact of domain diversity and syntactic diversity on difficulty. We use a combination of classifiers to identify the most difficult sentences and analyze their characteristics. We employ two ways of defining sentence difficulty. The first one is binary and labels a sentence as difficult if the classifiers fail to correctly predict the sentiment polarity. The second one is a six-level scale based on how many of the top five best-performing classifiers can correctly predict the sentiment polarity. We also define 9 linguistic features that, combined, aim at estimating the difficulty at sentence level.Keywords: sentiment analysis, difficulty, classification, machine learning
Procedia PDF Downloads 8919495 Requirements Gathering for Improved Software Usability and the Potential for Usage-Centred Design
Authors: Kholod J. Alotaibi, Andrew M. Gravell
Abstract:
Usability is an important software quality that is often neglected at the design stage. Although methods exist to incorporate elements of usability engineering, there is a need for more balanced usability focused methods that can enhance the experience of software usability for users. In this regard, the potential for Usage-Centered Design is explored with respect to requirements gathering and is shown to lead to high software usability besides other benefits. It achieves this through its focus on usage, defining essential use cases, by conducting task modeling, encouraging user collaboration, refining requirements, and so on. The requirements gathering process in UgCD is described in detail.Keywords: requirements gathering, usability, usage-centred design, computer science
Procedia PDF Downloads 35819494 Comprehensive Review of Ultralightweight Security Protocols
Authors: Prashansa Singh, Manjot Kaur, Rohit Bajaj
Abstract:
The proliferation of wireless sensor networks and Internet of Things (IoT) devices in the quickly changing digital landscape has highlighted the urgent need for strong security solutions that can handle these systems’ limited resources. A key solution to this problem is the emergence of ultralightweight security protocols, which provide strong security features while respecting the strict computational, energy, and memory constraints imposed on these kinds of devices. This in-depth analysis explores the field of ultralightweight security protocols, offering a thorough examination of their evolution, salient features, and the particular security issues they resolve. We carefully examine and contrast different protocols, pointing out their advantages and disadvantages as well as the compromises between resource limitations and security resilience. We also study these protocols’ application domains, including the Internet of Things, RFID systems, and wireless sensor networks, to name a few. In addition, the review highlights recent developments and advancements in the field, pointing out new trends and possible avenues for future research. This paper aims to be a useful resource for researchers, practitioners, and developers, guiding the design and implementation of safe, effective, and scalable systems in the Internet of Things era by providing a comprehensive overview of ultralightweight security protocols.Keywords: wireless sensor network, machine-to-machine, MQTT broker, server, ultralightweight, TCP/IP
Procedia PDF Downloads 8219493 PM₁₀ and PM2.5 Concentrations in Bangkok over Last 10 Years: Implications for Air Quality and Health
Authors: Tin Thongthammachart, Wanida Jinsart
Abstract:
Atmospheric particulate matter particles with a diameter less than 10 microns (PM₁₀) and less than 2.5 microns (PM₂.₅) have adverse health effect. The impact from PM was studied from both health and regulatory perspective. Ambient PM data was collected over ten years in Bangkok and vicinity areas of Thailand from 2007 to 2017. Statistical models were used to forecast PM concentrations from 2018 to 2020. Monitoring monthly data averaged concentration of PM₁₀ and PM₂.₅ were used as input to forecast the monthly average concentration of PM. The forecasting results were validated by root means square error (RMSE). The predicted results were used to determine hazard risk for the carcinogenic disease. The health risk values were interpolated with GIS with ordinary kriging technique to create hazard maps in Bangkok and vicinity area. GIS-based maps illustrated the variability of PM distribution and high-risk locations. These evaluated results could support national policy for the sake of human health.Keywords: PM₁₀, PM₂.₅, statistical models, atmospheric particulate matter
Procedia PDF Downloads 15919492 Deep Learning for Recommender System: Principles, Methods and Evaluation
Authors: Basiliyos Tilahun Betru, Charles Awono Onana, Bernabe Batchakui
Abstract:
Recommender systems have become increasingly popular in recent years, and are utilized in numerous areas. Nowadays many web services provide several information for users and recommender systems have been developed as critical element of these web applications to predict choice of preference and provide significant recommendations. With the help of the advantage of deep learning in modeling different types of data and due to the dynamic change of user preference, building a deep model can better understand users demand and further improve quality of recommendation. In this paper, deep neural network models for recommender system are evaluated. Most of deep neural network models in recommender system focus on the classical collaborative filtering user-item setting. Deep learning models demonstrated high level features of complex data can be learned instead of using metadata which can significantly improve accuracy of recommendation. Even though deep learning poses a great impact in various areas, applying the model to a recommender system have not been fully exploited and still a lot of improvements can be done both in collaborative and content-based approach while considering different contextual factors.Keywords: big data, decision making, deep learning, recommender system
Procedia PDF Downloads 47819491 Features of Calculating Structures for Frequent Weak Earthquakes
Authors: M. S. Belashov, A. V. Benin, Lin Hong, Sh. Sh. Nazarova, O. B. Sabirova, A. M. Uzdin, Lin Hong
Abstract:
The features of calculating structures for the action of weak earthquakes are analyzed. Earthquakes with a recurrence of 30 years and 50 years are considered. In the first case, the structure is to operate normally without damage after the earthquake. In the second case, damages are allowed that do not affect the possibility of the structure operation. Three issues are emphasized: setting elastic and damping characteristics of reinforced concrete, formalization of limit states, and combinations of loads. The dependence of damping on the reinforcement coefficient is estimated. When evaluating limit states, in addition to calculations for crack resistance and strength, a human factor, i.e., the possibility of panic among people, was considered. To avoid it, it is proposed to limit a floor-by-floor speed level in certain octave ranges. Proposals have been developed for estimating the coefficients of the combination of various loads with the seismic one. As an example, coefficients of combinations of seismic and ice loads are estimated. It is shown that for strong actions, the combination coefficients for different regions turn out to be close, while for weak actions, they may differ.Keywords: weak earthquake, frequent earthquake, damage, limit state, reinforcement, crack resistance, strength resistance, a floor-by-floor velocity, combination coefficients
Procedia PDF Downloads 8819490 Performance Comparison of Wideband Covariance Matrix Sparse Representation (W-CMSR) with Other Wideband DOA Estimation Methods
Authors: Sandeep Santosh, O. P. Sahu
Abstract:
In this paper, performance comparison of wideband covariance matrix sparse representation (W-CMSR) method with other existing wideband Direction of Arrival (DOA) estimation methods has been made.W-CMSR relies less on a priori information of the incident signal number than the ordinary subspace based methods.Consider the perturbation free covariance matrix of the wideband array output. The diagonal covariance elements are contaminated by unknown noise variance. The covariance matrix of array output is conjugate symmetric i.e its upper right triangular elements can be represented by lower left triangular ones.As the main diagonal elements are contaminated by unknown noise variance,slide over them and align the lower left triangular elements column by column to obtain a measurement vector.Simulation results for W-CMSR are compared with simulation results of other wideband DOA estimation methods like Coherent signal subspace method (CSSM), Capon, l1-SVD, and JLZA-DOA. W-CMSR separate two signals very clearly and CSSM, Capon, L1-SVD and JLZA-DOA fail to separate two signals clearly and an amount of pseudo peaks exist in the spectrum of L1-SVD.Keywords: W-CMSR, wideband direction of arrival (DOA), covariance matrix, electrical and computer engineering
Procedia PDF Downloads 47119489 Comparison between Deterministic and Probabilistic Stability Analysis, Featuring Consequent Risk Assessment
Authors: Isabela Moreira Queiroz
Abstract:
Slope stability analyses are largely carried out by deterministic methods and evaluated through a single security factor. Although it is known that the geotechnical parameters can present great dispersal, such analyses are considered fixed and known. The probabilistic methods, in turn, incorporate the variability of input key parameters (random variables), resulting in a range of values of safety factors, thus enabling the determination of the probability of failure, which is an essential parameter in the calculation of the risk (probability multiplied by the consequence of the event). Among the probabilistic methods, there are three frequently used methods in geotechnical society: FOSM (First-Order, Second-Moment), Rosenblueth (Point Estimates) and Monte Carlo. This paper presents a comparison between the results from deterministic and probabilistic analyses (FOSM method, Monte Carlo and Rosenblueth) applied to a hypothetical slope. The end was held to evaluate the behavior of the slope and consequent risk analysis, which is used to calculate the risk and analyze their mitigation and control solutions. It can be observed that the results obtained by the three probabilistic methods were quite close. It should be noticed that the calculation of the risk makes it possible to list the priority to the implementation of mitigation measures. Therefore, it is recommended to do a good assessment of the geological-geotechnical model incorporating the uncertainty in viability, design, construction, operation and closure by means of risk management.Keywords: probabilistic methods, risk assessment, risk management, slope stability
Procedia PDF Downloads 39219488 Structural Health Monitoring and Damage Structural Identification Using Dynamic Response
Authors: Reza Behboodian
Abstract:
Monitoring the structural health and diagnosing their damage in the early stages has always been one of the topics of concern. Nowadays, research on structural damage detection methods based on vibration analysis is very extensive. Moreover, these methods can be used as methods of permanent and timely inspection of structures and prevent further damage to structures. Non-destructive methods are the low-cost and economical methods for determining the damage of structures. In this research, a non-destructive method for detecting and identifying the failure location in structures based on dynamic responses resulting from time history analysis is proposed. When the structure is damaged due to the reduction of stiffness, and due to the applied loads, the displacements in different parts of the structure were increased. In the proposed method, the damage position is determined based on the calculation of the strain energy difference in each member of the damaged structure and the healthy structure at any time. Defective members of the structure are indicated by the amount of strain energy relative to the healthy state. The results indicated that the proper accuracy and performance of the proposed method for identifying failure in structures.Keywords: failure, time history analysis, dynamic response, strain energy
Procedia PDF Downloads 13319487 Optimization of Machine Learning Regression Results: An Application on Health Expenditures
Authors: Songul Cinaroglu
Abstract:
Machine learning regression methods are recommended as an alternative to classical regression methods in the existence of variables which are difficult to model. Data for health expenditure is typically non-normal and have a heavily skewed distribution. This study aims to compare machine learning regression methods by hyperparameter tuning to predict health expenditure per capita. A multiple regression model was conducted and performance results of Lasso Regression, Random Forest Regression and Support Vector Machine Regression recorded when different hyperparameters are assigned. Lambda (λ) value for Lasso Regression, number of trees for Random Forest Regression, epsilon (ε) value for Support Vector Regression was determined as hyperparameters. Study results performed by using 'k' fold cross validation changed from 5 to 50, indicate the difference between machine learning regression results in terms of R², RMSE and MAE values that are statistically significant (p < 0.001). Study results reveal that Random Forest Regression (R² ˃ 0.7500, RMSE ≤ 0.6000 ve MAE ≤ 0.4000) outperforms other machine learning regression methods. It is highly advisable to use machine learning regression methods for modelling health expenditures.Keywords: machine learning, lasso regression, random forest regression, support vector regression, hyperparameter tuning, health expenditure
Procedia PDF Downloads 22619486 Umbilical Cord-Derived Cells in Corneal Epithelial Regeneration
Authors: Hasan Mahmud Reza
Abstract:
Extensive studies of the human umbilical cord, both basic and translational, over the last three decades have unveiled a plethora of information. The cord lining harbors at least two phenotypically different multipotent stem cells: mesenchymal stem cells (MSCs) and cord lining epithelial stem cells (CLECs). These cells exhibit a mixed genetic profiling of both embryonic and adult stem cells, hence display a broader stem features than cells from other sources. We have observed that umbilical cord-derived cells are immunologically privileged and non-tumorigenic by animal study. These cells are ethically acceptable, thus provides a significant advantage over other stem cells. The high proliferative capacity, viability, differentiation potential, and superior harvest of these cells have made them better candidates in comparison to contemporary adult stem cells. Following 30 replication cycles, these cells have been observed to retain their stemness, with their phenotype and karyotype intact. Transplantation of bioengineered CLEC sheets in limbal stem cell-deficient rabbit eyes resulted in regeneration of clear cornea with phenotypic expression of the normal cornea-specific epithelial cytokeratin markers. The striking features of low immunogenicity protecting self along with co-transplanted allografts from rejection largely define the transplantation potential of umbilical cord-derived stem cells.Keywords: cord lining epithelial stem cells, mesenchymal stem cell, regenerative medicine, umbilical cord
Procedia PDF Downloads 15619485 Modeling Engagement with Multimodal Multisensor Data: The Continuous Performance Test as an Objective Tool to Track Flow
Authors: Mohammad H. Taheri, David J. Brown, Nasser Sherkat
Abstract:
Engagement is one of the most important factors in determining successful outcomes and deep learning in students. Existing approaches to detect student engagement involve periodic human observations that are subject to inter-rater reliability. Our solution uses real-time multimodal multisensor data labeled by objective performance outcomes to infer the engagement of students. The study involves four students with a combined diagnosis of cerebral palsy and a learning disability who took part in a 3-month trial over 59 sessions. Multimodal multisensor data were collected while they participated in a continuous performance test. Eye gaze, electroencephalogram, body pose, and interaction data were used to create a model of student engagement through objective labeling from the continuous performance test outcomes. In order to achieve this, a type of continuous performance test is introduced, the Seek-X type. Nine features were extracted including high-level handpicked compound features. Using leave-one-out cross-validation, a series of different machine learning approaches were evaluated. Overall, the random forest classification approach achieved the best classification results. Using random forest, 93.3% classification for engagement and 42.9% accuracy for disengagement were achieved. We compared these results to outcomes from different models: AdaBoost, decision tree, k-Nearest Neighbor, naïve Bayes, neural network, and support vector machine. We showed that using a multisensor approach achieved higher accuracy than using features from any reduced set of sensors. We found that using high-level handpicked features can improve the classification accuracy in every sensor mode. Our approach is robust to both sensor fallout and occlusions. The single most important sensor feature to the classification of engagement and distraction was shown to be eye gaze. It has been shown that we can accurately predict the level of engagement of students with learning disabilities in a real-time approach that is not subject to inter-rater reliability, human observation or reliant on a single mode of sensor input. This will help teachers design interventions for a heterogeneous group of students, where teachers cannot possibly attend to each of their individual needs. Our approach can be used to identify those with the greatest learning challenges so that all students are supported to reach their full potential.Keywords: affective computing in education, affect detection, continuous performance test, engagement, flow, HCI, interaction, learning disabilities, machine learning, multimodal, multisensor, physiological sensors, student engagement
Procedia PDF Downloads 9419484 Method to Create Signed Word - Application in Teaching and Learning Vietnamese Sign Language
Authors: Nguyen Thi Kim Thoa
Abstract:
Vietnam currently has about two million five hundred deaf/hard of hearing people. Although the issue of Vietnamese Sign Language (VSL) education has received attention from the State, there are still many issues that need to be resolved, such as policies, teacher training in both knowledge and teaching methods, education programs, and textbook compilation. Furthermore, the issue of research on VSL has not yet attracted the attention of linguists. Using the quantitative description method, the article will analyze, synthesize, and compare to find methods to create signed words in VSL, such as based on external shape characteristics, operational characteristics, operating methods, and basic meanings, from which we can see the special nature of signed words, the division of word types and the morphological meaning of creating new words through sign methods. From the results of this research, the aspect of ‘visual culture’ will be clarified in Vietnamese Deaf Culture. Through that, we also develop a number of vocabulary teaching methods (such as teaching vocabulary through a group of methods of forming signed words, teaching vocabulary using mind maps, and teaching vocabulary through culture...), with the aim of further improving the effectiveness of teaching and learning VSL in Vietnam. The research results also provide deaf people in Vietnam with a scientific and effective method of learning vocabulary, helping them quickly integrate into the community. The article will be a useful reference for linguists who want to research VSL.Keywords: Vietnamese sign language (VSL), signed word, teaching, method
Procedia PDF Downloads 3619483 The System of Uniform Criteria for the Characterization and Evaluation of Elements of Economic Structure: The Territory, Infrastructure, Processes, Technological Chains, the End Products
Authors: Aleksandr A. Gajour, Vladimir G. Merzlikin, Vladimir I. Veselov
Abstract:
This paper refers to the analysis of the characteristics of industrial and lifestyle facilities heat- energy objects as a part of the thermal envelope of Earth's surface for inclusion in any database of economic forecasting. The idealized model of the Earth's surface is discussed. This model gives the opportunity to obtain the energy equivalent for each element of terrain and world ocean. Energy efficiency criterion of comfortable human existence is introduced. Dynamics of changes of this criterion offers the possibility to simulate the possible technogenic catastrophes with the spontaneous industrial development of the certain Earth areas. Calculated model with the confirmed forecast of the Gulf Stream freezing in the polar regions in 2011 due to the heat-energy balance disturbance for the oceanic subsurface oil polluted layer is given. Two opposing trends of human development under limited and unlimited amount of heat-energy resources are analyzed.Keywords: Earth's surface, heat-energy consumption, energy criteria, technogenic catastrophes
Procedia PDF Downloads 402