Search results for: classification of patterns
3710 Application of MALDI-MS to Differentiate SARS-CoV-2 and Non-SARS-CoV-2 Symptomatic Infections in the Early and Late Phases of the Pandemic
Authors: Dmitriy Babenko, Sergey Yegorov, Ilya Korshukov, Aidana Sultanbekova, Valentina Barkhanskaya, Tatiana Bashirova, Yerzhan Zhunusov, Yevgeniya Li, Viktoriya Parakhina, Svetlana Kolesnichenko, Yeldar Baiken, Aruzhan Pralieva, Zhibek Zhumadilova, Matthew S. Miller, Gonzalo H. Hortelano, Anar Turmuhambetova, Antonella E. Chesca, Irina Kadyrova
Abstract:
Introduction: The rapidly evolving COVID-19 pandemic, along with the re-emergence of pathogens causing acute respiratory infections (ARI), has necessitated the development of novel diagnostic tools to differentiate various causes of ARI. MALDI-MS, due to its wide usage and affordability, has been proposed as a potential instrument for diagnosing SARS-CoV-2 versus non-SARS-CoV-2 ARI. The aim of this study was to investigate the potential of MALDI-MS in conjunction with a machine learning model to accurately distinguish between symptomatic infections caused by SARS-CoV-2 and non-SARS-CoV-2 during both the early and later phases of the pandemic. Furthermore, this study aimed to analyze mass spectrometry (MS) data obtained from nasal swabs of healthy individuals. Methods: We gathered mass spectra from 252 samples, comprising 108 SARS-CoV-2-positive samples obtained in 2020 (Covid 2020), 7 SARS-CoV- 2-positive samples obtained in 2023 (Covid 2023), 71 samples from symptomatic individuals without SARS-CoV-2 (Control non-Covid ARVI), and 66 samples from healthy individuals (Control healthy). All the samples were subjected to RT-PCR testing. For data analysis, we employed the caret R package to train and test seven machine-learning algorithms: C5.0, KNN, NB, RF, SVM-L, SVM-R, and XGBoost. We conducted a training process using a five-fold (outer) nested repeated (five times) ten-fold (inner) cross-validation with a randomized stratified splitting approach. Results: In this study, we utilized the Covid 2020 dataset as a case group and the non-Covid ARVI dataset as a control group to train and test various machine learning (ML) models. Among these models, XGBoost and SVM-R demonstrated the highest performance, with accuracy values of 0.97 [0.93, 0.97] and 0.95 [0.95; 0.97], specificity values of 0.86 [0.71; 0.93] and 0.86 [0.79; 0.87], and sensitivity values of 0.984 [0.984; 1.000] and 1.000 [0.968; 1.000], respectively. When examining the Covid 2023 dataset, the Naive Bayes model achieved the highest classification accuracy of 43%, while XGBoost and SVM-R achieved accuracies of 14%. For the healthy control dataset, the accuracy of the models ranged from 0.27 [0.24; 0.32] for k-nearest neighbors to 0.44 [0.41; 0.45] for the Support Vector Machine with a radial basis function kernel. Conclusion: Therefore, ML models trained on MALDI MS of nasopharyngeal swabs obtained from patients with Covid during the initial phase of the pandemic, as well as symptomatic non-Covid individuals, showed excellent classification performance, which aligns with the results of previous studies. However, when applied to swabs from healthy individuals and a limited sample of patients with Covid in the late phase of the pandemic, ML models exhibited lower classification accuracy.Keywords: SARS-CoV-2, MALDI-TOF MS, ML models, nasopharyngeal swabs, classification
Procedia PDF Downloads 1063709 Investigation of Different Stimulation Patterns to Reduce Muscle Fatigue during Functional Electrical Stimulation
Abstract:
Functional electrical stimulation (FES) is a commonly used technique in rehabilitation and often associated with rapid muscle fatigue which becomes the limiting factor in its applications. The objective of this study is to investigate the effects on the onset of fatigue of conventional synchronous stimulation, as well as asynchronous stimulation that mimic voluntary muscle activation targeting different motor units which are activated sequentially or randomly via multiple pairs of stimulation electrodes. We investigate three different approaches with various electrode configurations, as well as different patterns of stimulation applied to the gastrocnemius muscle: Conventional Synchronous Stimulation (CSS), Asynchronous Sequential Stimulation (ASS) and Asynchronous Random Stimulation (ARS). Stimulation was applied repeatedly for 300 ms followed by 700 ms of no-stimulation with 40 Hz effective frequency for all protocols. Ten able-bodied volunteers (28±3 years old) participated in this study. As fatigue indicators, we focused on the analysis of Normalized Fatigue Index (NFI), Fatigue Time Interval (FTI) and pre-post Twitch-Tetanus Ratio (ΔTTR). The results demonstrated that ASS and ARS give higher NFI and longer FTI confirming less fatigue for asynchronous stimulation. In addition, ASS and ARS resulted in higher ΔTTR than conventional CSS. In this study, we proposed a randomly distributed stimulation method for the application of FES and investigated its suitability for reducing muscle fatigue compared to previously applied methods. The results validated that asynchronous stimulation reduces fatigue, and indicates that random stimulation may improve fatigue resistance in some conditions.Keywords: asynchronous stimulation, electrode configuration, functional electrical stimulation (FES), muscle fatigue, pattern stimulation, random stimulation, sequential stimulation, synchronous stimulation
Procedia PDF Downloads 3063708 Application of Multilayer Perceptron and Markov Chain Analysis Based Hybrid-Approach for Predicting and Monitoring the Pattern of LULC Using Random Forest Classification in Jhelum District, Punjab, Pakistan
Authors: Basit Aftab, Zhichao Wang, Feng Zhongke
Abstract:
Land Use and Land Cover Change (LULCC) is a critical environmental issue that has significant effects on biodiversity, ecosystem services, and climate change. This study examines the spatiotemporal dynamics of land use and land cover (LULC) across a three-decade period (1992–2022) in a district area. The goal is to support sustainable land management and urban planning by utilizing the combination of remote sensing, GIS data, and observations from Landsat satellites 5 and 8 to provide precise predictions of the trajectory of urban sprawl. In order to forecast the LULCC patterns, this study suggests a hybrid strategy that combines the Random Forest method with Multilayer Perceptron (MLP) and Markov Chain analysis. To predict the dynamics of LULC change for the year 2035, a hybrid technique based on multilayer Perceptron and Markov Chain Model Analysis (MLP-MCA) was employed. The area of developed land has increased significantly, while the amount of bare land, vegetation, and forest cover have all decreased. This is because the principal land types have changed due to population growth and economic expansion. The study also discovered that between 1998 and 2023, the built-up area increased by 468 km² as a result of the replacement of natural resources. It is estimated that 25.04% of the study area's urbanization will be increased by 2035. The performance of the model was confirmed with an overall accuracy of 90% and a kappa coefficient of around 0.89. It is important to use advanced predictive models to guide sustainable urban development strategies. It provides valuable insights for policymakers, land managers, and researchers to support sustainable land use planning, conservation efforts, and climate change mitigation strategies.Keywords: land use land cover, Markov chain model, multi-layer perceptron, random forest, sustainable land, remote sensing.
Procedia PDF Downloads 323707 Optimization Based Extreme Learning Machine for Watermarking of an Image in DWT Domain
Authors: RAM PAL SINGH, VIKASH CHAUDHARY, MONIKA VERMA
Abstract:
In this paper, we proposed the implementation of optimization based Extreme Learning Machine (ELM) for watermarking of B-channel of color image in discrete wavelet transform (DWT) domain. ELM, a regularization algorithm, works based on generalized single-hidden-layer feed-forward neural networks (SLFNs). However, hidden layer parameters, generally called feature mapping in context of ELM need not to be tuned every time. This paper shows the embedding and extraction processes of watermark with the help of ELM and results are compared with already used machine learning models for watermarking.Here, a cover image is divide into suitable numbers of non-overlapping blocks of required size and DWT is applied to each block to be transformed in low frequency sub-band domain. Basically, ELM gives a unified leaning platform with a feature mapping, that is, mapping between hidden layer and output layer of SLFNs, is tried for watermark embedding and extraction purpose in a cover image. Although ELM has widespread application right from binary classification, multiclass classification to regression and function estimation etc. Unlike SVM based algorithm which achieve suboptimal solution with high computational complexity, ELM can provide better generalization performance results with very small complexity. Efficacy of optimization method based ELM algorithm is measured by using quantitative and qualitative parameters on a watermarked image even though image is subjected to different types of geometrical and conventional attacks.Keywords: BER, DWT, extreme leaning machine (ELM), PSNR
Procedia PDF Downloads 3103706 Regular or Irregular: An Investigation of Medicine Consumption Pattern with Poisson Mixture Model
Authors: Lichung Jen, Yi Chun Liu, Kuan-Wei Lee
Abstract:
Fruitful data has been accumulated in database nowadays and is commonly used as support for decision-making. In the healthcare industry, hospital, for instance, ordering pharmacy inventory is one of the key decision. With large drug inventory, the current cost increases and its expiration dates might lead to future issue, such as drug disposal and recycle. In contrast, underestimating demand of the pharmacy inventory, particularly standing drugs, affects the medical treatment and possibly hospital reputation. Prescription behaviour of hospital physicians is one of the critical factor influencing this decision, particularly irregular prescription behaviour. If a drug’s usage amount in the month is irregular and less than the regular usage, it may cause the trend of subsequent stockpiling. On the contrary, if a drug has been prescribed often than expected, it may result in insufficient inventory. We proposed a hierarchical Bayesian mixture model with two components to identify physicians’ regular/irregular prescription patterns with probabilities. Heterogeneity of hospital is considered in our proposed hierarchical Bayes model. The result suggested that modeling the prescription patterns of physician is beneficial for estimating the order quantity of medication and pharmacy inventory management of the hospital. Managerial implication and future research are discussed.Keywords: hierarchical Bayesian model, poission mixture model, medicines prescription behavior, irregular behavior
Procedia PDF Downloads 1273705 Information Technology Impacts on the Supply Chain Performance: Case Study Approach
Authors: Kajal Zarei
Abstract:
Supply chain management is becoming an increasingly important issue in many businesses today. In such circumstances, a number of reasons such as management deficiency in different segments of the supply chain, lack of streamlined processes, resistance to change the current systems and technologies, and lack of advanced information system have paved the ground to ask for innovative research studies. To this end, information technology (IT) is becoming a major driver to overcome the supply chain limitations and deficiencies. The emergence of IT has provided an excellent opportunity for redefining the supply chain to be more effective and competitive. This paper has investigated the IT impact on two-digit industry codes in the International Standard Industrial Classification (ISIC) that are operating in four groups of the supply chains. Firstly, the primary fields of the supply chain were investigated, and then paired comparisons of different industry parts were accomplished. Using experts' ideas and Analytical Hierarchy Process (AHP), the status of industrial activities in Kurdistan Province in Iran was determined. The results revealed that manufacturing and inventory fields have been more important compared to other fields of the supply chain. In addition, IT has had greater impact on food and beverage industry, chemical industry, wood industry, wood products, and production of basic metals. The results indicated the need to IT awareness in supply chain management; in other words, IT applications needed to be developed for the identified industries.Keywords: supply chain, information technology, analytical hierarchy process, two-digit codes, international standard industrial classification
Procedia PDF Downloads 2793704 The Effect of Interpersonal Relationships on Eating Patterns and Physical Activity among Asian-American and European-American Adolescents
Authors: Jamil Lane, Jason Freeman
Abstract:
Background: The role of interpersonal relationships is vital predictors of adolescents’ eating habits, exercise activity, and health problems including obesity. The effect of interpersonal relationships (i.e. family, friends, and intimate partners) on individual health behaviors and development have gained considerable attention during the past 10 years. Teenagers eating habits and exercise activities are established through a dynamic course involving internal and external factors such as food preferences, body weight perception, and parental and peer influence. When conceptualizing one’s interpersonal relationships, it is important to understand that how one relates to others is shaped by their culture. East-Asian culture has been characterized as collectivistic, which describes the significant role intergroup relationships play in their construction of the self. Cultures found in North America, on the other hand, can be characterized as individualistic, meaning that these cultures encourage individuals to prioritize their interest over the needs and want of their compatriots. Individuals from collectivistic cultures typically have stronger boundaries between in-group and out-group membership, whereas those from individualistic cultures see themselves as distinct and separate from strangers as well as family or friends. Objective: The purpose of this study is to examine the effect of collectivism and individualism on interpersonal relationships that shapes eating patterns and physical activity among Asian-American and European-American adolescents. Design/Methods: Analyses were based on data from the National Longitudinal Study of Adolescent Health, a nationally representative sample of adolescents in the United States who were surveyed from 1994 through 2008. This data will be used to examine interpersonal relationship factors that shape dietary intake and physical activity patterns within the Asian-American and European-American population in the United States. Factors relating to relationship strength, eating, and exercise behaviors were reported by participants in this first wave of data collection (1995). We plan to analyze our data using intragroup comparisons among those who identified as 'Asian-American' (n = 270) and 'White or European American' (n = 4,294) among the domains of positivity of peer influence and level of physical activity / healthy eating. Further, intergroup comparisons of these relationships will be made to extricate how the role positive peer influence in maintaining healthy eating and exercise habits differs with cultural variation. Results: We hypothesize that East-Asian participants with a higher degree of positivity in their peer and family relationships will experience a significantly greater rise in healthy eating and exercise behaviors than European-American participants with similar degrees of relationship positivity.Keywords: interpersonal relationships, eating patterns, physical activity, adolescent health
Procedia PDF Downloads 1963703 Merging Appeal to Ignorance, Composition, and Division Argument Schemes with Bayesian Networks
Authors: Kong Ngai Pei
Abstract:
The argument scheme approach to argumentation has two components. One is to identify the recurrent patterns of inferences used in everyday discourse. The second is to devise critical questions to evaluate the inferences in these patterns. Although this approach is intuitive and contains many insightful ideas, it has been noted to be not free of problems. One is that due to its disavowing the probability calculus, it cannot give the exact strength of an inference. In order to tackle this problem, thereby paving the way to a more complete normative account of argument strength, it has been proposed, the most promising way is to combine the scheme-based approach with Bayesian networks (BNs). This paper pursues this line of thought, attempting to combine three common schemes, Appeal to Ignorance, Composition, and Division, with BNs. In the first part, it is argued that most (if not all) formulations of the critical questions corresponding to these schemes in the current argumentation literature are incomplete and not very informative. To remedy these flaws, more thorough and precise formulations of these questions are provided. In the second part, how to use graphical idioms (e.g. measurement and synthesis idioms) to translate the schemes as well as their corresponding critical questions to graphical structure of BNs, and how to define probability tables of the nodes using functions of various sorts are shown. In the final part, it is argued that many misuses of these schemes, traditionally called fallacies with the same names as the schemes, can indeed be adequately accounted for by the BN models proposed in this paper.Keywords: appeal to ignorance, argument schemes, Bayesian networks, composition, division
Procedia PDF Downloads 2853702 Computational Linguistic Implications of Gender Bias: Machines Reflect Misogyny in Society
Authors: Irene Yi
Abstract:
Machine learning, natural language processing, and neural network models of language are becoming more and more prevalent in the fields of technology and linguistics today. Training data for machines are at best, large corpora of human literature and at worst, a reflection of the ugliness in society. Computational linguistics is a growing field dealing with such issues of data collection for technological development. Machines have been trained on millions of human books, only to find that in the course of human history, derogatory and sexist adjectives are used significantly more frequently when describing females in history and literature than when describing males. This is extremely problematic, both as training data, and as the outcome of natural language processing. As machines start to handle more responsibilities, it is crucial to ensure that they do not take with them historical sexist and misogynistic notions. This paper gathers data and algorithms from neural network models of language having to deal with syntax, semantics, sociolinguistics, and text classification. Computational analysis on such linguistic data is used to find patterns of misogyny. Results are significant in showing the existing intentional and unintentional misogynistic notions used to train machines, as well as in developing better technologies that take into account the semantics and syntax of text to be more mindful and reflect gender equality. Further, this paper deals with the idea of non-binary gender pronouns and how machines can process these pronouns correctly, given its semantic and syntactic context. This paper also delves into the implications of gendered grammar and its effect, cross-linguistically, on natural language processing. Languages such as French or Spanish not only have rigid gendered grammar rules, but also historically patriarchal societies. The progression of society comes hand in hand with not only its language, but how machines process those natural languages. These ideas are all extremely vital to the development of natural language models in technology, and they must be taken into account immediately.Keywords: computational analysis, gendered grammar, misogynistic language, neural networks
Procedia PDF Downloads 1193701 A Novel Approach for the Analysis of Ground Water Quality by Using Classification Rules and Water Quality Index
Authors: Kamakshaiah Kolli, R. Seshadri
Abstract:
Water is a key resource in all economic activities ranging from agriculture to industry. Only a tiny fraction of the planet's abundant water is available to us as fresh water. Assessment of water quality has always been paramount in the field of environmental quality management. It is the foundation for health, hygiene, progress and prosperity. With ever increasing pressure of human population, there is severe stress on water resources. Therefore efficient water management is essential to civil society for betterment of quality of life. The present study emphasizes on the groundwater quality, sources of ground water contamination, variation of groundwater quality and its spatial distribution. The bases for groundwater quality assessment are groundwater bodies and representative monitoring network enabling determination of chemical status of groundwater body. For this study, water samples were collected from various areas of the entire corporation area of Guntur. Water is required for all living organisms of which 1.7% is available as ground water. Water has no calories or any nutrients, but essential for various metabolic activities in our body. Chemical and physical parameters can be tested for identifying the portability of ground water. Electrical conductivity, pH, alkalinity, Total Alkalinity, TDS, Calcium, Magnesium, Sodium, Potassium, Chloride, and Sulphate of the ground water from Guntur district: Different areas of the District were analyzed. Our aim is to check, if the ground water from the above areas are potable or not. As multivariate are present, Data mining technique using JRIP rules was employed for classifying the ground water.Keywords: groundwater, water quality standards, potability, data mining, JRIP, PCA, classification
Procedia PDF Downloads 4283700 An ANOVA-based Sequential Forward Channel Selection Framework for Brain-Computer Interface Application based on EEG Signals Driven by Motor Imagery
Authors: Forouzan Salehi Fergeni
Abstract:
Converting the movement intents of a person into commands for action employing brain signals like electroencephalogram signals is a brain-computer interface (BCI) system. When left or right-hand motions are imagined, different patterns of brain activity appear, which can be employed as BCI signals for control. To make better the brain-computer interface (BCI) structures, effective and accurate techniques for increasing the classifying precision of motor imagery (MI) based on electroencephalography (EEG) are greatly needed. Subject dependency and non-stationary are two features of EEG signals. So, EEG signals must be effectively processed before being used in BCI applications. In the present study, after applying an 8 to 30 band-pass filter, a car spatial filter is rendered for the purpose of denoising, and then, a method of analysis of variance is used to select more appropriate and informative channels from a category of a large number of different channels. After ordering channels based on their efficiencies, a sequential forward channel selection is employed to choose just a few reliable ones. Features from two domains of time and wavelet are extracted and shortlisted with the help of a statistical technique, namely the t-test. Finally, the selected features are classified with different machine learning and neural network classifiers being k-nearest neighbor, Probabilistic neural network, support-vector-machine, Extreme learning machine, decision tree, Multi-layer perceptron, and linear discriminant analysis with the purpose of comparing their performance in this application. Utilizing a ten-fold cross-validation approach, tests are performed on a motor imagery dataset found in the BCI competition III. Outcomes demonstrated that the SVM classifier got the greatest classification precision of 97% when compared to the other available approaches. The entire investigative findings confirm that the suggested framework is reliable and computationally effective for the construction of BCI systems and surpasses the existing methods.Keywords: brain-computer interface, channel selection, motor imagery, support-vector-machine
Procedia PDF Downloads 473699 The Relationship Between Walking and Sleep Quality Among Taiwanese High School Students
Authors: Lu Ruei Tsen
Abstract:
Among Taiwanese high school students today, as academic stress increases during adolescence, it has become a major factor contributing to poor sleep, resulting in adverse impacts on mental health and academic performance. This study investigates the relationship between walking and sleep quality among Taiwanese high school students by utilizing Apple Watches for data collection. Addressing concerns over adolescents' sleep patterns due to academic stress and digital distractions, this research fills a gap in understanding the specific demographic within the Taiwanese context. Employing a quantitative approach, data were collected from 23 participants aged 15 to 18, focusing on their walking habits tracked by Apple Watches and sleep quality measured by the Pittsburgh Sleep Quality Index (PSQI). The findings suggest a positive correlation between walking and sleep quality, particularly among females. However, unexpected results, such as disparities in sleep quality among different age groups, highlight the complexity of factors influencing sleep patterns. While limitations exist, including potential confounding variables and sample size, this study provides valuable insights for future research. Recommendations for further research include exploring gender differences and conducting longitudinal studies across diverse demographics. Overall, this research indicates that encouraging adolescents to be more physically active, like walking, can enhance sleep quality.Keywords: sleep quality, PSQI, sleep among adolescents, wearable devices
Procedia PDF Downloads 243698 Algorithm for Modelling Land Surface Temperature and Land Cover Classification and Their Interaction
Authors: Jigg Pelayo, Ricardo Villar, Einstine Opiso
Abstract:
The rampant and unintended spread of urban areas resulted in increasing artificial component features in the land cover types of the countryside and bringing forth the urban heat island (UHI). This paved the way to wide range of negative influences on the human health and environment which commonly relates to air pollution, drought, higher energy demand, and water shortage. Land cover type also plays a relevant role in the process of understanding the interaction between ground surfaces with the local temperature. At the moment, the depiction of the land surface temperature (LST) at city/municipality scale particularly in certain areas of Misamis Oriental, Philippines is inadequate as support to efficient mitigations and adaptations of the surface urban heat island (SUHI). Thus, this study purposely attempts to provide application on the Landsat 8 satellite data and low density Light Detection and Ranging (LiDAR) products in mapping out quality automated LST model and crop-level land cover classification in a local scale, through theoretical and algorithm based approach utilizing the principle of data analysis subjected to multi-dimensional image object model. The paper also aims to explore the relationship between the derived LST and land cover classification. The results of the presented model showed the ability of comprehensive data analysis and GIS functionalities with the integration of object-based image analysis (OBIA) approach on automating complex maps production processes with considerable efficiency and high accuracy. The findings may potentially lead to expanded investigation of temporal dynamics of land surface UHI. It is worthwhile to note that the environmental significance of these interactions through combined application of remote sensing, geographic information tools, mathematical morphology and data analysis can provide microclimate perception, awareness and improved decision-making for land use planning and characterization at local and neighborhood scale. As a result, it can aid in facilitating problem identification, support mitigations and adaptations more efficiently.Keywords: LiDAR, OBIA, remote sensing, local scale
Procedia PDF Downloads 2813697 Determining the Functionality of Urban Wildlife with Large Megafauna: A Case Study from Chobe District, Northern Botswana
Authors: Tempe S. F. Adams, Michael J. Chase, Tracey L. Rogers, Keith E. A. Leggett
Abstract:
Transfrontier wildlife corridors can be successful conservation tools, connecting protected areas and reducing the impact of habitat fragmentation on mobile species. Urban wildlife corridors have been proposed as a potential mitigation tool to facilitate the passage of elephants through towns without causing conflict with urban communities. However, because such corridors are typically narrow and close to human development, wildlife (particularly large mammals) may be less likely to use them. We used remote-sensor camera traps and global positioning system collars to identify the movement patterns of African elephants Loxondonta africana through narrow, urban corridors in Botswana. The corridors were in three types of human-dominated land-use designations with varying levels of human activity: agricultural, industrial and open space recreational land. We found that elephants used the corridors within all three land-use designations and we identified, using a model selection approach, that season, time of day and rainfall were important factors in determining the presence of elephants in the corridors. Elephants moved more slowly through the narrow corridors compared with their movement patterns through broader, wide-ranging corridors. Our results indicate that urban wildlife corridors are useful for facilitating elephants to pass through urban areas.Keywords: camera traps, conservation, human-elephant coexistence, land management, urban wildlife corridors
Procedia PDF Downloads 3023696 Time and Cost Prediction Models for Language Classification Over a Large Corpus on Spark
Authors: Jairson Barbosa Rodrigues, Paulo Romero Martins Maciel, Germano Crispim Vasconcelos
Abstract:
This paper presents an investigation of the performance impacts regarding the variation of five factors (input data size, node number, cores, memory, and disks) when applying a distributed implementation of Naïve Bayes for text classification of a large Corpus on the Spark big data processing framework. Problem: The algorithm's performance depends on multiple factors, and knowing before-hand the effects of each factor becomes especially critical as hardware is priced by time slice in cloud environments. Objectives: To explain the functional relationship between factors and performance and to develop linear predictor models for time and cost. Methods: the solid statistical principles of Design of Experiments (DoE), particularly the randomized two-level fractional factorial design with replications. This research involved 48 real clusters with different hardware arrangements. The metrics were analyzed using linear models for screening, ranking, and measurement of each factor's impact. Results: Our findings include prediction models and show some non-intuitive results about the small influence of cores and the neutrality of memory and disks on total execution time, and the non-significant impact of data input scale on costs, although notably impacts the execution time.Keywords: big data, design of experiments, distributed machine learning, natural language processing, spark
Procedia PDF Downloads 1183695 Application of GIS Techniques for Analysing Urban Built-Up Growth of Class-I Indian Cities: A Case Study of Surat
Authors: Purba Biswas, Priyanka Dey
Abstract:
Worldwide rapid urbanisation has accelerated city expansion in both developed and developing nations. This unprecedented urbanisation trend due to the increasing population and economic growth has caused challenges for the decision-makers in city planning and urban management. Metropolitan cities, class-I towns, and major urban centres undergo a continuous process of evolution due to interaction between socio-cultural and economic attributes. This constant evolution leads to urban expansion in all directions. Understanding the patterns and dynamics of urban built-up growth is crucial for policymakers, urban planners, and researchers, as it aids in resource management, decision-making, and the development of sustainable strategies to address the complexities associated with rapid urbanisation. Identifying spatio-temporal patterns of urban growth has emerged as a crucial challenge in monitoring and assessing present and future trends in urban development. Analysing urban growth patterns and tracking changes in land use is an important aspect of urban studies. This study analyses spatio-temporal urban transformations and land-use and land cover changes using remote sensing and GIS techniques. Built-up growth analysis has been done for the city of Surat as a case example, using the GIS tools of NDBI and GIS models of the Built-up Urban Density Index and Shannon Entropy Index to identify trends and the geographical direction of transformation from 2005 to 2020. Surat is one of the fastest-growing urban centres in both the state and the nation, ranking as the 4th fastest-growing city globally. This study analyses the dynamics of urban built-up area transformations both zone-wise and geographical direction-wise, in which their trend, rate, and magnitude were calculated for the period of 15 years. This study also highlights the need for analysing and monitoring the urban growth pattern of class-I cities in India using spatio-temporal and quantitative techniques like GIS for improved urban management.Keywords: urban expansion, built-up, geographic information system, remote sensing, Shannon’s entropy
Procedia PDF Downloads 723694 Using Predictive Analytics to Identify First-Year Engineering Students at Risk of Failing
Authors: Beng Yew Low, Cher Liang Cha, Cheng Yong Teoh
Abstract:
Due to a lack of continual assessment or grade related data, identifying first-year engineering students in a polytechnic education at risk of failing is challenging. Our experience over the years tells us that there is no strong correlation between having good entry grades in Mathematics and the Sciences and excelling in hardcore engineering subjects. Hence, identifying students at risk of failure cannot be on the basis of entry grades in Mathematics and the Sciences alone. These factors compound the difficulty of early identification and intervention. This paper describes the development of a predictive analytics model in the early detection of students at risk of failing and evaluates its effectiveness. Data from continual assessments conducted in term one, supplemented by data of student psychological profiles such as interests and study habits, were used. Three classification techniques, namely Logistic Regression, K Nearest Neighbour, and Random Forest, were used in our predictive model. Based on our findings, Random Forest was determined to be the strongest predictor with an Area Under the Curve (AUC) value of 0.994. Correspondingly, the Accuracy, Precision, Recall, and F-Score were also highest among these three classifiers. Using this Random Forest Classification technique, students at risk of failure could be identified at the end of term one. They could then be assigned to a Learning Support Programme at the beginning of term two. This paper gathers the results of our findings. It also proposes further improvements that can be made to the model.Keywords: continual assessment, predictive analytics, random forest, student psychological profile
Procedia PDF Downloads 1333693 Development of GIS-Based Geotechnical Guidance Maps for Prediction of Soil Bearing Capacity
Authors: Q. Toufeeq, R. Kauser, U. R. Jamil, N. Sohaib
Abstract:
Foundation design of a structure needs soil investigation to avoid failures due to settlements. This soil investigation is expensive and time-consuming. Developments of new residential societies involve huge leveling of large sites that is accompanied by heavy land filling. Poor practices of land fill for deep depths cause differential settlements and consolidations of underneath soil that sometimes result in the collapse of structures. The extent of filling remains unknown to the individual developer unless soil investigation is carried out. Soil investigation cannot be performed on each available site due to involved costs. However, fair estimate of bearing capacity can be made if such tests are already done in the surrounding areas. The geotechnical guidance maps can provide a fair assessment of soil properties. Previously, GIS-based approaches have been used to develop maps using extrapolation and interpolations techniques for bearing capacities, underground recharge, soil classification, geological hazards, landslide hazards, socio-economic, and soil liquefaction mapping. Standard penetration test (SPT) data of surrounding sites were already available. Google Earth is used for digitization of collected data. Few points were considered for data calibration and validation. Resultant Geographic information system (GIS)-based guidance maps are helpful to anticipate the bearing capacity in the real estate industry.Keywords: bearing capacity, soil classification, geographical information system, inverse distance weighted, radial basis function
Procedia PDF Downloads 1343692 An Automatic Bayesian Classification System for File Format Selection
Authors: Roman Graf, Sergiu Gordea, Heather M. Ryan
Abstract:
This paper presents an approach for the classification of an unstructured format description for identification of file formats. The main contribution of this work is the employment of data mining techniques to support file format selection with just the unstructured text description that comprises the most important format features for a particular organisation. Subsequently, the file format indentification method employs file format classifier and associated configurations to support digital preservation experts with an estimation of required file format. Our goal is to make use of a format specification knowledge base aggregated from a different Web sources in order to select file format for a particular institution. Using the naive Bayes method, the decision support system recommends to an expert, the file format for his institution. The proposed methods facilitate the selection of file format and the quality of a digital preservation process. The presented approach is meant to facilitate decision making for the preservation of digital content in libraries and archives using domain expert knowledge and specifications of file formats. To facilitate decision-making, the aggregated information about the file formats is presented as a file format vocabulary that comprises most common terms that are characteristic for all researched formats. The goal is to suggest a particular file format based on this vocabulary for analysis by an expert. The sample file format calculation and the calculation results including probabilities are presented in the evaluation section.Keywords: data mining, digital libraries, digital preservation, file format
Procedia PDF Downloads 4973691 Influence of Settlements and Human Activities on Beetle Diversity and Assemblage Structure at Small Islands of the Kepulauan Seribu Marine National Park and Nearby Java
Authors: Shinta Holdsworth, Jan Axmacher, Darren J. Mann
Abstract:
Beetles represent the most diverse insect taxon, and they contribute significantly to a wide range of vital ecological functions. Examples include decomposition by bark beetles, nitrogen recycling and dung processing by dung beetles or pest control by predatory ground beetles. Nonetheless, research into the distribution patterns, species richness and functional diversity of beetles particularly from tropical regions remains extremely limited. In our research, we aim to investigate the distribution and diversity patterns of beetles and the roles they play in small tropical island ecosystems in the Kepulauan Seribu Marine National Park and on Java. Our research furthermore provides insights into the effects anthropogenic activities have on the assemblage composition and diversity of beetles on the small islands. We recorded a substantial number of highly abundant small island species, including a substantial number of unique small island species across the study area, highlighting these islands’ potential importance for the regional conservation of genetic resources. The highly varied patterns observed in relation to the use of different trapping types - pitfall traps and flight interception traps (FITs) - underscores the need for complementary trapping strategies that combine multiple methods for beetle community surveys in tropical islands. The significant impacts of human activities have on the small island beetle faunas were also highlighted in our research. More island beetle species encountered in settlement than forest areas shows clear trend of positive links between anthropogenic activities and the overall beetle species richness. However, undisturbed forests harboured a high number of unique species, also in comparison to disturbed forests. Finally, our study suggests that, with regards to different feeding guilds, the diversity of herbivorous beetles on islands is strongly affected by the different levels of forest cover encountered.Keywords: beetle diversity, forest disturbance, island biogeography, island settlement
Procedia PDF Downloads 2183690 Finding the Theory of Riba Avoidance: A Scoping Review to Set the Research Agenda
Authors: Randa Ismail Sharafeddine
Abstract:
The Islamic economic system is distinctive in that it implicitly recognizes money as a separate, independent component of production capable of assuming risk and so entitled to the same reward as other Entrepreneurial Factors of Production (EFP). Conventional theory does not identify money capital explicitly as a component of production; rather, interest is recognized as a reward for capital, the interest rate is the cost of money capital, and it is also seen as a cost of physical capital. The conventional theory of production examines how diverse non-entrepreneurial resources (Land, Labor, and Capital) are selected; however, the economic theory community is largely unaware of the reasons why these resources choose to remain as non-entrepreneurial resources as opposed to becoming entrepreneurial resources. Should land, labor, and financial asset owners choose to work for others in return for rent, income, or interest, or should they engage in entrepreneurial risk-taking in order to profit. This is a decision made often in the actual world, but it has never been effectively treated in economic theory. This article will conduct a critical analysis of the conventional classification of factors of production and propose a classification for resource allocation and income distribution (Rent, Wages, Interest, and Profits) that is more rational, even within the conventional theoretical framework for evaluating and developing production and distribution theories. Money is an essential component of production in an Islamic economy, and it must be used to sustain economic activity.Keywords: financial capital, production theory, distribution theory, economic activity, riba avoidance, institution of participation
Procedia PDF Downloads 903689 Surface Flattening Assisted with 3D Mannequin Based on Minimum Energy
Authors: Shih-Wen Hsiao, Rong-Qi Chen, Chien-Yu Lin
Abstract:
The topic of surface flattening plays a vital role in the field of computer aided design and manufacture. Surface flattening enables the production of 2D patterns and it can be used in design and manufacturing for developing a 3D surface to a 2D platform, especially in fashion design. This study describes surface flattening based on minimum energy methods according to the property of different fabrics. Firstly, through the geometric feature of a 3D surface, the less transformed area can be flattened on a 2D platform by geodesic. Then, strain energy that has accumulated in mesh can be stably released by an approximate implicit method and revised error function. In some cases, cutting mesh to further release the energy is a common way to fix the situation and enhance the accuracy of the surface flattening, and this makes the obtained 2D pattern naturally generate significant cracks. When this methodology is applied to a 3D mannequin constructed with feature lines, it enhances the level of computer-aided fashion design. Besides, when different fabrics are applied to fashion design, it is necessary to revise the shape of a 2D pattern according to the properties of the fabric. With this model, the outline of 2D patterns can be revised by distributing the strain energy with different results according to different fabric properties. Finally, this research uses some common design cases to illustrate and verify the feasibility of this methodology.Keywords: surface flattening, strain energy, minimum energy, approximate implicit method, fashion design
Procedia PDF Downloads 3343688 Spread of Measles Disease in Indonesia with Susceptible Vaccinated Infected Recovered Model
Authors: Septiawan A. Saputro, Purnami Widyaningsih, Sutanto Sastraredja
Abstract:
Measles is a disease which can spread caused by a virus and has been a priority’s Ministry of Health in Indonesia to be solved. Each infected person can be recovered and get immunity so that the spread of the disease can be constructed with susceptible infected recovered (SIR). To prevent the spread of measles transmission, the Ministry of Health holds vaccinations program. The aims of the research are to derive susceptible vaccinated infected recovered (SVIR) model, to determine the patterns of disease spread with SVIR model, and also to apply the SVIR model on the spread of measles in Indonesia. Based on the article, it can be concluded that the spread model of measles with vaccinations, that is SVIR model. It is a first-order differential equation system. The patterns of disease spread is determined by solution of the model. Based on that model Indonesia will be a measles-free nation in 2186 with the average of vaccinations scope about 88% and the average score of vaccinations failure about 4.9%. If it is simulated as Ministry of Health new programs with the average of vaccinations scope about 95% and the average score of vaccinations failure about 3%, then Indonesia will be a measles-free nation in 2184. Even with the average of vaccinations scope about 100% and no failure of vaccinations, Indonesia will be a measles-free nation in 2183. Indonesia’s target as a measles-free nation in 2020 has not been reached.Keywords: measles, vaccination, susceptible infected recovered (SIR), susceptible vaccinated infected recovered (SVIR)
Procedia PDF Downloads 2473687 Q-Methodology to Identify Perceptions of Deceased Organ Donation in the UK
Authors: Reem Muaid, Thomas Chesney
Abstract:
Background: Attitude towards organ donation is predominantly positive in the UK; however, the donation rate remains low. To develop more effective interventions, this research aims to examine the behavioural barriers in organ donations using Q methodology to elicit patterns of overlap among different barriers and motivators. Method: A Q methodology study was conducted with 40 participants aged 19-64 who were asked to rank 47 statements on issues that are associated with organ donation. By-person factor analysis using Centroid method and Varimax rotation was conducted to bring out patterns in the way statements were ranked to obtain groupings of participants who had arranged the statements in similar fashion. Results: Four viewpoints were extracted: The Realist, the Optimist Hesitant, the Pessimist Determinant, and the Empathetic. Salient barriers to organ donation presented in each viewpoint suggest that perceived lack of knowledge, anxiety, mistrust in the healthcare system, and lack of cue to action are the main barriers to organ donation. Consensus statements suggest that religion and family agreement are inconsequential if the attitude to organ donation is well-formed. Conclusion: There are different attitudes around deceased organ donation that were uncovered using Q methodology. These results suggest that people respond to behavioural change campaigns differently depending on their own perceptions of organ donation. We argue that a paradigm shift in behavioural interventions is underpinned by understanding the overlapping yet distinctive nature of perceived perspectives.Keywords: organ donation, Q methodology, behavioural interventions, post Q Survey
Procedia PDF Downloads 893686 Multi-Temporal Urban Land Cover Mapping Using Spectral Indices
Authors: Mst Ilme Faridatul, Bo Wu
Abstract:
Multi-temporal urban land cover mapping is of paramount importance for monitoring urban sprawl and managing the ecological environment. For diversified urban activities, it is challenging to map land covers in a complex urban environment. Spectral indices have proved to be effective for mapping urban land covers. To improve multi-temporal urban land cover classification and mapping, we evaluate the performance of three spectral indices, e.g. modified normalized difference bare-land index (MNDBI), tasseled cap water and vegetation index (TCWVI) and shadow index (ShDI). The MNDBI is developed to evaluate its performance of enhancing urban impervious areas by separating bare lands. A tasseled cap index, TCWVI is developed to evaluate its competence to detect vegetation and water simultaneously. The ShDI is developed to maximize the spectral difference between shadows of skyscrapers and water and enhance water detection. First, this paper presents a comparative analysis of three spectral indices using Landsat Enhanced Thematic Mapper (ETM), Thematic Mapper (TM) and Operational Land Imager (OLI) data. Second, optimized thresholds of the spectral indices are imputed to classify land covers, and finally, their performance of enhancing multi-temporal urban land cover mapping is assessed. The results indicate that the spectral indices are competent to enhance multi-temporal urban land cover mapping and achieves an overall classification accuracy of 93-96%.Keywords: land cover, mapping, multi-temporal, spectral indices
Procedia PDF Downloads 1513685 Optimizing Agricultural Packaging in Fiji: Strategic Barrier Analysis Using Interpretive Structural Modeling and Cross-Impact Matrix Multiplication Applied to Classification
Authors: R. Ananthanarayanan, S. B. Nakula, D. R. Seenivasagam, J. Naua, B. Sharma
Abstract:
Product packaging is a critical component of production, trade, and marketing, playing numerous vital roles that often go unnoticed by consumers. Packaging is essential for maintaining the shelf life, quality assurance, and safety of both manufactured and agricultural products. For example, harvested produce or processed foods can quickly lose quality and freshness, making secure packaging crucial for preservation and safety throughout the food supply chain. In Fiji, agricultural packaging has primarily been managed by local companies for international trade, with gradual advancements in these practices. To further enhance the industry’s performance, this study examines the challenges and constraints hindering the optimization of agricultural packaging practices in Fiji. The study utilizes Multi-Criteria Decision Making (MCDM) tools, specifically Interpretive Structural Modeling (ISM) and Cross-Impact Matrix Multiplication Applied to Classification (MICMAC). ISM analyzes the hierarchical structure of barriers, categorizing them from the least to the most influential, while MICMAC classifies barriers based on their driving and dependence power. This approach helps identify the interrelationships between barriers, providing valuable insights for policymakers and decision-makers to propose innovative solutions for sustainable development in the agricultural packaging sector, ultimately shaping the future of packaging practices in Fiji.Keywords: agricultural packaging, barriers, ISM, MICMAC
Procedia PDF Downloads 263684 Valence and Arousal-Based Sentiment Analysis: A Comparative Study
Authors: Usama Shahid, Muhammad Zunnurain Hussain
Abstract:
This research paper presents a comprehensive analysis of a sentiment analysis approach that employs valence and arousal as its foundational pillars, in comparison to traditional techniques. Sentiment analysis is an indispensable task in natural language processing that involves the extraction of opinions and emotions from textual data. The valence and arousal dimensions, representing the intensity and positivity/negativity of emotions, respectively, enable the creation of four quadrants, each representing a specific emotional state. The study seeks to determine the impact of utilizing these quadrants to identify distinct emotional states on the accuracy and efficiency of sentiment analysis, in comparison to traditional techniques. The results reveal that the valence and arousal-based approach outperforms other approaches, particularly in identifying nuanced emotions that may be missed by conventional methods. The study's findings are crucial for applications such as social media monitoring and market research, where the accurate classification of emotions and opinions is paramount. Overall, this research highlights the potential of using valence and arousal as a framework for sentiment analysis and offers invaluable insights into the benefits of incorporating specific types of emotions into the analysis. These findings have significant implications for researchers and practitioners in the field of natural language processing, as they provide a basis for the development of more accurate and effective sentiment analysis tools.Keywords: sentiment analysis, valence and arousal, emotional states, natural language processing, machine learning, text analysis, sentiment classification, opinion mining
Procedia PDF Downloads 983683 Comprehensive Feature Extraction for Optimized Condition Assessment of Fuel Pumps
Authors: Ugochukwu Ejike Akpudo, Jank-Wook Hur
Abstract:
The increasing demand for improved productivity, maintainability, and reliability has prompted rapidly increasing research studies on the emerging condition-based maintenance concept- Prognostics and health management (PHM). Varieties of fuel pumps serve critical functions in several hydraulic systems; hence, their failure can have daunting effects on productivity, safety, etc. The need for condition monitoring and assessment of these pumps cannot be overemphasized, and this has led to the uproar in research studies on standard feature extraction techniques for optimized condition assessment of fuel pumps. By extracting time-based, frequency-based and the more robust time-frequency based features from these vibrational signals, a more comprehensive feature assessment (and selection) can be achieved for a more accurate and reliable condition assessment of these pumps. With the aid of emerging deep classification and regression algorithms like the locally linear embedding (LLE), we propose a method for comprehensive condition assessment of electromagnetic fuel pumps (EMFPs). Results show that the LLE as a comprehensive feature extraction technique yields better feature fusion/dimensionality reduction results for condition assessment of EMFPs against the use of single features. Also, unlike other feature fusion techniques, its capabilities as a fault classification technique were explored, and the results show an acceptable accuracy level using standard performance metrics for evaluation.Keywords: electromagnetic fuel pumps, comprehensive feature extraction, condition assessment, locally linear embedding, feature fusion
Procedia PDF Downloads 1153682 An Interdisciplinary Maturity Model for Accompanying Sustainable Digital Transformation Processes in a Smart Residential Quarter
Authors: Wesley Preßler, Lucie Schmidt
Abstract:
Digital transformation is playing an increasingly important role in the development of smart residential quarters. In order to accompany and steer this process and ultimately make the success of the transformation efforts measurable, it is helpful to use an appropriate maturity model. However, conventional maturity models for digital transformation focus primarily on the evaluation of processes and neglect the information and power imbalances between the stakeholders, which affects the validity of the results. The Multi-Generation Smart Community (mGeSCo) research project is developing an interdisciplinary maturity model that integrates the dimensions of digital literacy, interpretive patterns, and technology acceptance to address this gap. As part of the mGeSCo project, the technological development of selected dimensions in the Smart Quarter Jena-Lobeda (Germany) is being investigated. A specific maturity model, based on Cohen's Smart Cities Wheel, evaluates the central dimensions Working, Living, Housing and Caring. To improve the reliability and relevance of the maturity assessment, the factors Digital Literacy, Interpretive Patterns and Technology Acceptance are integrated into the developed model. The digital literacy dimension examines stakeholders' skills in using digital technologies, which influence their perception and assessment of technological maturity. Digital literacy is measured by means of surveys, interviews, and participant observation, using the European Commission's Digital Literacy Framework (DigComp) as a basis. Interpretations of digital technologies provide information about how individuals perceive technologies and ascribe meaning to them. However, these are not mere assessments, prejudices, or stereotyped perceptions but collective patterns, rules, attributions of meaning and the cultural repertoire that leads to these opinions and attitudes. Understanding these interpretations helps in assessing the overarching readiness of stakeholders to digitally transform a/their neighborhood. This involves examining people's attitudes, beliefs, and values about technology adoption, as well as their perceptions of the benefits and risks associated with digital tools. These insights provide important data for a holistic view and inform the steps needed to prepare individuals in the neighborhood for a digital transformation. Technology acceptance is another crucial factor for successful digital transformation to examine the willingness of individuals to adopt and use new technologies. Surveys or questionnaires based on Davis' Technology Acceptance Model can be used to complement interpretive patterns to measure neighborhood acceptance of digital technologies. Integrating the dimensions of digital literacy, interpretive patterns and technology acceptance enables the development of a roadmap with clear prerequisites for initiating a digital transformation process in the neighborhood. During the process, maturity is measured at different points in time and compared with changes in the aforementioned dimensions to ensure sustainable transformation. Participation, co-creation, and co-production are essential concepts for a successful and inclusive digital transformation in the neighborhood context. This interdisciplinary maturity model helps to improve the assessment and monitoring of sustainable digital transformation processes in smart residential quarters. It enables a more comprehensive recording of the factors that influence the success of such processes and supports the development of targeted measures to promote digital transformation in the neighborhood context.Keywords: digital transformation, interdisciplinary, maturity model, neighborhood
Procedia PDF Downloads 773681 A Study on Human Musculoskeletal Model for Cycle Fitting: Comparison with EMG
Authors: Yoon- Ho Shin, Jin-Seung Choi, Dong-Won Kang, Jeong-Woo Seo, Joo-Hack Lee, Ju-Young Kim, Dae-Hyeok Kim, Seung-Tae Yang, Gye-Rae Tack
Abstract:
It is difficult to study the effect of various variables on cycle fitting through actual experiment. To overcome such difficulty, the forward dynamics of a musculoskeletal model was applied to cycle fitting in this study. The measured EMG data were compared with the muscle activities of the musculoskeletal model through forward dynamics. EMG data were measured from five cyclists who do not have musculoskeletal diseases during three minutes pedaling with a constant load (150 W) and cadence (90 RPM). The muscles used for the analysis were the Vastus Lateralis (VL), Tibialis Anterior (TA), Bicep Femoris (BF), and Gastrocnemius Medial (GM). Person’s correlation coefficients of the muscle activity patterns, the peak timing of the maximum muscle activities, and the total muscle activities were calculated and compared. BIKE3D model of AnyBody (Anybodytech, Denmark) was used for the musculoskeletal model simulation. The comparisons of the actual experiments with the simulation results showed significant correlations in the muscle activity patterns (VL: 0.789, TA: 0.503, BF: 0.468, GM: 0.670). The peak timings of the maximum muscle activities were distributed at particular phases. The total muscle activities were compared with the normalized muscle activities, and the comparison showed about 10% difference in the VL (+10%), TA (+9.7%), and BF (+10%), excluding the GM (+29.4%). Thus, it can be concluded that muscle activities of model & experiment showed similar results. The results of this study indicated that it was possible to apply the simulation of further improved musculoskeletal model to cycle fitting.Keywords: musculoskeletal modeling, EMG, cycle fitting, simulation
Procedia PDF Downloads 566