Search results for: maturity classification
645 Identification of Hepatocellular Carcinoma Using Supervised Learning Algorithms
Authors: Sagri Sharma
Abstract:
Analysis of diseases integrating multi-factors increases the complexity of the problem and therefore, development of frameworks for the analysis of diseases is an issue that is currently a topic of intense research. Due to the inter-dependence of the various parameters, the use of traditional methodologies has not been very effective. Consequently, newer methodologies are being sought to deal with the problem. Supervised Learning Algorithms are commonly used for performing the prediction on previously unseen data. These algorithms are commonly used for applications in fields ranging from image analysis to protein structure and function prediction and they get trained using a known dataset to come up with a predictor model that generates reasonable predictions for the response to new data. Gene expression profiles generated by DNA analysis experiments can be quite complex since these experiments can involve hypotheses involving entire genomes. The application of well-known machine learning algorithm - Support Vector Machine - to analyze the expression levels of thousands of genes simultaneously in a timely, automated and cost effective way is thus used. The objectives to undertake the presented work are development of a methodology to identify genes relevant to Hepatocellular Carcinoma (HCC) from gene expression dataset utilizing supervised learning algorithms and statistical evaluations along with development of a predictive framework that can perform classification tasks on new, unseen data.Keywords: artificial intelligence, biomarker, gene expression datasets, hepatocellular carcinoma, machine learning, supervised learning algorithms, support vector machine
Procedia PDF Downloads 429644 Utility of Range of Motion Measurements on Classification of Athletes
Authors: Dhiraj Dolai, Rupayan Bhattacharya
Abstract:
In this study, a comparison of Range Of Motion (ROM) of middle and long-distance runners and swimmers has been made. The mobility of the various joints is essential for the quick movement of any sportsman. Knowledge of a ROM helps in preventing injuries, in repeating the movement, and in generating speed and power. ROM varies among individuals, and it is influenced by factors such as gender, age, and whether the motion is performed actively or passively. ROM for running and swimming, both performed with due consideration on speed, plays an important role. The time of generation of speed and mobility of the particular joints are very important for both kinds of athletes. The difficulties that happen during running and swimming in the direction of motion is changed. In this study, data were collected for a total of 102 subjects divided into three groups: control group (22), middle and long-distance runners (40), and swimmers (40), and their ages are between 12 to 18 years. The swimmers have higher ROM in shoulder joint flexion, extension, abduction, and adduction movement. Middle and long-distance runners have significantly greater ROM from Control Group in the left shoulder joint flexion with a 5.82 mean difference. Swimmers have significantly higher ROM from the Control Group in the left shoulder joint flexion with 24.84 mean difference and swimmers have significantly higher ROM from the Middle and Long distance runners in left shoulder flexion with 19.02 mean difference. The picture will be clear after a more detailed investigation.Keywords: range of motion, runners, swimmers, significance
Procedia PDF Downloads 129643 [Keynote Talk]: Determination of Metal Content in the Surface Sediments of the Istanbul Bosphorus Strait
Authors: Durata Haciu, Elif Sena Tekin, Gokce Ozturk, Patricia Ramey Balcı
Abstract:
Coastal zones are under increasing threat due to anthropogenic activities that introduce considerable pollutants such as heavy metals into marine ecosystems. As part of a larger experimental study examining species responses to contaminated marine sediments, surface sediments (top 5cm) were analysed for major trace elements at three locations in Istanbul Straight. Samples were randomly collected by divers (May 2018) using hand-corers from Istinye (n=4), Garipce (n=10) and Poyrazköy (n=6), at water depths of 4-8m. Twelve metals were examined: As, arsenic; Pb, lead; Cd, cadmium; Cr, chromium; Cu, Copper; Fe, Iron; Ni, Nickel; Zn, Zinc; V, vanadium; Mn, Manganese; Ba, Barium; and Ag, silver by wavelength-dispersive X-ray fluorescence spectrometry (WDXRF) and Inductively Coupled Plasma/Mass Spectroscopy (ICP/MS). Preliminary results indicate that the average concentrations of metals (mg kg⁻¹) varied considerably among locations. In general, concentrations were relatively lower at Garipce compared to either Istinye or Poyrazköy. For most metals mean concentrations were highest at Poyrazköy and Ag and Cd were below detection limits (exception= Ag in a few samples). While Cd and As were undetected in all stations, the concentrations of Fe and Ni fall in the criteria of moderately polluted range and the rest of the metals in the range of low polluted range as compared to Effects Range Low (ERL) and Effects Range median (ERM) values determined by US Environmental Protection Agency (EPA).Keywords: effect-range classification, ICP/MS, marine sediments, XRF
Procedia PDF Downloads 131642 Predictive Maintenance: Machine Condition Real-Time Monitoring and Failure Prediction
Authors: Yan Zhang
Abstract:
Predictive maintenance is a technique to predict when an in-service machine will fail so that maintenance can be planned in advance. Analytics-driven predictive maintenance is gaining increasing attention in many industries such as manufacturing, utilities, aerospace, etc., along with the emerging demand of Internet of Things (IoT) applications and the maturity of technologies that support Big Data storage and processing. This study aims to build an end-to-end analytics solution that includes both real-time machine condition monitoring and machine learning based predictive analytics capabilities. The goal is to showcase a general predictive maintenance solution architecture, which suggests how the data generated from field machines can be collected, transmitted, stored, and analyzed. We use a publicly available aircraft engine run-to-failure dataset to illustrate the streaming analytics component and the batch failure prediction component. We outline the contributions of this study from four aspects. First, we compare the predictive maintenance problems from the view of the traditional reliability centered maintenance field, and from the view of the IoT applications. When evolving to the IoT era, predictive maintenance has shifted its focus from ensuring reliable machine operations to improve production/maintenance efficiency via any maintenance related tasks. It covers a variety of topics, including but not limited to: failure prediction, fault forecasting, failure detection and diagnosis, and recommendation of maintenance actions after failure. Second, we review the state-of-art technologies that enable a machine/device to transmit data all the way through the Cloud for storage and advanced analytics. These technologies vary drastically mainly based on the power source and functionality of the devices. For example, a consumer machine such as an elevator uses completely different data transmission protocols comparing to the sensor units in an environmental sensor network. The former may transfer data into the Cloud via WiFi directly. The latter usually uses radio communication inherent the network, and the data is stored in a staging data node before it can be transmitted into the Cloud when necessary. Third, we illustrate show to formulate a machine learning problem to predict machine fault/failures. By showing a step-by-step process of data labeling, feature engineering, model construction and evaluation, we share following experiences: (1) what are the specific data quality issues that have crucial impact on predictive maintenance use cases; (2) how to train and evaluate a model when training data contains inter-dependent records. Four, we review the tools available to build such a data pipeline that digests the data and produce insights. We show the tools we use including data injection, streaming data processing, machine learning model training, and the tool that coordinates/schedules different jobs. In addition, we show the visualization tool that creates rich data visualizations for both real-time insights and prediction results. To conclude, there are two key takeaways from this study. (1) It summarizes the landscape and challenges of predictive maintenance applications. (2) It takes an example in aerospace with publicly available data to illustrate each component in the proposed data pipeline and showcases how the solution can be deployed as a live demo.Keywords: Internet of Things, machine learning, predictive maintenance, streaming data
Procedia PDF Downloads 386641 The Factors Affecting the Use of Massive Open Online Courses in Blended Learning by Lecturers in Universities
Authors: Taghreed Alghamdi, Wendy Hall, David Millard
Abstract:
Massive Open Online Courses (MOOCs) have recently gained widespread interest in the academic world, starting a wide range of discussion of a number of issues. One of these issues, using MOOCs in teaching and learning in the higher education by integrating MOOCs’ contents with traditional face-to-face activities in blended learning format, is called blended MOOCs (bMOOCs) and is intended not to replace traditional learning but to enhance students learning. Most research on MOOCs has focused on students’ perception and institutional threats whereas there is a lack of published research on academics’ experiences and practices. Thus, the first aim of the study is to develop a classification of blended MOOCs models by conducting a systematic literature review, classifying 19 different case studies, and identifying the broad types of bMOOCs models namely: Supplementary Model and Integrated Model. Thus, the analyses phase will emphasize on these different types of bMOOCs models in terms of adopting MOOCs by lecturers. The second aim of the study is to improve the understanding of lecturers’ acceptance of bMOOCs by investigate the factors that influence academics’ acceptance of using MOOCs in traditional learning by distributing an online survey to lecturers who participate in MOOCs platforms. These factors can help institutions to encourage their lecturers to integrate MOOCs with their traditional courses in universities.Keywords: acceptance, blended learning, blended MOOCs, higher education, lecturers, MOOCs, professors
Procedia PDF Downloads 131640 Comparative Study of Ni Catalysts Supported by Silica and Modified by Metal Additions Co and Ce for The Steam Reforming of Methane
Authors: Ali Zazi, Ouiza Cherifi
Abstract:
The Catalysts materials Ni-SiO₂, Ni-Co-SiO₂ and Ni-Ce-SiO₂ were synthetized by classical method impregnation and supported by silica. This involves combing the silica with an adequate rate of the solution of nickel nitrates, or nickel nitrate and cobalt nitrate, or nickel nitrate and cerium nitrate, mixed, dried and calcined at 700 ° c. These catalysts have been characterized by different physicochemical analysis techniques. The atomic absorption spectrometry indicates that the real contents of nickel, cerium and cobalt are close to the theoretical contents previously assumed, which let's say that the nitrate solutions have impregnated well the silica support. The BET results show that the surface area of the specific surfaces decreases slightly after impregnation with nickel nitrates or Co and Ce metals and a further slight decrease after the reaction. This is likely due to coke deposition. X-ray diffraction shows the presence of the different SiO₂ and NiO phases for all catalysts—theCoO phase for that promoted by Co and the Ce₂O₂ phase for that promoted by Ce. The methane steam reforming reaction was carried out on a quartz reactor in a fixed bed. Reactants and products of the reaction were analyzed by a gas chromatograph. This study shows that the metal addition of Cerium or Cobalt improves the majority of the catalytic performance of Ni for the steam reforming reaction of methane. And we conclude the classification of our Catalysts in order of decreasing activity and catalytic performances as follows: Ni-Ce / SiO₂ >Ni-Co / SiO₂> Ni / SiO₂ .Keywords: cerium, cobalt, heterogeneous catalysis, hydrogen, methane, steam reforming, synthesis gas
Procedia PDF Downloads 192639 Understanding Cognitive Fatigue From FMRI Scans With Self-supervised Learning
Authors: Ashish Jaiswal, Ashwin Ramesh Babu, Mohammad Zaki Zadeh, Fillia Makedon, Glenn Wylie
Abstract:
Functional magnetic resonance imaging (fMRI) is a neuroimaging technique that records neural activations in the brain by capturing the blood oxygen level in different regions based on the task performed by a subject. Given fMRI data, the problem of predicting the state of cognitive fatigue in a person has not been investigated to its full extent. This paper proposes tackling this issue as a multi-class classification problem by dividing the state of cognitive fatigue into six different levels, ranging from no-fatigue to extreme fatigue conditions. We built a spatio-temporal model that uses convolutional neural networks (CNN) for spatial feature extraction and a long short-term memory (LSTM) network for temporal modeling of 4D fMRI scans. We also applied a self-supervised method called MoCo (Momentum Contrast) to pre-train our model on a public dataset BOLD5000 and fine-tuned it on our labeled dataset to predict cognitive fatigue. Our novel dataset contains fMRI scans from Traumatic Brain Injury (TBI) patients and healthy controls (HCs) while performing a series of N-back cognitive tasks. This method establishes a state-of-the-art technique to analyze cognitive fatigue from fMRI data and beats previous approaches to solve this problem.Keywords: fMRI, brain imaging, deep learning, self-supervised learning, contrastive learning, cognitive fatigue
Procedia PDF Downloads 189638 Evaluation of Groundwater Quality in North-West Region of Punjab, India
Authors: Jeevan Jyoti Mohindroo, Umesh Kumar Garg
Abstract:
The district of Tarntaran is located25 km south of Amritsar city in Punjab State of Northwestern India. It is 5059 Sq. Km in area. It is surrounded by Amritsar in the North, Kapurthala in the East, and Ferozepur in the South and Pakistan in the West. Patti Town is a municipal council of the Tarntaran district of the Indian state of Punjab, located 45 km from Amritsar its geographical coordinates are 310 16' 51" north to 740 51' 25" East Longitude. The town spreads over an area of 50sq. Km. Moisture content is very less in the air, falling within the semiarid region and frequently facing water scarcity as well as water quality problems. The major sources of employment are agriculture, horticulture and animal husbandry engaging almost 80% of the workforce. Water samples are collected from 400 locations in 20 villages on the Patti –Khem Karan highway with 20 samples from each village, and were subjected to analysis of chemical characteristics. The type of water that predominates in the study area is Ca-Mg-HCO3 type, based on hydro-chemical analysis. Besides, suitability of water for irrigation is evaluated based on the sodium adsorption ratio (SAR), residual sodium carbonate, sodium percent and salinity hazard. Other Physico-chemical parameters such as pH, TDS, conductance, etc. were also determined using a water analysis kit. Analysis of water samples for heavy metal analysis was also carried out in the present study.Keywords: groundwater, chemical classification, SAR, RSC, USSL diagram
Procedia PDF Downloads 197637 Developing a Cloud Intelligence-Based Energy Management Architecture Facilitated with Embedded Edge Analytics for Energy Conservation in Demand-Side Management
Authors: Yu-Hsiu Lin, Wen-Chun Lin, Yen-Chang Cheng, Chia-Ju Yeh, Yu-Chuan Chen, Tai-You Li
Abstract:
Demand-Side Management (DSM) has the potential to reduce electricity costs and carbon emission, which are associated with electricity used in the modern society. A home Energy Management System (EMS) commonly used by residential consumers in a down-stream sector of a smart grid to monitor, control, and optimize energy efficiency to domestic appliances is a system of computer-aided functionalities as an energy audit for residential DSM. Implementing fault detection and classification to domestic appliances monitored, controlled, and optimized is one of the most important steps to realize preventive maintenance, such as residential air conditioning and heating preventative maintenance in residential/industrial DSM. In this study, a cloud intelligence-based green EMS that comes up with an Internet of Things (IoT) technology stack for residential DSM is developed. In the EMS, Arduino MEGA Ethernet communication-based smart sockets that module a Real Time Clock chip to keep track of current time as timestamps via Network Time Protocol are designed and implemented for readings of load phenomena reflecting on voltage and current signals sensed. Also, a Network-Attached Storage providing data access to a heterogeneous group of IoT clients via Hypertext Transfer Protocol (HTTP) methods is configured to data stores of parsed sensor readings. Lastly, a desktop computer with a WAMP software bundle (the Microsoft® Windows operating system, Apache HTTP Server, MySQL relational database management system, and PHP programming language) serves as a data science analytics engine for dynamic Web APP/REpresentational State Transfer-ful web service of the residential DSM having globally-Advanced Internet of Artificial Intelligence (AI)/Computational Intelligence. Where, an abstract computing machine, Java Virtual Machine, enables the desktop computer to run Java programs, and a mash-up of Java, R language, and Python is well-suited and -configured for AI in this study. Having the ability of sending real-time push notifications to IoT clients, the desktop computer implements Google-maintained Firebase Cloud Messaging to engage IoT clients across Android/iOS devices and provide mobile notification service to residential/industrial DSM. In this study, in order to realize edge intelligence that edge devices avoiding network latency and much-needed connectivity of Internet connections for Internet of Services can support secure access to data stores and provide immediate analytical and real-time actionable insights at the edge of the network, we upgrade the designed and implemented smart sockets to be embedded AI Arduino ones (called embedded AIduino). With the realization of edge analytics by the proposed embedded AIduino for data analytics, an Arduino Ethernet shield WizNet W5100 having a micro SD card connector is conducted and used. The SD library is included for reading parsed data from and writing parsed data to an SD card. And, an Artificial Neural Network library, ArduinoANN, for Arduino MEGA is imported and used for locally-embedded AI implementation. The embedded AIduino in this study can be developed for further applications in manufacturing industry energy management and sustainable energy management, wherein in sustainable energy management rotating machinery diagnostics works to identify energy loss from gross misalignment and unbalance of rotating machines in power plants as an example.Keywords: demand-side management, edge intelligence, energy management system, fault detection and classification
Procedia PDF Downloads 250636 Mineralogical and Geochemical Characteristics of Serpentinite-Derived Ni-Bearing Laterites from Fars Province, Iran: Implications for the Lateritization Process and Classification of Ni-Laterites
Authors: S. Rasti, M. A. Rajabzadeh
Abstract:
Nickel-bearing laterites occur as two parallel belts along Sedimentary Zagros Orogenic (SZO) and Metamorphic Sanandaj-Sirjan (MSS) petrostructural zones, Fars Province, south Iran. An undisturbed vertical profile of these laterites includes protolith, saprolite, clay, and oxide horizons from base to top. Highly serpentinized harzburgite with relicts of olivine and orthopyroxene is regarded as the source rock. The laterites are unusual in lacking a significant saprolite zone with little development of Ni-silicates. Hematite, saponite, dolomite, smectite and clinochlore increase, while calcite, olivine, lizardite and chrysotile decrease from saprolite to oxide zones. Smectite and clinochlore with minor calcite are the major minerals in clay zone. Contacts of different horizons in laterite profiles are gradual and characterized by a decrease in Mg concentration ranging from 18.1 to 9.3 wt.% in oxide and saprolite, respectively. The maximum Ni concentration is 0.34 wt.% (NiO) in the base of the oxide zone, and goethite is the major Ni-bearing phase. From saprolite to oxide horizons, Al2O3, K2O, TiO2, and CaO decrease, while SiO2, MnO, NiO, and Fe2O3 increase. Silica content reaches up to 45 wt.% in the upper part of the soil profile. There is a decrease in pH (8.44-8.17) and an increase in organic matter (0.28-0.59 wt.%) from base to top of the soils. The studied laterites are classified in the oxide clans which were derived from ophiolite ultramafic rocks under Mediterranean climate conditions.Keywords: Iran, laterite, mineralogy, ophiolite
Procedia PDF Downloads 332635 Identification of Vulnerable Zone Due to Cyclone-Induced Storm Surge in the Exposed Coast of Bangladesh
Authors: Mohiuddin Sakib, Fatin Nihal, Rabeya Akter, Anisul Haque, Munsur Rahman, Wasif-E-Elahi
Abstract:
Surge generating cyclones are one of the deadliest natural disasters that threaten the life of coastal environment and communities worldwide. Due to the geographic location, ‘low lying alluvial plain, geomorphologic characteristics and 710 kilometers exposed coastline, Bangladesh is considered as one of the greatest vulnerable country for storm surge flooding. Bay of Bengal is possessing the highest potential of creating storm surge inundation to the coastal areas. Bangladesh is the most exposed country to tropical cyclone with an average of four cyclone striking every years. Frequent cyclone landfall made the country one of the worst sufferer within the world for cyclone induced storm surge flooding and casualties. During the years from 1797 to 2009 Bangladesh has been hit by 63 severe cyclones with strengths of different magnitudes. Though detailed studies were done focusing on the specific cyclone like Sidr or Aila, no study was conducted where vulnerable areas of exposed coast were identified based on the strength of cyclones. This study classifies the vulnerable areas of the exposed coast based on storm surge inundation depth and area due to cyclones of varying strengths. Classification of the exposed coast based on hazard induced cyclonic vulnerability will help the decision makers to take appropriate policies for reducing damage and loss.Keywords: cyclone, landfall, storm surge, exposed coastline, vulnerability
Procedia PDF Downloads 399634 HIV and AIDS in Kosovo, Stigma Persist!
Authors: Luljeta Gashi, Naser Ramadani, Zana Deva, Dafina Gexha-Bunjaku
Abstract:
The official HIV/AIDS data in Kosovo are based on HIV case reporting from health-care services, the blood transfusion system and Voluntary Counselling and Testing centres. Between 1986 and 2014, are reported 95 HIV and AIDS cases, of which 49 were AIDS, 46 HIV and 40 deaths. The majority (69%) of cases were men, age group 25 to 34 (37%) and route of transmission is: heterosexual (90%), MSM (7%), vertical transmission (2%) and IDU (1%). Based on existing data and the UNAIDS classification system, Kosovo is currently still categorised as having a low-level HIV epidemic. Even though with a low HIV prevalence, Kosovo faces a number of threatening factors, including increased number of drug users, a stigmatized and discriminated MSM community, high percentage of youth among general population (57% of the population under the age of 25), with changing social norms and especially the sexual ones. Methods: Data collection was done using self administered structured questionnaires amongst 249 high school students. Data were analysed using the Statistical Package for Social Sciences (SPSS). Results: The findings revealed that 68% of students know that HIV transmission can be reduced by having sex with only one uninfected partner who has no other partners, 94% know that the risk of getting HIV can be reduced by using a condom every time they have sex, 68% know that a person cannot get HIV from mosquito bites, 81% know that they cannot get HIV by sharing food with someone who is infected and 46% know that a healthy looking person can have HIV. Conclusions: Seventy one percent of high school students correctly identify ways of preventing the sexual transmission of HIV and who reject the major misconceptions about HIV transmission. The findings of the study indicate a need for more health education and promotion.Keywords: Kosovo, KPAR, HIV, high school
Procedia PDF Downloads 477633 Melanoma and Non-Melanoma, Skin Lesion Classification, Using a Deep Learning Model
Authors: Shaira L. Kee, Michael Aaron G. Sy, Myles Joshua T. Tan, Hezerul Abdul Karim, Nouar AlDahoul
Abstract:
Skin diseases are considered the fourth most common disease, with melanoma and non-melanoma skin cancer as the most common type of cancer in Caucasians. The alarming increase in Skin Cancer cases shows an urgent need for further research to improve diagnostic methods, as early diagnosis can significantly improve the 5-year survival rate. Machine Learning algorithms for image pattern analysis in diagnosing skin lesions can dramatically increase the accuracy rate of detection and decrease possible human errors. Several studies have shown the diagnostic performance of computer algorithms outperformed dermatologists. However, existing methods still need improvements to reduce diagnostic errors and generate efficient and accurate results. Our paper proposes an ensemble method to classify dermoscopic images into benign and malignant skin lesions. The experiments were conducted using the International Skin Imaging Collaboration (ISIC) image samples. The dataset contains 3,297 dermoscopic images with benign and malignant categories. The results show improvement in performance with an accuracy of 88% and an F1 score of 87%, outperforming other existing models such as support vector machine (SVM), Residual network (ResNet50), EfficientNetB0, EfficientNetB4, and VGG16.Keywords: deep learning - VGG16 - efficientNet - CNN – ensemble – dermoscopic images - melanoma
Procedia PDF Downloads 81632 Using Autoencoder as Feature Extractor for Malware Detection
Authors: Umm-E-Hani, Faiza Babar, Hanif Durad
Abstract:
Malware-detecting approaches suffer many limitations, due to which all anti-malware solutions have failed to be reliable enough for detecting zero-day malware. Signature-based solutions depend upon the signatures that can be generated only when malware surfaces at least once in the cyber world. Another approach that works by detecting the anomalies caused in the environment can easily be defeated by diligently and intelligently written malware. Solutions that have been trained to observe the behavior for detecting malicious files have failed to cater to the malware capable of detecting the sandboxed or protected environment. Machine learning and deep learning-based approaches greatly suffer in training their models with either an imbalanced dataset or an inadequate number of samples. AI-based anti-malware solutions that have been trained with enough samples targeted a selected feature vector, thus ignoring the input of leftover features in the maliciousness of malware just to cope with the lack of underlying hardware processing power. Our research focuses on producing an anti-malware solution for detecting malicious PE files by circumventing the earlier-mentioned shortcomings. Our proposed framework, which is based on automated feature engineering through autoencoders, trains the model over a fairly large dataset. It focuses on the visual patterns of malware samples to automatically extract the meaningful part of the visual pattern. Our experiment has successfully produced a state-of-the-art accuracy of 99.54 % over test data.Keywords: malware, auto encoders, automated feature engineering, classification
Procedia PDF Downloads 72631 The Identification of Combined Genomic Expressions as a Diagnostic Factor for Oral Squamous Cell Carcinoma
Authors: Ki-Yeo Kim
Abstract:
Trends in genetics are transforming in order to identify differential coexpressions of correlated gene expression rather than the significant individual gene. Moreover, it is known that a combined biomarker pattern improves the discrimination of a specific cancer. The identification of the combined biomarker is also necessary for the early detection of invasive oral squamous cell carcinoma (OSCC). To identify the combined biomarker that could improve the discrimination of OSCC, we explored an appropriate number of genes in a combined gene set in order to attain the highest level of accuracy. After detecting a significant gene set, including the pre-defined number of genes, a combined expression was identified using the weights of genes in a gene set. We used the Principal Component Analysis (PCA) for the weight calculation. In this process, we used three public microarray datasets. One dataset was used for identifying the combined biomarker, and the other two datasets were used for validation. The discrimination accuracy was measured by the out-of-bag (OOB) error. There was no relation between the significance and the discrimination accuracy in each individual gene. The identified gene set included both significant and insignificant genes. One of the most significant gene sets in the classification of normal and OSCC included MMP1, SOCS3 and ACOX1. Furthermore, in the case of oral dysplasia and OSCC discrimination, two combined biomarkers were identified. The combined genomic expression achieved better performance in the discrimination of different conditions than in a single significant gene. Therefore, it could be expected that accurate diagnosis for cancer could be possible with a combined biomarker.Keywords: oral squamous cell carcinoma, combined biomarker, microarray dataset, correlated genes
Procedia PDF Downloads 423630 Synergistic Effect of Platelet-Rich Plasma with Hyaluronic Acid Injection Following Arthrocentesis to Reduce Pain and Improve Function in Temporomandibular joint (TMJ) Osteoarthritis
Authors: Ayman Hegab
Abstract:
Increasing evidence supports the use of platelet-rich plasma (PRP) combined with hyaluronic acid (HA) for the treatment of knee osteoarthritis, which effectively promotes cartilage repair. This study aimed to determine whether injection of PRP+HA following arthrocentesis reduces pain and improves maximum incisal opening. This was a single-blind, prospective, randomized control study. The patients were selected based on the Hegab classification: Group I: patients treated with arthrocentesis followed by a single PRP injection; Group II (Control): patients treated with arthrocentesis followed by a single HA injection; and Group III: patients treated with arthrocentesis followed by a single PRP+HA combination injection. The primary predictor variable was the medication used for injection. The primary outcome variables were the maximum voluntary mouth opening and pain index scores. The secondary outcome variable was joint sounds. All outcome variables were assessed and compared among the three groups at baseline and at 1-, 3-, 6-, and 12-month intervals. Other variables, including patients’ age and sex, were evaluated in relation to the patient outcomes. Injecting PRP+HA showed statistically significant improvement in the primary and secondary treatment outcomes over PRP or HA injection throughout the study period (P<0.005). Injection of PRP+HA following arthrocentesis had significant long-term clinical efficacy regarding pain relief that was considered the main concern of both the patient and clinician.Keywords: TMJ, HA, PRP, osteoarthritis
Procedia PDF Downloads 7629 DCDNet: Lightweight Document Corner Detection Network Based on Attention Mechanism
Authors: Kun Xu, Yuan Xu, Jia Qiao
Abstract:
The document detection plays an important role in optical character recognition and text analysis. Because the traditional detection methods have weak generalization ability, and deep neural network has complex structure and large number of parameters, which cannot be well applied in mobile devices, this paper proposes a lightweight Document Corner Detection Network (DCDNet). DCDNet is a two-stage architecture. The first stage with Encoder-Decoder structure adopts depthwise separable convolution to greatly reduce the network parameters. After introducing the Feature Attention Union (FAU) module, the second stage enhances the feature information of spatial and channel dim and adaptively adjusts the size of receptive field to enhance the feature expression ability of the model. Aiming at solving the problem of the large difference in the number of pixel distribution between corner and non-corner, Weighted Binary Cross Entropy Loss (WBCE Loss) is proposed to define corner detection problem as a classification problem to make the training process more efficient. In order to make up for the lack of Dataset of document corner detection, a Dataset containing 6620 images named Document Corner Detection Dataset (DCDD) is made. Experimental results show that the proposed method can obtain fast, stable and accurate detection results on DCDD.Keywords: document detection, corner detection, attention mechanism, lightweight
Procedia PDF Downloads 354628 Commuters Trip Purpose Decision Tree Based Model of Makurdi Metropolis, Nigeria and Strategic Digital City Project
Authors: Emmanuel Okechukwu Nwafor, Folake Olubunmi Akintayo, Denis Alcides Rezende
Abstract:
Decision tree models are versatile and interpretable machine learning algorithms widely used for both classification and regression tasks, which can be related to cities, whether physical or digital. The aim of this research is to assess how well decision tree algorithms can predict trip purposes in Makurdi, Nigeria, while also exploring their connection to the strategic digital city initiative. The research methodology involves formalizing household demographic and trips information datasets obtained from extensive survey process. Modelling and Prediction were achieved using Python Programming Language and the evaluation metrics like R-squared and mean absolute error were used to assess the decision tree algorithm's performance. The results indicate that the model performed well, with accuracies of 84% and 68%, and low MAE values of 0.188 and 0.314, on training and validation data, respectively. This suggests the model can be relied upon for future prediction. The conclusion reiterates that This model will assist decision-makers, including urban planners, transportation engineers, government officials, and commuters, in making informed decisions on transportation planning and management within the framework of a strategic digital city. Its application will enhance the efficiency, sustainability, and overall quality of transportation services in Makurdi, Nigeria.Keywords: decision tree algorithm, trip purpose, intelligent transport, strategic digital city, travel pattern, sustainable transport
Procedia PDF Downloads 20627 Use of Data of the Remote Sensing for Spatiotemporal Analysis Land Use Changes in the Eastern Aurès (Algeria)
Authors: A. Bouzekri, H. Benmassaud
Abstract:
Aurès region is one of the arid and semi-arid areas that have suffered climate crises and overexploitation of natural resources they have led to significant land degradation. The use of remote sensing data allowed us to analyze the land and its spatiotemporal changes in the Aurès between 1987 and 2013, for this work, we adopted a method of analysis based on the exploitation of the images satellite Landsat TM 1987 and Landsat OLI 2013, from the supervised classification likelihood coupled with field surveys of the mission of May and September of 2013. Using ENVI EX software by the superposition of the ground cover maps from 1987 and 2013, one can extract a spatial map change of different land cover units. The results show that between 1987 and 2013 vegetation has suffered negative changes are the significant degradation of forests and steppe rangelands, and sandy soils and bare land recorded a considerable increase. The spatial change map land cover units between 1987 and 2013 allows us to understand the extensive or regressive orientation of vegetation and soil, this map shows that dense forests give his place to clear forests and steppe vegetation develops from a degraded forest vegetation and bare, sandy soils earn big steppe surfaces that explain its remarkable extension. The analysis of remote sensing data highlights the profound changes in our environment over time and quantitative monitoring of the risk of desertification.Keywords: remote sensing, spatiotemporal, land use, Aurès
Procedia PDF Downloads 335626 Development of an Omaha System-Based Remote Intervention Program for Work-Related Musculoskeletal Disorders (WMSDs) Among Front-Line Nurses
Authors: Tianqiao Zhang, Ye Tian, Yanliang Yin, Yichao Tian, Suzhai Tian, Weige Sun, Shuhui Gong, Limei Tang, Ruoliang Tang
Abstract:
Introduction: Healthcare workers, especially the nurses all over the world, are highly vulnerable to work-related musculoskeletal disorders (WMSDs), experiencing high rates of neck, shoulder, and low back injuries, due to the unfavorable working conditions. To reduce WMSDs among nursing personnel, many workplace interventions have been developed and implemented. Unfortunately, the ongoing Covid-19 (SARS-CoV-2) pandemic has posed great challenges to the ergonomic practices and interventions in healthcare facilities, particularly the hospitals, since current Covid-19 mitigation measures, such as social distancing and working remotely, has substantially minimized in-person gatherings and trainings. On the other hand, hospitals throughout the world have been short-staffed, resulting in disturbance of shift scheduling and more importantly, the increased job demand among the available caregivers, particularly the doctors and nurses. With the latest development in communication technology, remote intervention measures have been developed as an alternative, without the necessity of in-person meetings. The Omaha System (OS) is a standardized classification system for nursing practices, including a problem classification system, an intervention system, and an outcome evaluation system. This paper describes the development of an OS-based ergonomic intervention program. Methods: First, a comprehensive literature search was performed among worldwide electronic databases, including PubMed, Web of Science, Cochrane Library, China National Knowledge Infrastructure (CNKI), between journal inception to May 2020, resulting in a total of 1,418 scientific articles. After two independent screening processes, the final knowledge pool included eleven randomized controlled trial studies to develop the draft of the intervention program with Omaha intervention subsystem as the framework. After the determination of sample size needed for statistical power and the potential loss to follow-up, a total of 94 nurses from eight clinical departments agreed to provide written, informed consent to participate in the study, which were subsequently assigned into two random groups (i.e., intervention vs. control). A subgroup of twelve nurses were randomly selected to participate in a semi-structured interview, during which their general understanding and awareness of musculoskeletal disorders and potential interventions was assessed. Then, the first draft was modified to reflect the findings from these interviews. Meanwhile, the tentative program schedule was also assessed. Next, two rounds of consultation were conducted among experts in nursing management, occupational health, psychology, and rehabilitation, to further adjust and finalize the intervention program. The control group had access to all the information and exercise modules at baseline, while an interdisciplinary research team was formed and supervised the implementation of the on-line intervention program through multiple social media groups. Outcome measures of this comparative study included biomechanical load assessed by the Quick Exposure Check and stresses due to awkward body postures. Results and Discussion: Modification to the draft included (1) supplementing traditional Chinese medicine practices, (2) adding the use of assistive patient handling equipment, and (3) revising the on-line training method. Information module should be once a week, lasting about 20 to 30 minutes, for a total of 6 weeks, while the exercise module should be 5 times a week, each lasting about 15 to 20 minutes, for a total of 6 weeks.Keywords: ergonomic interventions, musculoskeletal disorders (MSDs), omaha system, nurses, Covid-19
Procedia PDF Downloads 182625 Application of 3D Apparel CAD for Costume Reproduction
Authors: Zi Y. Kang, Tracy D. Cassidy, Tom Cassidy
Abstract:
3D apparel CAD is one of the remarkable products in advanced technology which enables intuitive design, visualisation and evaluation of garments through stereoscopic drape simulation. The progressive improvements of 3D apparel CAD have led to the creation of more realistic clothing simulation which is used not only in design development but also in presentation, promotion and communication for fashion as well as other industries such as film, game and social network services. As a result, 3D clothing technology is becoming more ubiquitous in human culture and lives today. This study considers that such phenomenon implies that the technology has reached maturity and it is time to inspect the status of current technology and to explore its potential uses in ways to create cultural values to further move forward. For this reason, this study aims to generate virtual costumes as culturally significant objects using 3D apparel CAD and to assess its capability, applicability and attitudes of the audience towards clothing simulation through comparison with physical counterparts. Since the access to costume collection is often limited due to the conservative issues, the technology may make valuable contribution by democratization of culture and knowledge for museums and its audience. This study is expected to provide foundation knowledge for development of clothing technology and for expanding its boundary of practical uses. To prevent any potential damage, two replicas of the costumes in the 1860s and 1920s at the Museum of London were chosen as samples. Their structural, visual and physical characteristics were measured and collected using patterns, scanned images of fabrics and objective fabric measurements with scale, KES-F (Kawabata Evaluation System of Fabrics) and Titan. Commercial software, DC Suite 5.0 was utilised to create virtual costumes applying collected data and the following outcomes were produced for the evaluation: Images of virtual costumes and video clips showing static and dynamic simulation. Focus groups were arranged with fashion design students and the public for evaluation which exposed the outcomes together with physical samples, fabrics swatches and photographs. The similarities, application and acceptance of virtual costumes were estimated through discussion and a questionnaire. The findings show that the technology has the capability to produce realistic or plausible simulation but expression of some factors such as details and capability of light material requires improvements. While the use of virtual costumes was viewed as more interesting and futuristic replacements to physical objects by the public group, the fashion student group noted more differences in detail and preferred physical garments highlighting the absence of tangibility. However, the advantages and potential of virtual costumes as effective and useful visual references for educational and exhibitory purposes were underlined by both groups. Although 3D apparel CAD has sufficient capacity to assist garment design process, it has limits in identical replication and more study on accurate reproduction of details and drape is needed for its technical improvements. Nevertheless, the virtual costumes in this study demonstrated the possibility of the technology to contribute to cultural and knowledgeable value creation through its applicability and as an interesting way to offer 3D visual information.Keywords: digital clothing technology, garment simulation, 3D Apparel CAD, virtual costume
Procedia PDF Downloads 221624 Comparative Evaluation of Accuracy of Selected Machine Learning Classification Techniques for Diagnosis of Cancer: A Data Mining Approach
Authors: Rajvir Kaur, Jeewani Anupama Ginige
Abstract:
With recent trends in Big Data and advancements in Information and Communication Technologies, the healthcare industry is at the stage of its transition from clinician oriented to technology oriented. Many people around the world die of cancer because the diagnosis of disease was not done at an early stage. Nowadays, the computational methods in the form of Machine Learning (ML) are used to develop automated decision support systems that can diagnose cancer with high confidence in a timely manner. This paper aims to carry out the comparative evaluation of a selected set of ML classifiers on two existing datasets: breast cancer and cervical cancer. The ML classifiers compared in this study are Decision Tree (DT), Support Vector Machine (SVM), k-Nearest Neighbor (k-NN), Logistic Regression, Ensemble (Bagged Tree) and Artificial Neural Networks (ANN). The evaluation is carried out based on standard evaluation metrics Precision (P), Recall (R), F1-score and Accuracy. The experimental results based on the evaluation metrics show that ANN showed the highest-level accuracy (99.4%) when tested with breast cancer dataset. On the other hand, when these ML classifiers are tested with the cervical cancer dataset, Ensemble (Bagged Tree) technique gave better accuracy (93.1%) in comparison to other classifiers.Keywords: artificial neural networks, breast cancer, classifiers, cervical cancer, f-score, machine learning, precision, recall
Procedia PDF Downloads 276623 Dimensional Investigation of Food Addiction in Individuals Who Have Undergone Bariatric Surgery
Authors: Ligia Florio, João Mauricio Castaldelli-Maia
Abstract:
Background: Food addiction (FA) emerged in the 1990s as a possible contributor to the increasing prevalence of obesity and overweight, in conjunction with changing food environments and mental health conditions. However, FA is not yet listed as one of the disorders in the DSM-5 and/or the ICD-11. Although there are controversies and debates in the literature about the classification and construct of FA, the most common approach to access it is the use of a research tool - the Yale Food Addiction Scale (YFAS) - which approximates the concept of FA to the concept diagnosis of dependence on psychoactive substances. There is a need to explore the dimensional phenotypes accessed by YFAS in different population groups for a better understanding and scientific support of FA diagnoses. Methods: The primary objective of this project was to investigate the construct validity of the FA concept by mYFAS 2.0 in individuals who underwent bariatric surgery (n = 100) at the Hospital Estadual Mário Covas since 2011. Statistical analyzes were conducted using the STATA software. In this sense, structural or factor validity was the type of construct validity investigated using exploratory factor analysis (EFA) and item response theory (IRT) techniques. Results: EFA showed that the one-dimensional model was the most parsimonious. The IRT showed that all criteria contributed to the latent structure, presenting discrimination values greater than 0.5, with most presenting values greater than 2. Conclusion: This study reinforces a FA dimension in patients who underwent bariatric surgery. Within this dimension, we identified the most severe and discriminating criteria for the diagnosis of FA.Keywords: obesity, food addiction, bariatric surgery, regain
Procedia PDF Downloads 76622 Analysis of Brain Activities due to Differences in Running Shoe Properties
Authors: Kei Okubo, Yosuke Kurihara, Takashi Kaburagi, Kajiro Watanabe
Abstract:
Many of the ever-growing elderly population require exercise, such as running, for health management. One important element of a runner’s training is the choice of shoes for exercise; shoes are important because they provide the interface between the feet and road. When we purchase shoes, we may instinctively choose a pair after trying on many different pairs of shoes. Selecting the shoes instinctively may work, but it does not guarantee a suitable fit for running activities. Therefore, if we could select suitable shoes for each runner from the viewpoint of brain activities, it would be helpful for validating shoe selection. In this paper, we describe how brain activities show different characteristics during particular task, corresponding to different properties of shoes. Using five subjects, we performed a verification experiment, applying weight, softness, and flexibility as shoe properties. In order to affect the shoe property’s differences to the brain, subjects run for ten min. Before and after running, subjects conducted a paced auditory serial addition task (PASAT) as the particular task; and the subjects’ brain activities during the PASAT are evaluated based on oxyhemoglobin and deoxyhemoglobin relative concentration changes, measured by near-infrared spectroscopy (NIRS). When the brain works actively, oxihemoglobin and deoxyhemoglobin concentration drastically changes; therefore, we calculate the maximum values of concentration changes. In order to normalize relative concentration changes after running, the maximum value are divided by before running maximum value as evaluation parameters. The classification of the groups of shoes is expressed on a self-organizing map (SOM). As a result, deoxyhemoglobin can make clusters for two of the three types of shoes.Keywords: brain activities, NIRS, PASAT, running shoes
Procedia PDF Downloads 373621 Pattern Discovery from Student Feedback: Identifying Factors to Improve Student Emotions in Learning
Authors: Angelina A. Tzacheva, Jaishree Ranganathan
Abstract:
Interest in (STEM) Science Technology Engineering Mathematics education especially Computer Science education has seen a drastic increase across the country. This fuels effort towards recruiting and admitting a diverse population of students. Thus the changing conditions in terms of the student population, diversity and the expected teaching and learning outcomes give the platform for use of Innovative Teaching models and technologies. It is necessary that these methods adapted should also concentrate on raising quality of such innovations and have positive impact on student learning. Light-Weight Team is an Active Learning Pedagogy, which is considered to be low-stake activity and has very little or no direct impact on student grades. Emotion plays a major role in student’s motivation to learning. In this work we use the student feedback data with emotion classification using surveys at a public research institution in the United States. We use Actionable Pattern Discovery method for this purpose. Actionable patterns are patterns that provide suggestions in the form of rules to help the user achieve better outcomes. The proposed method provides meaningful insight in terms of changes that can be incorporated in the Light-Weight team activities, resources utilized in the course. The results suggest how to enhance student emotions to a more positive state, in particular focuses on the emotions ‘Trust’ and ‘Joy’.Keywords: actionable pattern discovery, education, emotion, data mining
Procedia PDF Downloads 98620 Perceiving Interpersonal Conflict and the Big Five Personality Traits
Authors: Emily Rivera, Toni DiDona
Abstract:
The Big Five personality traits is a hierarchical classification of personality traits that applies factor analysis to a personality survey data in order to describe human personality using five broad dimensions: Extraversion, Agreeableness, Conscientiousness, Neuroticism, and Openness (Fetvadjiev & Van de Vijer, 2015). Research shows that personality constructs underline individual differences in processing conflict and interpersonal relations. (Graziano et al., 1996). This research explores the understudied correlation between the Big Five personality traits and perceived interpersonal conflict in the workplace. It revises social psychological literature on Big Five personality traits within a social context and discusses organizational development journal articles on the perceived efficacy of conflict tactics and approach to interpersonal relationships. The study also presents research undertaken on a survey group of 867 subjects over the age of 18 that were recruited by means of convenience sampling through social media, email, and text messaging. The central finding of this study is that only two of the Big Five personality traits had a significant correlation with perceiving interpersonal conflict in the workplace. Individuals who score higher on agreeableness and neuroticism, perceive more interpersonal conflict in the workplace compared to those that score lower on each dimension. The relationship between both constructs is worthy of research due to its everyday frequency and unique individual psycho-social consequences. This multimethod research associated the Big Five personality dimensions to interpersonal conflict. Its findings that can be utilized to further understand social cognition, person perception, complex social behavior and social relationships in the work environment.Keywords: five-factor model, interpersonal conflict, personality, The Big Five personality traits
Procedia PDF Downloads 157619 Sub-Pixel Mapping Based on New Mixed Interpolation
Authors: Zeyu Zhou, Xiaojun Bi
Abstract:
Due to the limited environmental parameters and the limited resolution of the sensor, the universal existence of the mixed pixels in the process of remote sensing images restricts the spatial resolution of the remote sensing images. Sub-pixel mapping technology can effectively improve the spatial resolution. As the bilinear interpolation algorithm inevitably produces the edge blur effect, which leads to the inaccurate sub-pixel mapping results. In order to avoid the edge blur effect that affects the sub-pixel mapping results in the interpolation process, this paper presents a new edge-directed interpolation algorithm which uses the covariance adaptive interpolation algorithm on the edge of the low-resolution image and uses bilinear interpolation algorithm in the low-resolution image smooth area. By using the edge-directed interpolation algorithm, the super-resolution of the image with low resolution is obtained, and we get the percentage of each sub-pixel under a certain type of high-resolution image. Then we rely on the probability value as a soft attribute estimate and carry out sub-pixel scale under the ‘hard classification’. Finally, we get the result of sub-pixel mapping. Through the experiment, we compare the algorithm and the bilinear algorithm given in this paper to the results of the sub-pixel mapping method. It is found that the sub-pixel mapping method based on the edge-directed interpolation algorithm has better edge effect and higher mapping accuracy. The results of the paper meet our original intention of the question. At the same time, the method does not require iterative computation and training of samples, making it easier to implement.Keywords: remote sensing images, sub-pixel mapping, bilinear interpolation, edge-directed interpolation
Procedia PDF Downloads 229618 Evaluating Surface Water Quality Using WQI, Trend Analysis, and Cluster Classification in Kebir Rhumel Basin, Algeria
Authors: Lazhar Belkhiri, Ammar Tiri, Lotfi Mouni, Fatma Elhadj Lakouas
Abstract:
This study evaluates the surface water quality in the Kebir Rhumel Basin by analyzing hydrochemical parameters. To assess spatial and temporal variations in water quality, we applied the Water Quality Index (WQI), Mann-Kendall (MK) trend analysis, and hierarchical cluster analysis (HCA). Monthly measurements of eleven hydrochemical parameters were collected across eight stations from January 2016 to December 2020. Calcium and sulfate emerged as the dominant cation and anion, respectively. WQI analysis indicated a high incidence of poor water quality at stations Ain Smara (AS), Beni Haroune (BH), Grarem (GR), and Sidi Khalifa (SK), where 89.5%, 90.6%, 78.2%, and 62.7% of samples, respectively, fell into this category. The MK trend analysis revealed a significant upward trend in WQI at Oued Boumerzoug (ON) and SK stations, signaling temporal deterioration in these areas. HCA grouped the dataset into three clusters, covering approximately 22%, 30%, and 48% of the months, respectively. Within these clusters, specific stations exhibited elevated WQI values: GR and ON in the first cluster, OB and SK in the second, and AS, BH, El Milia (EM), and Hammam Grouz (HG) in the third. Furthermore, approximately 38%, 41%, and 38% of samples in clusters one, two, and three, respectively, were classified as having poor water quality. These findings provide essential insights for policymakers in formulating strategies to restore and manage surface water quality in the region.Keywords: surface water quality, water quality index (WQI), Mann-Kendall Trend Analysis, hierarchical cluster analysis (HCA), spatial-temporal distribution, Kebir Rhumel Basin
Procedia PDF Downloads 16617 An Overview of Electronic Waste as Aggregate in Concrete
Authors: S. R. Shamili, C. Natarajan, J. Karthikeyan
Abstract:
Rapid growth of world population and widespread urbanization has remarkably increased the development of the construction industry which caused a huge demand for sand and gravels. Environmental problems occur when the rate of extraction of sand, gravels, and other materials exceeds the rate of generation of natural resources; therefore, an alternative source is essential to replace the materials used in concrete. Now-a-days, electronic products have become an integral part of daily life which provides more comfort, security, and ease of exchange of information. These electronic waste (E-Waste) materials have serious human health concerns and require extreme care in its disposal to avoid any adverse impacts. Disposal or dumping of these E-Wastes also causes major issues because it is highly complex to handle and often contains highly toxic chemicals such as lead, cadmium, mercury, beryllium, brominates flame retardants (BFRs), polyvinyl chloride (PVC), and phosphorus compounds. Hence, E-Waste can be incorporated in concrete to make a sustainable environment. This paper deals with the composition, preparation, properties, classification of E-Waste. All these processes avoid dumping to landfills whilst conserving natural aggregate resources, and providing a better environmental option. This paper also provides a detailed literature review on the behaviour of concrete with incorporation of E-Wastes. Many research shows the strong possibility of using E-Waste as a substitute of aggregates eventually it reduces the use of natural aggregates in concrete.Keywords: dumping, electronic waste, landfill, toxic chemicals
Procedia PDF Downloads 169616 Improving Activity Recognition Classification of Repetitious Beginner Swimming Using a 2-Step Peak/Valley Segmentation Method with Smoothing and Resampling for Machine Learning
Authors: Larry Powell, Seth Polsley, Drew Casey, Tracy Hammond
Abstract:
Human activity recognition (HAR) systems have shown positive performance when recognizing repetitive activities like walking, running, and sleeping. Water-based activities are a reasonably new area for activity recognition. However, water-based activity recognition has largely focused on supporting the elite and competitive swimming population, which already has amazing coordination and proper form. Beginner swimmers are not perfect, and activity recognition needs to support the individual motions to help beginners. Activity recognition algorithms are traditionally built around short segments of timed sensor data. Using a time window input can cause performance issues in the machine learning model. The window’s size can be too small or large, requiring careful tuning and precise data segmentation. In this work, we present a method that uses a time window as the initial segmentation, then separates the data based on the change in the sensor value. Our system uses a multi-phase segmentation method that pulls all peaks and valleys for each axis of an accelerometer placed on the swimmer’s lower back. This results in high recognition performance using leave-one-subject-out validation on our study with 20 beginner swimmers, with our model optimized from our final dataset resulting in an F-Score of 0.95.Keywords: time window, peak/valley segmentation, feature extraction, beginner swimming, activity recognition
Procedia PDF Downloads 123