Search results for: Bal Deep Sharma
2143 Graph Clustering Unveiled: ClusterSyn - A Machine Learning Framework for Predicting Anti-Cancer Drug Synergy Scores
Authors: Babak Bahri, Fatemeh Yassaee Meybodi, Changiz Eslahchi
Abstract:
In the pursuit of effective cancer therapies, the exploration of combinatorial drug regimens is crucial to leverage synergistic interactions between drugs, thereby improving treatment efficacy and overcoming drug resistance. However, identifying synergistic drug pairs poses challenges due to the vast combinatorial space and limitations of experimental approaches. This study introduces ClusterSyn, a machine learning (ML)-powered framework for classifying anti-cancer drug synergy scores. ClusterSyn employs a two-step approach involving drug clustering and synergy score prediction using a fully connected deep neural network. For each cell line in the training dataset, a drug graph is constructed, with nodes representing drugs and edge weights denoting synergy scores between drug pairs. Drugs are clustered using the Markov clustering (MCL) algorithm, and vectors representing the similarity of drug pairs to each cluster are input into the deep neural network for synergy score prediction (synergy or antagonism). Clustering results demonstrate effective grouping of drugs based on synergy scores, aligning similar synergy profiles. Subsequently, neural network predictions and synergy scores of the two drugs on others within their clusters are used to predict the synergy score of the considered drug pair. This approach facilitates comparative analysis with clustering and regression-based methods, revealing the superior performance of ClusterSyn over state-of-the-art methods like DeepSynergy and DeepDDS on diverse datasets such as Oniel and Almanac. The results highlight the remarkable potential of ClusterSyn as a versatile tool for predicting anti-cancer drug synergy scores.Keywords: drug synergy, clustering, prediction, machine learning., deep learning
Procedia PDF Downloads 812142 Comparative Analysis of Predictive Models for Customer Churn Prediction in the Telecommunication Industry
Authors: Deepika Christopher, Garima Anand
Abstract:
To determine the best model for churn prediction in the telecom industry, this paper compares 11 machine learning algorithms, namely Logistic Regression, Support Vector Machine, Random Forest, Decision Tree, XGBoost, LightGBM, Cat Boost, AdaBoost, Extra Trees, Deep Neural Network, and Hybrid Model (MLPClassifier). It also aims to pinpoint the top three factors that lead to customer churn and conducts customer segmentation to identify vulnerable groups. According to the data, the Logistic Regression model performs the best, with an F1 score of 0.6215, 81.76% accuracy, 68.95% precision, and 56.57% recall. The top three attributes that cause churn are found to be tenure, Internet Service Fiber optic, and Internet Service DSL; conversely, the top three models in this article that perform the best are Logistic Regression, Deep Neural Network, and AdaBoost. The K means algorithm is applied to establish and analyze four different customer clusters. This study has effectively identified customers that are at risk of churn and may be utilized to develop and execute strategies that lower customer attrition.Keywords: attrition, retention, predictive modeling, customer segmentation, telecommunications
Procedia PDF Downloads 582141 KCBA, A Method for Feature Extraction of Colonoscopy Images
Authors: Vahid Bayrami Rad
Abstract:
In recent years, the use of artificial intelligence techniques, tools, and methods in processing medical images and health-related applications has been highlighted and a lot of research has been done in this regard. For example, colonoscopy and diagnosis of colon lesions are some cases in which the process of diagnosis of lesions can be improved by using image processing and artificial intelligence algorithms, which help doctors a lot. Due to the lack of accurate measurements and the variety of injuries in colonoscopy images, the process of diagnosing the type of lesions is a little difficult even for expert doctors. Therefore, by using different software and image processing, doctors can be helped to increase the accuracy of their observations and ultimately improve their diagnosis. Also, by using automatic methods, the process of diagnosing the type of disease can be improved. Therefore, in this paper, a deep learning framework called KCBA is proposed to classify colonoscopy lesions which are composed of several methods such as K-means clustering, a bag of features and deep auto-encoder. Finally, according to the experimental results, the proposed method's performance in classifying colonoscopy images is depicted considering the accuracy criterion.Keywords: colorectal cancer, colonoscopy, region of interest, narrow band imaging, texture analysis, bag of feature
Procedia PDF Downloads 572140 CO₂ Storage Capacity Assessment of Deep Saline Aquifers in Malaysia
Authors: Radzuan Junin, Dayang Zulaika A. Hasbollah
Abstract:
The increasing amount of greenhouse gasses in the atmosphere recently has become one of the discussed topics in relation with world’s concern on climate change. Developing countries’ emissions (such as Malaysia) are now seen to surpass developed country’s emissions due to rapid economic development growth in recent decades. This paper presents the potential storage sites suitability and storage capacity assessment for CO2 sequestration in sedimentary basins of Malaysia. This study is the first of its kind that made an identification of potential storage sites and assessment of CO2 storage capacity within the deep saline aquifers in the country. The CO2 storage capacity in saline formation assessment was conducted based on the method for quick assessment of CO2 storage capacity in closed, and semi-closed saline formations modified to suit the geology setting of Malaysia. Then, an integrated approach that involved geographic information systems (GIS) analysis and field data assessment was adopted to provide the potential storage sites and its capacity for CO2 sequestration. This study concentrated on the assessment of major sedimentary basins in Malaysia both onshore and offshore where potential geological formations which CO2 could be stored exist below 800 meters and where suitable sealing formations are present. Based on regional study and amount of data available, there are 14 sedimentary basins all around Malaysia that has been identified as potential CO2 storage. Meanwhile, from the screening and ranking exercises, it is obvious that Malay Basin, Central Luconia Province, West Baram Delta and Balingian Province are respectively ranked as the top four in the ranking system for CO2 storage. 27% of sedimentary basins in Malaysia were evaluated as high potential area for CO2 storage. This study should provide a basis for further work to reduce the uncertainty in these estimates and also provide support to policy makers on future planning of carbon capture and sequestration (CCS) projects in Malaysia.Keywords: CO₂ storage, deep saline aquifer, GIS, sedimentary basin
Procedia PDF Downloads 3612139 SEM Image Classification Using CNN Architectures
Authors: Güzi̇n Ti̇rkeş, Özge Teki̇n, Kerem Kurtuluş, Y. Yekta Yurtseven, Murat Baran
Abstract:
A scanning electron microscope (SEM) is a type of electron microscope mainly used in nanoscience and nanotechnology areas. Automatic image recognition and classification are among the general areas of application concerning SEM. In line with these usages, the present paper proposes a deep learning algorithm that classifies SEM images into nine categories by means of an online application to simplify the process. The NFFA-EUROPE - 100% SEM data set, containing approximately 21,000 images, was used to train and test the algorithm at 80% and 20%, respectively. Validation was carried out using a separate data set obtained from the Middle East Technical University (METU) in Turkey. To increase the accuracy in the results, the Inception ResNet-V2 model was used in view of the Fine-Tuning approach. By using a confusion matrix, it was observed that the coated-surface category has a negative effect on the accuracy of the results since it contains other categories in the data set, thereby confusing the model when detecting category-specific patterns. For this reason, the coated-surface category was removed from the train data set, hence increasing accuracy by up to 96.5%.Keywords: convolutional neural networks, deep learning, image classification, scanning electron microscope
Procedia PDF Downloads 1262138 Meta Mask Correction for Nuclei Segmentation in Histopathological Image
Authors: Jiangbo Shi, Zeyu Gao, Chen Li
Abstract:
Nuclei segmentation is a fundamental task in digital pathology analysis and can be automated by deep learning-based methods. However, the development of such an automated method requires a large amount of data with precisely annotated masks which is hard to obtain. Training with weakly labeled data is a popular solution for reducing the workload of annotation. In this paper, we propose a novel meta-learning-based nuclei segmentation method which follows the label correction paradigm to leverage data with noisy masks. Specifically, we design a fully conventional meta-model that can correct noisy masks by using a small amount of clean meta-data. Then the corrected masks are used to supervise the training of the segmentation model. Meanwhile, a bi-level optimization method is adopted to alternately update the parameters of the main segmentation model and the meta-model. Extensive experimental results on two nuclear segmentation datasets show that our method achieves the state-of-the-art result. In particular, in some noise scenarios, it even exceeds the performance of training on supervised data.Keywords: deep learning, histopathological image, meta-learning, nuclei segmentation, weak annotations
Procedia PDF Downloads 1412137 Construction of Strain Distribution Profiles of EDD Steel at Elevated Temperatures
Authors: K. Eshwara Prasad, R. Raman Goud, Swadesh Kumar Singh, N. Sateesh
Abstract:
In the present work forming limit diagrams and strain distribution profile diagrams for extra deep drawing steel at room and elevated temperatures have been determined experimentally by conducting stretch forming experiments by using designed and fabricated warm stretchforming tooling setup. With the help of forming Limit Diagrams (FLDs) and strain distribution profile diagrams the formability of Extra Deep Drawing steel has been analyzed and co-related with mechanical properties like strain hardening COEFFICIENT (n) and normal anisotropy (r−).Mechanical properties of EDD steel from room temperature to 4500C were determined and discussed the impact of temperature on the properties like work hardening exponent (n) anisotropy(r-) and strength coefficient of the material. Also the fractured surfaces after stretching have undergone the some metallurgical investigations and attempt has been made to co-relate with the formability of EDD steel sheets. They are co-related and good agreement with FLDs at various temperatures.Keywords: FLD, microhardness, strain distribution profile, stretch forming
Procedia PDF Downloads 3262136 Application of Deep Learning and Ensemble Methods for Biomarker Discovery in Diabetic Nephropathy through Fibrosis and Propionate Metabolism Pathways
Authors: Oluwafunmibi Omotayo Fasanya, Augustine Kena Adjei
Abstract:
Diabetic nephropathy (DN) is a major complication of diabetes, with fibrosis and propionate metabolism playing critical roles in its progression. Identifying biomarkers linked to these pathways may provide novel insights into DN diagnosis and treatment. This study aims to identify biomarkers associated with fibrosis and propionate metabolism in DN. Analyze the biological pathways and regulatory mechanisms of these biomarkers. Develop a machine learning model to predict DN-related biomarkers and validate their functional roles. Publicly available transcriptome datasets related to DN (GSE96804 and GSE104948) were obtained from the GEO database (https://www.ncbi.nlm.nih.gov/gds), and 924 propionate metabolism-related genes (PMRGs) and 656 fibrosis-related genes (FRGs) were identified. The analysis began with the extraction of DN-differentially expressed genes (DN-DEGs) and propionate metabolism-related DEGs (PM-DEGs), followed by the intersection of these with fibrosis-related genes to identify key intersected genes. Instead of relying on traditional models, we employed a combination of deep neural networks (DNNs) and ensemble methods such as Gradient Boosting Machines (GBM) and XGBoost to enhance feature selection and biomarker discovery. Recursive feature elimination (RFE) was coupled with these advanced algorithms to refine the selection of the most critical biomarkers. Functional validation was conducted using convolutional neural networks (CNN) for gene set enrichment and immunoinfiltration analysis, revealing seven significant biomarkers—SLC37A4, ACOX2, GPD1, ACE2, SLC9A3, AGT, and PLG. These biomarkers are involved in critical biological processes such as fatty acid metabolism and glomerular development, providing a mechanistic link to DN progression. Furthermore, a TF–miRNA–mRNA regulatory network was constructed using natural language processing models to identify 8 transcription factors and 60 miRNAs that regulate these biomarkers, while a drug–gene interaction network revealed potential therapeutic targets such as UROKINASE–PLG and ATENOLOL–AGT. This integrative approach, leveraging deep learning and ensemble models, not only enhances the accuracy of biomarker discovery but also offers new perspectives on DN diagnosis and treatment, specifically targeting fibrosis and propionate metabolism pathways.Keywords: diabetic nephropathy, deep neural networks, gradient boosting machines (GBM), XGBoost
Procedia PDF Downloads 122135 The Genesis of the Anomalous Sernio Fan (Valtellina, Northern Italy)
Authors: Erika De Finis, Paola Gattinoni, Laura Scesi
Abstract:
Massive rock avalanches formed some of the largest landslide deposits on Earth and they represent one of the major geohazards in high-relief mountains. This paper interprets a very large sedimentary fan (the Sernio fan, Valtellina, Northern Italy), located 20 Km SW from Val Pola Rock avalanche (1987), as the deposit of a partial collapse of a Deep Seated Gravitational Slope Deformation (DSGSD), afterwards eroded and buried by debris flows. The proposed emplacement sequence has been reconstructed based on geomorphological, structural and mechanical evidences. The Sernio fan is actually considered anomalous with reference to the very high ratio between the fan area (about 4.5km2) and the basin area (about 3km2). The morphology of the fan area is characterised by steep slopes (dip about 20%) and the fan apex is extended for 1.8 km inside the small catchment basin. This sedimentary fan was originated by a landslide that interested a part of a large deep-seated gravitational slope deformation, involving a wide area of about 55 km². The main controlling factor is tectonic and it is related to the proximity to regional fault systems and the consequent occurrence of fault weak rocks (GSI locally lower than 10 with compressive stress lower than 20MPa). Moreover, the fan deposit shows sedimentary evidences of recent debris flow events. The best current explanation of the Sernio fan involves an initial failure of some hundreds of Mm3. The run-out was quite limited because of the morphology of Valtellina’s valley floor, and the deposit filled the main valley forming a landslide dam, as confirmed by the lacustrine deposits detected upstream the fan. Nowadays the debris flow events represent the main hazard in the study area.Keywords: anomalous sedimentary fans, deep seated gravitational slope deformation, Italy, rock avalanche
Procedia PDF Downloads 4772134 Peg@GDF3:TB3+ – Rb Nanocomposites for Deep-Seated X-Ray Induced Photodynamic Therapy in Oncology
Authors: E.A. Kuchma
Abstract:
Photodynamic therapy (PDT) is considered an alternative and minimally invasive cancer treatment modality compared to chemotherapy and radiation therapy. PDT includes three main components: a photosensitizer (PS), oxygen, and a light source. PS is injected into the patient's body and then selectively accumulates in the tumor. However, the light used in PDT (spectral range 400–700 nm) is limited to superficial lesions, and the light penetration depth does not exceed a few cm. The problem of PDT (poor visible light transmission) can be solved by using X-rays. The penetration depth of X-rays is ten times greater than that of visible light. Therefore, X-ray radiation easily penetrates through the tissues of the body. The aim of this work is to develop universal nanocomposites for X-ray photodynamic therapy of deep and superficial tumors using scintillation nanoparticles of gadolinium fluoride (GdF3), doped with Tb3+, coated with a biocompatible coating (PEG) and photosensitizer RB (Rose Bengal). PEG@GdF3:Tb3+(15%) – RB could be used as an effective X-ray, UV, and photoluminescent mediator to excite a photosensitizer for generating reactive oxygen species (ROS) to kill tumor cells via photodynamic therapy. GdF3 nanoparticles can also be used as contrast agents for computed tomography (CT) and magnetic resonance imaging (MRI).Keywords: X-ray induced photodynamic therapy, scintillating nanoparticle, radiosensitizer, photosensitizer
Procedia PDF Downloads 812133 A Comparison of Convolutional Neural Network Architectures for the Classification of Alzheimer’s Disease Patients Using MRI Scans
Authors: Tomas Premoli, Sareh Rowlands
Abstract:
In this study, we investigate the impact of various convolutional neural network (CNN) architectures on the accuracy of diagnosing Alzheimer’s disease (AD) using patient MRI scans. Alzheimer’s disease is a debilitating neurodegenerative disorder that affects millions worldwide. Early, accurate, and non-invasive diagnostic methods are required for providing optimal care and symptom management. Deep learning techniques, particularly CNNs, have shown great promise in enhancing this diagnostic process. We aim to contribute to the ongoing research in this field by comparing the effectiveness of different CNN architectures and providing insights for future studies. Our methodology involved preprocessing MRI data, implementing multiple CNN architectures, and evaluating the performance of each model. We employed intensity normalization, linear registration, and skull stripping for our preprocessing. The selected architectures included VGG, ResNet, and DenseNet models, all implemented using the Keras library. We employed transfer learning and trained models from scratch to compare their effectiveness. Our findings demonstrated significant differences in performance among the tested architectures, with DenseNet201 achieving the highest accuracy of 86.4%. Transfer learning proved to be helpful in improving model performance. We also identified potential areas for future research, such as experimenting with other architectures, optimizing hyperparameters, and employing fine-tuning strategies. By providing a comprehensive analysis of the selected CNN architectures, we offer a solid foundation for future research in Alzheimer’s disease diagnosis using deep learning techniques. Our study highlights the potential of CNNs as a valuable diagnostic tool and emphasizes the importance of ongoing research to develop more accurate and effective models.Keywords: Alzheimer’s disease, convolutional neural networks, deep learning, medical imaging, MRI
Procedia PDF Downloads 742132 Strain DistributionProfiles of EDD Steel at Elevated Temperatures
Authors: Eshwara Prasad Koorapati, R. Raman Goud, Swadesh Kumar Singh
Abstract:
In the present work forming limit diagrams and strain distribution profile diagrams for extra deep drawing steel at room and elevated temperatures have been determined experimentally by conducting stretch forming experiments by using designed and fabricated warm stretch forming tooling setup. With the help of forming Limit Diagrams (FLDs) and strain distribution profile diagrams the formability of Extra Deep Drawing steel has been analyzed and co-related with mechanical properties like strain hardening coefficient (n) and normal anisotropy (r−).Mechanical properties of EDD steel from room temperature to 4500 C were determined and discussed the impact of temperature on the properties like work hardening exponent (n) anisotropy (r-) and strength coefficient of the material. Also, the fractured surfaces after stretching have undergone the some metallurgical investigations and attempt has been made to co-relate with the formability of EDD steel sheets. They are co-related and good agreement with FLDs at various temperatures.Keywords: FLD, micro hardness, strain distribution profile, stretch forming
Procedia PDF Downloads 4222131 Reinforcement Learning for Self Driving Racing Car Games
Authors: Adam Beaunoyer, Cory Beaunoyer, Mohammed Elmorsy, Hanan Saleh
Abstract:
This research aims to create a reinforcement learning agent capable of racing in challenging simulated environments with a low collision count. We present a reinforcement learning agent that can navigate challenging tracks using both a Deep Q-Network (DQN) and a Soft Actor-Critic (SAC) method. A challenging track includes curves, jumps, and varying road widths throughout. Using open-source code on Github, the environment used in this research is based on the 1995 racing game WipeOut. The proposed reinforcement learning agent can navigate challenging tracks rapidly while maintaining low racing completion time and collision count. The results show that the SAC model outperforms the DQN model by a large margin. We also propose an alternative multiple-car model that can navigate the track without colliding with other vehicles on the track. The SAC model is the basis for the multiple-car model, where it can complete the laps quicker than the single-car model but has a higher collision rate with the track wall.Keywords: reinforcement learning, soft actor-critic, deep q-network, self-driving cars, artificial intelligence, gaming
Procedia PDF Downloads 492130 Generating Synthetic Chest X-ray Images for Improved COVID-19 Detection Using Generative Adversarial Networks
Authors: Muneeb Ullah, Daishihan, Xiadong Young
Abstract:
Deep learning plays a crucial role in identifying COVID-19 and preventing its spread. To improve the accuracy of COVID-19 diagnoses, it is important to have access to a sufficient number of training images of CXRs (chest X-rays) depicting the disease. However, there is currently a shortage of such images. To address this issue, this paper introduces COVID-19 GAN, a model that uses generative adversarial networks (GANs) to generate realistic CXR images of COVID-19, which can be used to train identification models. Initially, a generator model is created that uses digressive channels to generate images of CXR scans for COVID-19. To differentiate between real and fake disease images, an efficient discriminator is developed by combining the dense connectivity strategy and instance normalization. This approach makes use of their feature extraction capabilities on CXR hazy areas. Lastly, the deep regret gradient penalty technique is utilized to ensure stable training of the model. With the use of 4,062 grape leaf disease images, the Leaf GAN model successfully produces 8,124 COVID-19 CXR images. The COVID-19 GAN model produces COVID-19 CXR images that outperform DCGAN and WGAN in terms of the Fréchet inception distance. Experimental findings suggest that the COVID-19 GAN-generated CXR images possess noticeable haziness, offering a promising approach to address the limited training data available for COVID-19 model training. When the dataset was expanded, CNN-based classification models outperformed other models, yielding higher accuracy rates than those of the initial dataset and other augmentation techniques. Among these models, ImagNet exhibited the best recognition accuracy of 99.70% on the testing set. These findings suggest that the proposed augmentation method is a solution to address overfitting issues in disease identification and can enhance identification accuracy effectively.Keywords: classification, deep learning, medical images, CXR, GAN.
Procedia PDF Downloads 1002129 Groundwater Level Prediction Using hybrid Particle Swarm Optimization-Long-Short Term Memory Model and Performance Evaluation
Authors: Sneha Thakur, Sanjeev Karmakar
Abstract:
This paper proposed hybrid Particle Swarm Optimization (PSO) – Long-Short Term Memory (LSTM) model for groundwater level prediction. The evaluation of the performance is realized using the parameters: root mean square error (RMSE) and mean absolute error (MAE). Ground water level forecasting will be very effective for planning water harvesting. Proper calculation of water level forecasting can overcome the problem of drought and flood to some extent. The objective of this work is to develop a ground water level forecasting model using deep learning technique integrated with optimization technique PSO by applying 29 years data of Chhattisgarh state, In-dia. It is important to find the precise forecasting in case of ground water level so that various water resource planning and water harvesting can be managed effectively.Keywords: long short-term memory, particle swarm optimization, prediction, deep learning, groundwater level
Procedia PDF Downloads 792128 Learning from Small Amount of Medical Data with Noisy Labels: A Meta-Learning Approach
Authors: Gorkem Algan, Ilkay Ulusoy, Saban Gonul, Banu Turgut, Berker Bakbak
Abstract:
Computer vision systems recently made a big leap thanks to deep neural networks. However, these systems require correctly labeled large datasets in order to be trained properly, which is very difficult to obtain for medical applications. Two main reasons for label noise in medical applications are the high complexity of the data and conflicting opinions of experts. Moreover, medical imaging datasets are commonly tiny, which makes each data very important in learning. As a result, if not handled properly, label noise significantly degrades the performance. Therefore, a label-noise-robust learning algorithm that makes use of the meta-learning paradigm is proposed in this article. The proposed solution is tested on retinopathy of prematurity (ROP) dataset with a very high label noise of 68%. Results show that the proposed algorithm significantly improves the classification algorithm's performance in the presence of noisy labels.Keywords: deep learning, label noise, robust learning, meta-learning, retinopathy of prematurity
Procedia PDF Downloads 1622127 Deep Learning Based Fall Detection Using Simplified Human Posture
Authors: Kripesh Adhikari, Hamid Bouchachia, Hammadi Nait-Charif
Abstract:
Falls are one of the major causes of injury and death among elderly people aged 65 and above. A support system to identify such kind of abnormal activities have become extremely important with the increase in ageing population. Pose estimation is a challenging task and to add more to this, it is even more challenging when pose estimations are performed on challenging poses that may occur during fall. Location of the body provides a clue where the person is at the time of fall. This paper presents a vision-based tracking strategy where available joints are grouped into three different feature points depending upon the section they are located in the body. The three feature points derived from different joints combinations represents the upper region or head region, mid-region or torso and lower region or leg region. Tracking is always challenging when a motion is involved. Hence the idea is to locate the regions in the body in every frame and consider it as the tracking strategy. Grouping these joints can be beneficial to achieve a stable region for tracking. The location of the body parts provides a crucial information to distinguish normal activities from falls.Keywords: fall detection, machine learning, deep learning, pose estimation, tracking
Procedia PDF Downloads 1892126 Investigation of Buddhology Reflected from Wall Paintings in Sri Lanka
Authors: R. G. D Jayawardena
Abstract:
The Buddha was known by great wise men from 6th century B.C up to date as a superhuman being born in the world beyond the omnipotent. The Buddha’s doctrinal descriptions reflect his deep enlightenment about imperial and metaphysical knowledge. Buddhology undertaken for this study is an unexposed subject in metaphysical points. The Buddhist wall painting in Sri Lanka depicts deep metaphysical meaning than its simple perspective of estheticism. Buddhology, in some perspectives, has been interpreted as a complete natural science discovered by the Buddha to teach the way of honorable living in perfect happiness and peace of mind till death. Such interpretations which emphasized are based on textual studies. The Buddhology conducted through literal tradition is depicted in wall paintings in Sri Lanka are in visual art with specific techniques rules. The Buddhology, which is investigated on wall paintings, portrays the Buddha in the form of a superhuman being and as an unparalleled person among the Devas, Brahmas, Yakshas, Maras, and humans. The Buddha concept is known to Sri Lankan Buddhists as a person attained to full awakening of wisdom. In personality, the Buddha is depicted as a supernormal person in the world and a rare birth. In brief, the paper will discuss and illustrate the Buddha’s transcendental position and the reality of what he experienced and its authenticity.Keywords: Buddhology, Metaphysic, Sri Lanka, paintings
Procedia PDF Downloads 2072125 Morphological Processing of Punjabi Text for Sentiment Analysis of Farmer Suicides
Authors: Jaspreet Singh, Gurvinder Singh, Prabhsimran Singh, Rajinder Singh, Prithvipal Singh, Karanjeet Singh Kahlon, Ravinder Singh Sawhney
Abstract:
Morphological evaluation of Indian languages is one of the burgeoning fields in the area of Natural Language Processing (NLP). The evaluation of a language is an eminent task in the era of information retrieval and text mining. The extraction and classification of knowledge from text can be exploited for sentiment analysis and morphological evaluation. This study coalesce morphological evaluation and sentiment analysis for the task of classification of farmer suicide cases reported in Punjab state of India. The pre-processing of Punjabi text involves morphological evaluation and normalization of Punjabi word tokens followed by the training of proposed model using deep learning classification on Punjabi language text extracted from online Punjabi news reports. The class-wise accuracies of sentiment prediction for four negatively oriented classes of farmer suicide cases are 93.85%, 88.53%, 83.3%, and 95.45% respectively. The overall accuracy of sentiment classification obtained using proposed framework on 275 Punjabi text documents is found to be 90.29%.Keywords: deep neural network, farmer suicides, morphological processing, punjabi text, sentiment analysis
Procedia PDF Downloads 3282124 Missing Link Data Estimation with Recurrent Neural Network: An Application Using Speed Data of Daegu Metropolitan Area
Authors: JaeHwan Yang, Da-Woon Jeong, Seung-Young Kho, Dong-Kyu Kim
Abstract:
In terms of ITS, information on link characteristic is an essential factor for plan or operation. But in practical cases, not every link has installed sensors on it. The link that does not have data on it is called “Missing Link”. The purpose of this study is to impute data of these missing links. To get these data, this study applies the machine learning method. With the machine learning process, especially for the deep learning process, missing link data can be estimated from present link data. For deep learning process, this study uses “Recurrent Neural Network” to take time-series data of road. As input data, Dedicated Short-range Communications (DSRC) data of Dalgubul-daero of Daegu Metropolitan Area had been fed into the learning process. Neural Network structure has 17 links with present data as input, 2 hidden layers, for 1 missing link data. As a result, forecasted data of target link show about 94% of accuracy compared with actual data.Keywords: data estimation, link data, machine learning, road network
Procedia PDF Downloads 5102123 Comparison of Deep Brain Stimulation Targets in Parkinson's Disease: A Systematic Review
Authors: Hushyar Azari
Abstract:
Aim and background: Deep brain stimulation (DBS) is regarded as an important therapeutic choice for Parkinson's disease (PD). The two most common targets for DBS are the subthalamic nucleus (STN) and globus pallidus (GPi). This review was conducted to compare the clinical effectiveness of these two targets. Methods: A systematic literature search in electronic databases: Embase, Cochrane Library and PubMed were restricted to English language publications 2010 to 2021. Specified MeSH terms were searched in all databases. Studies which evaluated the Unified Parkinson's Disease Rating Scale (UPDRS) III were selected by meeting the following criteria: (1) compared both GPi and STN DBS; (2) had at least three months follow-up period; (3)at least five participants in each group; (4)conducted after 2010. Study quality assessment was performed using the Modified Jadad Scale. Results: 3577 potentially relevant articles were identified, of these, 3569 were excluded based on title and abstract, duplicate and unsuitable article removal. Eight articles satisfied the inclusion criteria and were scrutinized (458 PD patients). According to Modified Jadad Scale, the majority of included studies had low evidence quality which was a limitation of this review. 5 studies reported no statistically significant between-group difference for improvements in UPDRS ш scores. At the same time, there were some results in terms of pain, action tremor, rigidity, and urinary symptoms, which indicated that STN DBS might be a better choice. Regarding the adverse effects, GPi was superior. Conclusion: It is clear that other larger randomized clinical trials with longer follow-up periods and control groups are needed to decide which target is more efficient for deep brain stimulation in Parkinson’s disease and imposes fewer adverse effects on the patients. Meanwhile, STN seems more reasonable according to the results of this systematic review.Keywords: brain stimulation, globus pallidus, Parkinson's disease, subthalamic nucleus
Procedia PDF Downloads 1792122 Thermosonic Devulcanization of Waste Ground Rubber Tires by Quaternary Ammonium-Based Ternary Deep Eutectic Solvents and the Effect of α-Hydrogen
Authors: Ricky Saputra, Rashmi Walvekar, Mohammad Khalid
Abstract:
Landfills, water contamination, and toxic gas emission are a few impacts faced by the environment due to the increasing number of αof waste rubber tires (WRT). In spite of such concerning issue, only minimal efforts are taken to reclaim or recycle these wastes as their products are generally not-profitable for companies. Unlike the typical reclamation process, devulcanization is a method to selectively cleave sulfidic bonds within vulcanizates to avoid polymeric scissions that compromise elastomer’s mechanical and tensile properties. The process also produces devulcanizates that are re-processable similar to virgin rubber. Often, a devulcanizing agent is needed. In the current study, novel and sustainable ammonium chloride-based ternary deep eutectic solvents (TDES), with a different number of α-hydrogens, were utilised to devulcanize ground rubber tire (GRT) as an effort to implement green chemistry to tackle such issue. 40-mesh GRT were soaked for 1 day with different TDESs and sonicated at 37-80 kHz for 60-120 mins and heated at 100-140oC for 30-90 mins. Devulcanizates were then filtered, dried, and evaluated based on the percentage of by means of Flory-Rehner calculation and swelling index. The result shows that an increasing number of α-Hs increases the degree of devulcanization, and the value achieved was around eighty-percent, thirty percent higher than the typical industrial-autoclave method. Resulting bondages of devulcanizates were also analysed by Fourier transform infrared spectrometer (FTIR), Horikx fitting, and thermogravimetric analyser (TGA). The earlier two confirms only sulfidic scissions were experienced by GRT through the treatment, while the latter proves the absence or negligibility of carbon-chains scission.Keywords: ammonium, sustainable, deep eutectic solvent, α-hydrogen, waste rubber tire
Procedia PDF Downloads 1272121 Wasting Human and Computer Resources
Authors: Mária Csernoch, Piroska Biró
Abstract:
The legends about “user-friendly” and “easy-to-use” birotical tools (computer-related office tools) have been spreading and misleading end-users. This approach has led us to the extremely high number of incorrect documents, causing serious financial losses in the creating, modifying, and retrieving processes. Our research proved that there are at least two sources of this underachievement: (1) The lack of the definition of the correctly edited, formatted documents. Consequently, end-users do not know whether their methods and results are correct or not. They are not aware of their ignorance. They are so ignorant that their ignorance does not allow them to realize their lack of knowledge. (2) The end-users’ problem-solving methods. We have found that in non-traditional programming environments end-users apply, almost exclusively, surface approach metacognitive methods to carry out their computer related activities, which are proved less effective than deep approach methods. Based on these findings we have developed deep approach methods which are based on and adapted from traditional programming languages. In this study, we focus on the most popular type of birotical documents, the text-based documents. We have provided the definition of the correctly edited text, and based on this definition, adapted the debugging method known in programming. According to the method, before the realization of text editing, a thorough debugging of already existing texts and the categorization of errors are carried out. With this method in advance to real text editing users learn the requirements of text-based documents and also of the correctly formatted text. The method has been proved much more effective than the previously applied surface approach methods. The advantages of the method are that the real text handling requires much less human and computer sources than clicking aimlessly in the GUI (Graphical User Interface), and the data retrieval is much more effective than from error-prone documents.Keywords: deep approach metacognitive methods, error-prone birotical documents, financial losses, human and computer resources
Procedia PDF Downloads 3822120 AI-Based Techniques for Online Social Media Network Sentiment Analysis: A Methodical Review
Authors: A. M. John-Otumu, M. M. Rahman, O. C. Nwokonkwo, M. C. Onuoha
Abstract:
Online social media networks have long served as a primary arena for group conversations, gossip, text-based information sharing and distribution. The use of natural language processing techniques for text classification and unbiased decision-making has not been far-fetched. Proper classification of this textual information in a given context has also been very difficult. As a result, we decided to conduct a systematic review of previous literature on sentiment classification and AI-based techniques that have been used in order to gain a better understanding of the process of designing and developing a robust and more accurate sentiment classifier that can correctly classify social media textual information of a given context between hate speech and inverted compliments with a high level of accuracy by assessing different artificial intelligence techniques. We evaluated over 250 articles from digital sources like ScienceDirect, ACM, Google Scholar, and IEEE Xplore and whittled down the number of research to 31. Findings revealed that Deep learning approaches such as CNN, RNN, BERT, and LSTM outperformed various machine learning techniques in terms of performance accuracy. A large dataset is also necessary for developing a robust sentiment classifier and can be obtained from places like Twitter, movie reviews, Kaggle, SST, and SemEval Task4. Hybrid Deep Learning techniques like CNN+LSTM, CNN+GRU, CNN+BERT outperformed single Deep Learning techniques and machine learning techniques. Python programming language outperformed Java programming language in terms of sentiment analyzer development due to its simplicity and AI-based library functionalities. Based on some of the important findings from this study, we made a recommendation for future research.Keywords: artificial intelligence, natural language processing, sentiment analysis, social network, text
Procedia PDF Downloads 1162119 Demystifying Board Games for Teachers
Authors: Shilpa Sharma, Lakshmi Ganesh, Mantra Gurumurthy, Shweta Sharma
Abstract:
Board games provide affordances of 21st-century skills like collaboration, critical thinking, and strategy. Board games such as chess, Catan, Battleship, Scrabble, and Taboo can enhance learning in these areas. While board games are popular in informal child settings, their use in formal K-12 education is limited. To encourage teachers to incorporate board games, it's essential to grasp their perceptions and tailor professional development programs accordingly. This paper aims to explore teacher attitudes toward board games and propose interventions to motivate teachers to integrate and create board games in the classroom. A user study was conceived, designed, and administered with teachers (n=38) to understand their experience in playing board games and using board games in the classroom. Purposive sampling was employed as the questionnaire was floated to teacher groups that the authors were aware of. The teachers taught in K-12 affordable private schools. The majority of them had experience ranging from 2-5 years. The questionnaire consisted of questions on teacher perceptions and beliefs of board game usage in the classroom. From the responses, it was observed that ~90% of teachers, though they had experience of playing board games, rarely did it translate to using board games in the classroom. Additionally, it was observed that translating learning objectives to board game objectives is the key factor that teachers consider while using board games in the classroom. Based on the results from the questionnaire, a professional development workshop was co-designed with the objective of motivating teachers to design, create and use board games in the classroom. The workshop is based on the principles of gamification. This is to ensure that the teachers experience a board game in a learning context. Additionally, the workshop is based on the principles of andragogy, such as agency, pertinence, and relevance. The workshop will begin by modifying and reusing known board games in the learning context so that the teachers do not find it difficult and daunting. The intention is to verify the face validity and content validity of the workshop design, orchestration and content with experienced teacher development professionals and education researchers. The results from this study will be published in the full paper.Keywords: board games, professional development, teacher motivation, teacher perception
Procedia PDF Downloads 1102118 Circle Work as a Relational Praxis to Facilitate Collaborative Learning within Higher Education: A Decolonial Pedagogical Framework for Teaching and Learning in the Virtual Classroom
Authors: Jennifer Nutton, Gayle Ployer, Ky Scott, Jenny Morgan
Abstract:
Working in a circle within higher education creates a decolonial space of mutual respect, responsibility, and reciprocity that facilitates collaborative learning and deep connections among learners and instructors. This approach is beyond simply facilitating a group in a circle but opens the door to creating a sacred space connecting each member to the land, to the Indigenous peoples who have taken care of the lands since time immemorial, to one another, and to one’s own positionality. These deep connections not only center human knowledges and relationships but also acknowledges responsibilities to land. Working in a circle as a relational pedagogical praxis also disrupts institutional power dynamics by creating a space of collaborative learning and deep connections in the classroom. Inherent within circle work is to facilitate connections not just academically but emotionally, physically, culturally, and spiritually. Recent literature supports the use of online talking circles, finding that it can offer a more relational and experiential learning environment, which is often absent in the virtual world and has been made more evident and necessary since the pandemic. These deeper experiences of learning and connection, rooted in both knowledge and the land, can then be shared with openness and vulnerability with one another, facilitating growth and change. This process of beginning with the land is critical to ensure we have the grounding to obstruct the ongoing realities of colonialism. The authors, who identify as both Indigenous and non-Indigenous, as both educators and learners, reflect on their teaching and learning experiences in circle. They share a relational pedagogical praxis framework that has been successful in educating future social workers, environmental activists, and leaders in social and human services, health, legal and political fields.Keywords: circle work, relational pedagogies, decolonization, distance education
Procedia PDF Downloads 762117 Healthy Feeding and Drinking Troughs for Profitable Intensive Deep-Litter Poultry Farming
Authors: Godwin Ojochogu Adejo, Evelyn UnekwuOjo Adejo, Sunday UnenwOjo Adejo
Abstract:
The mainstream contemporary approach to controlling the impact of diseases among poultry birds rely largely on curative measures through the administration of drugs to infected birds. Most times as observed in the deep liter poultry farming system, entire flocks including uninfected birds receive the treatment they do not need. As such, unguarded use of chemical drugs and antibiotics has led to wastage and accumulation of chemical residues in poultry products with associated health hazards to humans. However, wanton and frequent drug usage in poultry is avoidable if feeding and drinking equipment are designed to curb infection transmission among birds. Using toxicological assays as guide and with efficiency and simplicity in view, two newly field-tested and recently patented equipments called 'healthy liquid drinking trough (HDT)' and 'healthy feeding trough (HFT)' that systematically eliminate contamination of the feeding and drinking channels, thereby, curbing wide-spread infection and transmission of diseases in the (intensive) deep litter poultry farming system were designed. Upon combined usage, they automatically and drastically reduced both the amount and frequency of antibiotics use in poultry by over > 50%. Additionally, they conferred optimization of feed and water utilization/elimination of wastage by > 80%, reduced labour by > 70%, reduced production cost by about 15%, and reduced chemical residues in poultry meat or eggs by > 85%. These new and cheap technologies which require no energy input are likely to elevate safety of poultry products for consumers' health, increase marketability locally and for export, and increase output and profit especially among poultry farmers and poor people who keep poultry or inevitably utilize poultry products in developing countries.Keywords: healthy, trough, toxicological, assay-guided, poultry
Procedia PDF Downloads 1572116 Spectrogram Pre-Processing to Improve Isotopic Identification to Discriminate Gamma and Neutrons Sources
Authors: Mustafa Alhamdi
Abstract:
Industrial application to classify gamma rays and neutron events is investigated in this study using deep machine learning. The identification using a convolutional neural network and recursive neural network showed a significant improvement in predication accuracy in a variety of applications. The ability to identify the isotope type and activity from spectral information depends on feature extraction methods, followed by classification. The features extracted from the spectrum profiles try to find patterns and relationships to present the actual spectrum energy in low dimensional space. Increasing the level of separation between classes in feature space improves the possibility to enhance classification accuracy. The nonlinear nature to extract features by neural network contains a variety of transformation and mathematical optimization, while principal component analysis depends on linear transformations to extract features and subsequently improve the classification accuracy. In this paper, the isotope spectrum information has been preprocessed by finding the frequencies components relative to time and using them as a training dataset. Fourier transform implementation to extract frequencies component has been optimized by a suitable windowing function. Training and validation samples of different isotope profiles interacted with CdTe crystal have been simulated using Geant4. The readout electronic noise has been simulated by optimizing the mean and variance of normal distribution. Ensemble learning by combing voting of many models managed to improve the classification accuracy of neural networks. The ability to discriminate gamma and neutron events in a single predication approach using deep machine learning has shown high accuracy using deep learning. The paper findings show the ability to improve the classification accuracy by applying the spectrogram preprocessing stage to the gamma and neutron spectrums of different isotopes. Tuning deep machine learning models by hyperparameter optimization of neural network models enhanced the separation in the latent space and provided the ability to extend the number of detected isotopes in the training database. Ensemble learning contributed significantly to improve the final prediction.Keywords: machine learning, nuclear physics, Monte Carlo simulation, noise estimation, feature extraction, classification
Procedia PDF Downloads 1512115 Mentha piperita Formulations in Natural Deep Eutectic Solvents: Phenolic Profile and Biological Activity
Authors: Tatjana Jurić, Bojana Blagojević, Denis Uka, Ružica Ždero Pavlović, Boris M. Popović
Abstract:
Natural deep eutectic solvents (NADES) represent a class of modern systems that have been developed as a green alternative to toxic organic solvents, which are commonly used as extraction media. It has been considered that hydrogen bonding is the main interaction leading to the formation of NADES. The aim of this study was phytochemical characterization and determination of the antioxidant and antibacterial activity of Mentha piperita leaf extracts obtained by six choline chloride-based NADES. NADES were prepared by mixing choline chloride with different hydrogen bond donors in 1:1 molar ratio following the addition of 30% (w/w) water. The mixtures were then heated (60 °C) and stirred (650 rpm) until the clear homogenous liquids were obtained. The Mentha piperita extracts were prepared by mixing 75 mg of peppermint leaves with 1 mL of NADES following by the heating and stirring (60 °C, 650 rpm) within 30 min. The content of six phenolics in extracts was determined using HPLC-PDA. The dominant compounds presented in peppermint leaves - rosmarinic acid and luteolin 7-O-glucoside, were extracted by NADES at a similar level as 70% ethanol. The microdilution method was applied to test the antibacterial activity of extracts. Compared with 70% ethanol, all NADES systems showed higher antibacterial activity towards Pseudomonas aeruginosa (Gram -), Staphylococcus aureus (Gram +), Escherichia coli (Gram -), and Salmonella enterica (Gram -), especially NADES containing organic acids. The majority of NADES extracts showed a better ability to neutralize DPPH radical than conventional solvent and similar ability to reduce Fe3+ to Fe2+ ions in FRAP assay. The obtained results introduce NADES systems as the novel, sustainable, and low-cost solvents with a variety of applications.Keywords: antibacterial activity, antioxidant activity, green extraction, natural deep eutectic solvents, polyphenols
Procedia PDF Downloads 1862114 Deep Learning Prediction of Residential Radon Health Risk in Canada and Sweden to Prevent Lung Cancer Among Non-Smokers
Authors: Selim M. Khan, Aaron A. Goodarzi, Joshua M. Taron, Tryggve Rönnqvist
Abstract:
Indoor air quality, a prime determinant of health, is strongly influenced by the presence of hazardous radon gas within the built environment. As a health issue, dangerously high indoor radon arose within the 20th century to become the 2nd leading cause of lung cancer. While the 21st century building metrics and human behaviors have captured, contained, and concentrated radon to yet higher and more hazardous levels, the issue is rapidly worsening in Canada. It is established that Canadians in the Prairies are the 2nd highest radon-exposed population in the world, with 1 in 6 residences experiencing 0.2-6.5 millisieverts (mSv) radiation per week, whereas the Canadian Nuclear Safety Commission sets maximum 5-year occupational limits for atomic workplace exposure at only 20 mSv. This situation is also deteriorating over time within newer housing stocks containing higher levels of radon. Deep machine learning (LSTM) algorithms were applied to analyze multiple quantitative and qualitative features, determine the most important contributory factors, and predicted radon levels in the known past (1990-2020) and projected future (2021-2050). The findings showed gradual downwards patterns in Sweden, whereas it would continue to go from high to higher levels in Canada over time. The contributory factors found to be the basement porosity, roof insulation depthness, R-factor, and air dynamics of the indoor environment related to human window opening behaviour. Building codes must consider including these factors to ensure adequate indoor ventilation and healthy living that can prevent lung cancer in non-smokers.Keywords: radon, building metrics, deep learning, LSTM prediction model, lung cancer, canada, sweden
Procedia PDF Downloads 112