Search results for: hyperspectral image classification using tree search algorithm
8518 Application of Model Tree in the Prediction of TBM Rate of Penetration with Synthetic Minority Oversampling Technique
Authors: Ehsan Mehryaar
Abstract:
The rate of penetration is (RoP) one of the vital factors in the cost and time of tunnel boring projects; therefore, predicting it can lead to a substantial increase in the efficiency of the project. RoP is heavily dependent geological properties of the project site and TBM properties. In this study, 151-point data from Queen’s water tunnel is collected, which includes unconfined compression strength, peak slope index, angle with weak planes, and distance between planes of weaknesses. Since the size of the data is small, it was observed that it is imbalanced. To solve that problem synthetic minority oversampling technique is utilized. The model based on the model tree is proposed, where each leaf consists of a support vector machine model. Proposed model performance is then compared to existing empirical equations in the literature.Keywords: Model tree, SMOTE, rate of penetration, TBM(tunnel boring machine), SVM
Procedia PDF Downloads 1748517 Fast Return Path Planning for Agricultural Autonomous Terrestrial Robot in a Known Field
Authors: Carlo Cernicchiaro, Pedro D. Gaspar, Martim L. Aguiar
Abstract:
The agricultural sector is becoming more critical than ever in view of the expected overpopulation of the Earth. The introduction of robotic solutions in this field is an increasingly researched topic to make the most of the Earth's resources, thus going to avoid the problems of wear and tear of the human body due to the harsh agricultural work, and open the possibility of a constant careful processing 24 hours a day. This project is realized for a terrestrial autonomous robot aimed to navigate in an orchard collecting fallen peaches below the trees. When it receives the signal indicating the low battery, it has to return to the docking station where it will replace its battery and then return to the last work point and resume its routine. Considering a preset path in orchards with tree rows with variable length by which the robot goes iteratively using the algorithm D*. In case of low battery, the D* algorithm is still used to determine the fastest return path to the docking station as well as to come back from the docking station to the last work point. MATLAB simulations were performed to analyze the flexibility and adaptability of the developed algorithm. The simulation results show an enormous potential for adaptability, particularly in view of the irregularity of orchard field, since it is not flat and undergoes modifications over time from fallen branch as well as from other obstacles and constraints. The D* algorithm determines the best route in spite of the irregularity of the terrain. Moreover, in this work, it will be shown a possible solution to improve the initial points tracking and reduce time between movements.Keywords: path planning, fastest return path, agricultural autonomous terrestrial robot, docking station
Procedia PDF Downloads 1358516 Advances of Image Processing in Precision Agriculture: Using Deep Learning Convolution Neural Network for Soil Nutrient Classification
Authors: Halimatu S. Abdullahi, Ray E. Sheriff, Fatima Mahieddine
Abstract:
Agriculture is essential to the continuous existence of human life as they directly depend on it for the production of food. The exponential rise in population calls for a rapid increase in food with the application of technology to reduce the laborious work and maximize production. Technology can aid/improve agriculture in several ways through pre-planning and post-harvest by the use of computer vision technology through image processing to determine the soil nutrient composition, right amount, right time, right place application of farm input resources like fertilizers, herbicides, water, weed detection, early detection of pest and diseases etc. This is precision agriculture which is thought to be solution required to achieve our goals. There has been significant improvement in the area of image processing and data processing which has being a major challenge. A database of images is collected through remote sensing, analyzed and a model is developed to determine the right treatment plans for different crop types and different regions. Features of images from vegetations need to be extracted, classified, segmented and finally fed into the model. Different techniques have been applied to the processes from the use of neural network, support vector machine, fuzzy logic approach and recently, the most effective approach generating excellent results using the deep learning approach of convolution neural network for image classifications. Deep Convolution neural network is used to determine soil nutrients required in a plantation for maximum production. The experimental results on the developed model yielded results with an average accuracy of 99.58%.Keywords: convolution, feature extraction, image analysis, validation, precision agriculture
Procedia PDF Downloads 3188515 Spatial Interactions Between Earthworm Abundance and Tree Growth Characteristics in Western Niger Delta
Authors: Olatunde Sunday Eludoyin, Charles Obiechina Olisa
Abstract:
The study examined the spatial interactions between earthworm abundance (EA) and tree growth characteristics in ecological belts of Western Niger Delta, Nigeria. Eight 20m x 20m quadrat were delimited in the natural vegetation in each of the rainforest (RF), mangrove (M), fresh water swamp (FWS), and guinea savanna (GS) ecological belts to gather data about the tree species (TS) characteristics which included individual number of tree species (IN), diversity (Di), density (De) and richness (Ri). Three quadrats of 1m x 1m were delineated in each of the 20m x 20m quadrats to collect earthworm species the topsoil (0-15cm), and subsoil (15-30cm) and were taken to laboratory for further analysis. Descriptive statistics and inferential statistics were used for data analysis. Findings showed that a total of 19 earthworm species was found, with 58.5% individual species recorded in the topsoil and 41.5% recorded in the subsoil. The total population ofEudriliuseugeniae was predominantly highest in both topsoil (38.4%) and subsoil (27.1%). The total population of individual species of earthworm was least in GS in the topsoil (11.9%) and subsoil (8.4%). A total of 40 different species of TS was recorded, of which 55.5% were recorded in FWS, while RF was significantly highest in the species diversity(0.5971). Regression analysis revealed that Ri, IN, DBH, Di, and De of trees explained 65.9% of the variability of EA in the topsoil, while 46.9 % of the variability of earthworm abundance was explained by the floristic parameters in the subsoil.Similarly, correlation statistics revealed that in the topsoil, EA is positively and significantly correlated with Ri (r=0.35; p<0.05), IN (r=0.523; p<0.05) and De (r=0.469; p<0.05) while DBH was negatively and significantly correlated with earthworm abundance (r=-0.437; p<0.05). In the subsoil, only Ri and DBH correlated significantly with EA. The study concluded that EA in the study locations was highly influenced by tree growth species especially Ri, IN, DBH, Di, and De. The study recommended that the TSabundance should be improved in the study locations to ensure the survival of earthworms for ecosystem functions.Keywords: interactions, earthworm abundance, tree growth, ecological zones, western niger delta
Procedia PDF Downloads 1028514 Solid Phase Micro-Extraction/Gas Chromatography-Mass Spectrometry Study of Volatile Compounds from Strawberry Tree and Autumn Heather Honeys
Authors: Marinos Xagoraris, Elisavet Lazarou, Eleftherios Alissandrakis, Christos S. Pappas, Petros A. Tarantilis
Abstract:
Strawberry tree (Arbutus unedo L.) and autumn heather (Erica manipuliflora Salisb.) are important beekeeping plants of Greece. Six monofloral honeys (four strawberry tree, two autumn heather) were analyzed by means of Solid Phase Micro-Extraction (SPME, 60 min, 60 oC) followed by Gas Chromatography coupled to Mass Spectrometry (GC-MS) for the purpose of assessing the botanical origin. A Divinylbenzene/Carboxen/Polydimethylsiloxane (DVB/CAR/PDMS) fiber was employed, and benzophenone was used as internal standard. The volatile compounds with higher concentration (μg/ g of honey expressed as benzophenone) from strawberry tree honey samples, were α-isophorone (2.50-8.12); 3,4,5-trimethyl-phenol (0.20-4.62); 2-hydroxy-isophorone (0.06-0.53); 4-oxoisophorone (0.38-0.46); and β-isophorone (0.02-0.43). Regarding heather honey samples, the most abundant compounds were 1-methoxy-4-propyl-benzene (1.22-1.40); p-anisaldehyde (0.97-1.28); p-anisic acid (0.35-0.58); 2-furaldehyde (0.52-0.57); and benzaldehyde (0.41-0.56). Norisoprenoids are potent floral markers for strawberry-tree honey. β-isophorone is found exclusively in the volatile fraction of this type of honey, while also α-isophorone, 4-oxoisophorone and 2-hydroxy-isophorone could be considered as additional marker compounds. The analysis of autumn heather honey revealed that phenolic compounds are the most abundant and p-anisaldehyde; 1-methoxy-4-propyl-benzene; and p-anisic acid could serve as potent marker compounds. In conclusion, marker compounds for the determination of the botanical origin for these honeys could be identified as several norisoprenoids and phenolic components were found exclusively or in higher concentrations compared to common Greek honey varieties.Keywords: SPME/GC-MS, volatile compounds, heather honey, strawberry tree honey
Procedia PDF Downloads 2008513 A Review of Data Visualization Best Practices: Lessons for Open Government Data Portals
Authors: Bahareh Ansari
Abstract:
Background: The Open Government Data (OGD) movement in the last decade has encouraged many government organizations around the world to make their data publicly available to advance democratic processes. But current open data platforms have not yet reached to their full potential in supporting all interested parties. To make the data useful and understandable for everyone, scholars suggested that opening the data should be supplemented by visualization. However, different visualizations of the same information can dramatically change an individual’s cognitive and emotional experience in working with the data. This study reviews the data visualization literature to create a list of the methods empirically tested to enhance users’ performance and experience in working with a visualization tool. This list can be used in evaluating the OGD visualization practices and informing the future open data initiatives. Methods: Previous reviews of visualization literature categorized the visualization outcomes into four categories including recall/memorability, insight/comprehension, engagement, and enjoyment. To identify the papers, a search for these outcomes was conducted in the abstract of the publications of top-tier visualization venues including IEEE Transactions for Visualization and Computer Graphics, Computer Graphics, and proceedings of the CHI Conference on Human Factors in Computing Systems. The search results are complemented with a search in the references of the identified articles, and a search for 'open data visualization,' and 'visualization evaluation' keywords in the IEEE explore and ACM digital libraries. Articles are included if they provide empirical evidence through conducting controlled user experiments, or provide a review of these empirical studies. The qualitative synthesis of the studies focuses on identification and classifying the methods, and the conditions under which they are examined to positively affect the visualization outcomes. Findings: The keyword search yields 760 studies, of which 30 are included after the title/abstract review. The classification of the included articles shows five distinct methods: interactive design, aesthetic (artistic) style, storytelling, decorative elements that do not provide extra information including text, image, and embellishment on the graphs), and animation. Studies on decorative elements show consistency on the positive effects of these elements on user engagement and recall but are less consistent in their examination of the user performance. This inconsistency could be attributable to the particular data type or specific design method used in each study. The interactive design studies are consistent in their findings of the positive effect on the outcomes. Storytelling studies show some inconsistencies regarding the design effect on user engagement, enjoyment, recall, and performance, which could be indicative of the specific conditions required for the use of this method. Last two methods, aesthetics and animation, have been less frequent in the included articles, and provide consistent positive results on some of the outcomes. Implications for e-government: Review of the visualization best-practice methods show that each of these methods is beneficial under specific conditions. By using these methods in a potentially beneficial condition, OGD practices can promote a wide range of individuals to involve and work with the government data and ultimately engage in government policy-making procedures.Keywords: best practices, data visualization, literature review, open government data
Procedia PDF Downloads 1078512 A Distinct Method Based on Mamba-Unet for Brain Tumor Image Segmentation
Authors: Djallel Bouamama, Yasser R. Haddadi
Abstract:
Accurate brain tumor segmentation is crucial for diagnosis and treatment planning, yet it remains a challenging task due to the variability in tumor shapes and intensities. This paper introduces a distinct approach to brain tumor image segmentation by leveraging an advanced architecture known as Mamba-Unet. Building on the well-established U-Net framework, Mamba-Unet incorporates distinct design enhancements to improve segmentation performance. Our proposed method integrates a multi-scale attention mechanism and a hybrid loss function to effectively capture fine-grained details and contextual information in brain MRI scans. We demonstrate that Mamba-Unet significantly enhances segmentation accuracy compared to conventional U-Net models by utilizing a comprehensive dataset of annotated brain MRI scans. Quantitative evaluations reveal that Mamba-Unet surpasses traditional U-Net architectures and other contemporary segmentation models regarding Dice coefficient, sensitivity, and specificity. The improvements are attributed to the method's ability to manage class imbalance better and resolve complex tumor boundaries. This work advances the state-of-the-art in brain tumor segmentation and holds promise for improving clinical workflows and patient outcomes through more precise and reliable tumor detection.Keywords: brain tumor classification, image segmentation, CNN, U-NET
Procedia PDF Downloads 418511 Random Subspace Neural Classifier for Meteor Recognition in the Night Sky
Authors: Carlos Vera, Tetyana Baydyk, Ernst Kussul, Graciela Velasco, Miguel Aparicio
Abstract:
This article describes the Random Subspace Neural Classifier (RSC) for the recognition of meteors in the night sky. We used images of meteors entering the atmosphere at night between 8:00 p.m.-5: 00 a.m. The objective of this project is to classify meteor and star images (with stars as the image background). The monitoring of the sky and the classification of meteors are made for future applications by scientists. The image database was collected from different websites. We worked with RGB-type images with dimensions of 220x220 pixels stored in the BitMap Protocol (BMP) format. Subsequent window scanning and processing were carried out for each image. The scan window where the characteristics were extracted had the size of 20x20 pixels with a scanning step size of 10 pixels. Brightness, contrast and contour orientation histograms were used as inputs for the RSC. The RSC worked with two classes and classified into: 1) with meteors and 2) without meteors. Different tests were carried out by varying the number of training cycles and the number of images for training and recognition. The percentage error for the neural classifier was calculated. The results show a good RSC classifier response with 89% correct recognition. The results of these experiments are presented and discussed.Keywords: contour orientation histogram, meteors, night sky, RSC neural classifier, stars
Procedia PDF Downloads 1408510 Deep Learning-Based Classification of 3D CT Scans with Real Clinical Data; Impact of Image format
Authors: Maryam Fallahpoor, Biswajeet Pradhan
Abstract:
Background: Artificial intelligence (AI) serves as a valuable tool in mitigating the scarcity of human resources required for the evaluation and categorization of vast quantities of medical imaging data. When AI operates with optimal precision, it minimizes the demand for human interpretations and, thereby, reduces the burden on radiologists. Among various AI approaches, deep learning (DL) stands out as it obviates the need for feature extraction, a process that can impede classification, especially with intricate datasets. The advent of DL models has ushered in a new era in medical imaging, particularly in the context of COVID-19 detection. Traditional 2D imaging techniques exhibit limitations when applied to volumetric data, such as Computed Tomography (CT) scans. Medical images predominantly exist in one of two formats: neuroimaging informatics technology initiative (NIfTI) and digital imaging and communications in medicine (DICOM). Purpose: This study aims to employ DL for the classification of COVID-19-infected pulmonary patients and normal cases based on 3D CT scans while investigating the impact of image format. Material and Methods: The dataset used for model training and testing consisted of 1245 patients from IranMehr Hospital. All scans shared a matrix size of 512 × 512, although they exhibited varying slice numbers. Consequently, after loading the DICOM CT scans, image resampling and interpolation were performed to standardize the slice count. All images underwent cropping and resampling, resulting in uniform dimensions of 128 × 128 × 60. Resolution uniformity was achieved through resampling to 1 mm × 1 mm × 1 mm, and image intensities were confined to the range of (−1000, 400) Hounsfield units (HU). For classification purposes, positive pulmonary COVID-19 involvement was designated as 1, while normal images were assigned a value of 0. Subsequently, a U-net-based lung segmentation module was applied to obtain 3D segmented lung regions. The pre-processing stage included normalization, zero-centering, and shuffling. Four distinct 3D CNN models (ResNet152, ResNet50, DensNet169, and DensNet201) were employed in this study. Results: The findings revealed that the segmentation technique yielded superior results for DICOM images, which could be attributed to the potential loss of information during the conversion of original DICOM images to NIFTI format. Notably, ResNet152 and ResNet50 exhibited the highest accuracy at 90.0%, and the same models achieved the best F1 score at 87%. ResNet152 also secured the highest Area under the Curve (AUC) at 0.932. Regarding sensitivity and specificity, DensNet201 achieved the highest values at 93% and 96%, respectively. Conclusion: This study underscores the capacity of deep learning to classify COVID-19 pulmonary involvement using real 3D hospital data. The results underscore the significance of employing DICOM format 3D CT images alongside appropriate pre-processing techniques when training DL models for COVID-19 detection. This approach enhances the accuracy and reliability of diagnostic systems for COVID-19 detection.Keywords: deep learning, COVID-19 detection, NIFTI format, DICOM format
Procedia PDF Downloads 898509 ACOPIN: An ACO Algorithm with TSP Approach for Clustering Proteins in Protein Interaction Networks
Authors: Jamaludin Sallim, Rozlina Mohamed, Roslina Abdul Hamid
Abstract:
In this paper, we proposed an Ant Colony Optimization (ACO) algorithm together with Traveling Salesman Problem (TSP) approach to investigate the clustering problem in Protein Interaction Networks (PIN). We named this combination as ACOPIN. The purpose of this work is two-fold. First, to test the efficacy of ACO in clustering PIN and second, to propose the simple generalization of the ACO algorithm that might allow its application in clustering proteins in PIN. We split this paper to three main sections. First, we describe the PIN and clustering proteins in PIN. Second, we discuss the steps involved in each phase of ACO algorithm. Finally, we present some results of the investigation with the clustering patterns.Keywords: ant colony optimization algorithm, searching algorithm, protein functional module, protein interaction network
Procedia PDF Downloads 6138508 Secure Transfer of Medical Images Using Hybrid Encryption
Authors: Boukhatem Mohamed Belkaid, Lahdi Mourad
Abstract:
In this paper, we propose a new encryption system for security issues medical images. The hybrid encryption scheme is based on AES and RSA algorithms to validate the three security services are authentication, integrity, and confidentiality. Privacy is ensured by AES, authenticity is ensured by the RSA algorithm. Integrity is assured by the basic function of the correlation between adjacent pixels. Our system generates a unique password every new session of encryption, that will be used to encrypt each frame of the medical image basis to strengthen and ensure his safety. Several metrics have been used for various tests of our analysis. For the integrity test, we noticed the efficiencies of our system and how the imprint cryptographic changes at reception if a change affects the image in the transmission channel.Keywords: AES, RSA, integrity, confidentiality, authentication, medical images, encryption, decryption, key, correlation
Procedia PDF Downloads 4438507 Frequent Itemset Mining Using Rough-Sets
Authors: Usman Qamar, Younus Javed
Abstract:
Frequent pattern mining is the process of finding a pattern (a set of items, subsequences, substructures, etc.) that occurs frequently in a data set. It was proposed in the context of frequent itemsets and association rule mining. Frequent pattern mining is used to find inherent regularities in data. What products were often purchased together? Its applications include basket data analysis, cross-marketing, catalog design, sale campaign analysis, Web log (click stream) analysis, and DNA sequence analysis. However, one of the bottlenecks of frequent itemset mining is that as the data increase the amount of time and resources required to mining the data increases at an exponential rate. In this investigation a new algorithm is proposed which can be uses as a pre-processor for frequent itemset mining. FASTER (FeAture SelecTion using Entropy and Rough sets) is a hybrid pre-processor algorithm which utilizes entropy and rough-sets to carry out record reduction and feature (attribute) selection respectively. FASTER for frequent itemset mining can produce a speed up of 3.1 times when compared to original algorithm while maintaining an accuracy of 71%.Keywords: rough-sets, classification, feature selection, entropy, outliers, frequent itemset mining
Procedia PDF Downloads 4378506 A Greedy Alignment Algorithm Supporting Medication Reconciliation
Authors: David Tresner-Kirsch
Abstract:
Reconciling patient medication lists from multiple sources is a critical task supporting the safe delivery of patient care. Manual reconciliation is a time-consuming and error-prone process, and recently attempts have been made to develop efficiency- and safety-oriented automated support for professionals performing the task. An important capability of any such support system is automated alignment – finding which medications from a list correspond to which medications from a different source, regardless of misspellings, naming differences (e.g. brand name vs. generic), or changes in treatment (e.g. switching a patient from one antidepressant class to another). This work describes a new algorithmic solution to this alignment task, using a greedy matching approach based on string similarity, edit distances, concept extraction and normalization, and synonym search derived from the RxNorm nomenclature. The accuracy of this algorithm was evaluated against a gold-standard corpus of 681 medication records; this evaluation found that the algorithm predicted alignments with 99% precision and 91% recall. This performance is sufficient to support decision support applications for medication reconciliation.Keywords: clinical decision support, medication reconciliation, natural language processing, RxNorm
Procedia PDF Downloads 2868505 Hull Detection from Handwritten Digit Image
Authors: Sriraman Kothuri, Komal Teja Mattupalli
Abstract:
In this paper we proposed a novel algorithm for recognizing hulls in a hand written digits. This is an extension to the work on “Digit Recognition Using Freeman Chain code”. In order to find out the hulls in a user given digit it is necessary to follow three steps. Those are pre-processing, Boundary Extraction and at last apply the Hull Detection system in a way to attain the better results. The detection of Hull Regions is mainly intended to increase the machine learning capability in detection of characters or digits. This can also extend this in order to get the hull regions and their intensities in Black Holes in Space Exploration.Keywords: chain code, machine learning, hull regions, hull recognition system, SASK algorithm
Procedia PDF Downloads 4028504 Evaluating the Performance of Color Constancy Algorithm
Authors: Damanjit Kaur, Avani Bhatia
Abstract:
Color constancy is significant for human vision since color is a pictorial cue that helps in solving different visions tasks such as tracking, object recognition, or categorization. Therefore, several computational methods have tried to simulate human color constancy abilities to stabilize machine color representations. Two different kinds of methods have been used, i.e., normalization and constancy. While color normalization creates a new representation of the image by canceling illuminant effects, color constancy directly estimates the color of the illuminant in order to map the image colors to a canonical version. Color constancy is the capability to determine colors of objects independent of the color of the light source. This research work studies the most of the well-known color constancy algorithms like white point and gray world.Keywords: color constancy, gray world, white patch, modified white patch
Procedia PDF Downloads 3218503 Radar-Based Classification of Pedestrian and Dog Using High-Resolution Raw Range-Doppler Signatures
Authors: C. Mayr, J. Periya, A. Kariminezhad
Abstract:
In this paper, we developed a learning framework for the classification of vulnerable road users (VRU) by their range-Doppler signatures. The frequency-modulated continuous-wave (FMCW) radar raw data is first pre-processed to obtain robust object range-Doppler maps per coherent time interval. The complex-valued range-Doppler maps captured from our outdoor measurements are further fed into a convolutional neural network (CNN) to learn the classification. This CNN has gone through a hyperparameter optimization process for improved learning. By learning VRU range-Doppler signatures, the three classes 'pedestrian', 'dog', and 'noise' are classified with an average accuracy of almost 95%. Interestingly, this classification accuracy holds for a combined longitudinal and lateral object trajectories.Keywords: machine learning, radar, signal processing, autonomous driving
Procedia PDF Downloads 2468502 Using Satellite Images Datasets for Road Intersection Detection in Route Planning
Authors: Fatma El-Zahraa El-Taher, Ayman Taha, Jane Courtney, Susan Mckeever
Abstract:
Understanding road networks plays an important role in navigation applications such as self-driving vehicles and route planning for individual journeys. Intersections of roads are essential components of road networks. Understanding the features of an intersection, from a simple T-junction to larger multi-road junctions, is critical to decisions such as crossing roads or selecting the safest routes. The identification and profiling of intersections from satellite images is a challenging task. While deep learning approaches offer the state-of-the-art in image classification and detection, the availability of training datasets is a bottleneck in this approach. In this paper, a labelled satellite image dataset for the intersection recognition problem is presented. It consists of 14,692 satellite images of Washington DC, USA. To support other users of the dataset, an automated download and labelling script is provided for dataset replication. The challenges of construction and fine-grained feature labelling of a satellite image dataset is examined, including the issue of how to address features that are spread across multiple images. Finally, the accuracy of the detection of intersections in satellite images is evaluated.Keywords: satellite images, remote sensing images, data acquisition, autonomous vehicles
Procedia PDF Downloads 1478501 An Algorithm for the Map Labeling Problem with Two Kinds of Priorities
Authors: Noboru Abe, Yoshinori Amai, Toshinori Nakatake, Sumio Masuda, Kazuaki Yamaguchi
Abstract:
We consider the problem of placing labels of the points on a plane. For each point, its position, the size of its label and a priority are given. Moreover, several candidates of its label positions are prespecified, and each of such label positions is assigned a priority. The objective of our problem is to maximize the total sum of priorities of placed labels and their points. By refining a labeling algorithm that can use these priorities, we propose a new heuristic algorithm which is more suitable for treating the assigned priorities.Keywords: map labeling, greedy algorithm, heuristic algorithm, priority
Procedia PDF Downloads 4348500 Classification of Poverty Level Data in Indonesia Using the Naïve Bayes Method
Authors: Anung Style Bukhori, Ani Dijah Rahajoe
Abstract:
Poverty poses a significant challenge in Indonesia, requiring an effective analytical approach to understand and address this issue. In this research, we applied the Naïve Bayes classification method to examine and classify poverty data in Indonesia. The main focus is on classifying data using RapidMiner, a powerful data analysis platform. The analysis process involves data splitting to train and test the classification model. First, we collected and prepared a poverty dataset that includes various factors such as education, employment, and health..The experimental results indicate that the Naïve Bayes classification model can provide accurate predictions regarding the risk of poverty. The use of RapidMiner in the analysis process offers flexibility and efficiency in evaluating the model's performance. The classification produces several values to serve as the standard for classifying poverty data in Indonesia using Naive Bayes. The accuracy result obtained is 40.26%, with a moderate recall result of 35.94%, a high recall result of 63.16%, and a low recall result of 38.03%. The precision for the moderate class is 58.97%, for the high class is 17.39%, and for the low class is 58.70%. These results can be seen from the graph below.Keywords: poverty, classification, naïve bayes, Indonesia
Procedia PDF Downloads 618499 Drone Classification Using Classification Methods Using Conventional Model With Embedded Audio-Visual Features
Authors: Hrishi Rakshit, Pooneh Bagheri Zadeh
Abstract:
This paper investigates the performance of drone classification methods using conventional DCNN with different hyperparameters, when additional drone audio data is embedded in the dataset for training and further classification. In this paper, first a custom dataset is created using different images of drones from University of South California (USC) datasets and Leeds Beckett university datasets with embedded drone audio signal. The three well-known DCNN architectures namely, Resnet50, Darknet53 and Shufflenet are employed over the created dataset tuning their hyperparameters such as, learning rates, maximum epochs, Mini Batch size with different optimizers. Precision-Recall curves and F1 Scores-Threshold curves are used to evaluate the performance of the named classification algorithms. Experimental results show that Resnet50 has the highest efficiency compared to other DCNN methods.Keywords: drone classifications, deep convolutional neural network, hyperparameters, drone audio signal
Procedia PDF Downloads 1048498 Melaleuca alternifolia Fibre Composites: Effect of Different Type of Fibre on Mechanical and Physical Properties
Authors: Sahari Japar, Rodney Jammy, M. A. Maleque
Abstract:
The fabrication of melaleuca alternifolia fibre reinforced thermoplastic starch composites was successfully done. This paper aims to show the effect of melaleuca alternifolia fibres on mechanical and physical properties of composites by using starch as a matrix. The fibres were extracted from three different part i.e. tea tree trunk (TTT), tea tree bunch (TTB) and tea tree leaf (TTL) and combined with tapioca starch by casting method. All composites showed superior mechanical properties in comparison to TS. The addition of 5% (v/v) fibres as a filler to TS led to the improvement in young’s modulus by 350% for TTB/TS, 282% for TTT/TS and 220% for TTL/TS. The tensile strength also increased to 34.39% for TTL/TS, 82.80% for TTB/TS and 203.18% for TTT/TS respectively. The trend can be correlated to the amount of cellulose in the fibres. For physical properties, it can be seen that, with the addition of fibres, the water absorption and swelling of composites decreased. The addition of melaleuca alternifolia fibre improved mechanical and physical properties of thermoplastic starch composites.Keywords: melaleuca alternifolia, fibre, starch, mechanical, physical
Procedia PDF Downloads 4008497 Impact Analysis of Cultivation of Jatropha Tree on Fuel Prices and Environment
Authors: Saba Arif, Anam Nadeem, Roman Kalvin, Muzaffar Ali, Burhan Ali, Juntakan Taweekun
Abstract:
Globally transportation sector accounts for around 25% of energy demand and nearly 62% of oil consumed. Therefore, new energy sources are required to introduce for this huge demand replenishment of depleting conventional energy sources. Currently, biofuels such as Jatropha trees as an energy carrier for transportation sector are being utilized effectively round the globe. However, climate conditions at low altitudes with an average annual temperature above 20 degrees Celsius and rainfall of 300-1000mm are considered the most suitable environment for the efficient growth of Jatropha trees. The current study is providing a theoretical survey-based analysis to investigate the effect of rate of cultivation of jatropha trees on the reduction of fuel prices and its environmental benefits. The resulted study shows that jatropha tree’s 100 kg seeds give 80kg oil and the conversion process cost is very small as 890 PKR. Moreover, the extraction of oil from Jatropha tree is tax-free compared to other fuels. The analysis proved very essential for potential assessment of Jatropha regarding future energy fuel for transportation sector at global level. Additionally, it can be very beneficial for increment in the total amount of transportation fuel in Pakistan.Keywords: jatropha tree, environmental impact, energy contents, theoretical survey
Procedia PDF Downloads 2208496 Lifting Wavelet Transform and Singular Values Decomposition for Secure Image Watermarking
Authors: Siraa Ben Ftima, Mourad Talbi, Tahar Ezzedine
Abstract:
In this paper, we present a technique of secure watermarking of grayscale and color images. This technique consists in applying the Singular Value Decomposition (SVD) in LWT (Lifting Wavelet Transform) domain in order to insert the watermark image (grayscale) in the host image (grayscale or color image). It also uses signature in the embedding and extraction steps. The technique is applied on a number of grayscale and color images. The performance of this technique is proved by the PSNR (Pick Signal to Noise Ratio), the MSE (Mean Square Error) and the SSIM (structural similarity) computations.Keywords: lifting wavelet transform (LWT), sub-space vectorial decomposition, secure, image watermarking, watermark
Procedia PDF Downloads 2778495 Assessing the Theoretical Suitability of Sentinel-2 and Worldview-3 Data for Hydrocarbon Mapping of Spill Events, Using Hydrocarbon Spectral Slope Model
Authors: K. Tunde Olagunju, C. Scott Allen, Freek Van Der Meer
Abstract:
Identification of hydrocarbon oil in remote sensing images is often the first step in monitoring oil during spill events. Most remote sensing methods adopt techniques for hydrocarbon identification to achieve detection in order to model an appropriate cleanup program. Identification on optical sensors does not only allow for detection but also for characterization and quantification. Until recently, in optical remote sensing, quantification and characterization are only potentially possible using high-resolution laboratory and airborne imaging spectrometers (hyperspectral data). Unlike multispectral, hyperspectral data are not freely available, as this data category is mainly obtained via airborne survey at present. In this research, two (2) operational high-resolution multispectral satellites (WorldView-3 and Sentinel-2) are theoretically assessed for their suitability for hydrocarbon characterization, using the hydrocarbon spectral slope model (HYSS). This method utilized the two most persistent hydrocarbon diagnostic/absorption features at 1.73 µm and 2.30 µm for hydrocarbon mapping on multispectral data. In this research, spectra measurement of seven (7) different hydrocarbon oils (crude and refined oil) taken on ten (10) different substrates with the use of laboratory ASD Fieldspec were convolved to Sentinel-2 and WorldView-3 resolution, using their full width half maximum (FWHM) parameter. The resulting hydrocarbon slope values obtained from the studied samples enable clear qualitative discrimination of most hydrocarbons, despite the presence of different background substrates, particularly on WorldView-3. Due to close conformity of central wavelengths and narrow bandwidths to key hydrocarbon bands used in HYSS, the statistical significance for qualitative analysis on WorldView-3 sensors for all studied hydrocarbon oil returned with 95% confidence level (P-value ˂ 0.01), except for Diesel. Using multifactor analysis of variance (MANOVA), the discriminating power of HYSS is statistically significant for most hydrocarbon-substrate combinations on Sentinel-2 and WorldView-3 FWHM, revealing the potential of these two operational multispectral sensors as rapid response tools for hydrocarbon mapping. One notable exception is highly transmissive hydrocarbons on Sentinel-2 data due to the non-conformity of spectral bands with key hydrocarbon absorptions and the relatively coarse bandwidth (> 100 nm).Keywords: hydrocarbon, oil spill, remote sensing, hyperspectral, multispectral, hydrocarbon-substrate combination, Sentinel-2, WorldView-3
Procedia PDF Downloads 2168494 Application of Fuzzy Approach to the Vibration Fault Diagnosis
Authors: Jalel Khelil
Abstract:
In order to improve reliability of Gas Turbine machine especially its generator equipment, a fault diagnosis system based on fuzzy approach is proposed. Three various methods namely K-NN (K-nearest neighbors), F-KNN (Fuzzy K-nearest neighbors) and FNM (Fuzzy nearest mean) are adopted to provide the measurement of relative strength of vibration defaults. Both applications consist of two major steps: Feature extraction and default classification. 09 statistical features are extracted from vibration signals. 03 different classes are used in this study which describes vibrations condition: Normal, unbalance defect, and misalignment defect. The use of the fuzzy approaches and the classification results are discussed. Results show that these approaches yield high successful rates of vibration default classification.Keywords: fault diagnosis, fuzzy classification k-nearest neighbor, vibration
Procedia PDF Downloads 4688493 A Relative Analysis of Carbon and Dust Uptake by Important Tree Species in Tehran, Iran
Authors: Sahar Elkaee Behjati
Abstract:
Air pollution, particularly with dust, is one of the biggest issues Tehran is dealing with, and the city's green space which consists of trees has a critical role in absorption of it. The question this study aimed to investigate was which tree species the highest uptake capacity of the dust and carbon have suspended in the air. On this basis, 30 samples of trees from two different districts in Tehran were collected, and after washing and centrifuging, the samples were oven dried. The results of the study revealed that Ulmus minor had the highest amount of deposited dust in both districts. In addition, it was found that in Chamran district Ailanthus altissima and in Gandi district Ulmus minor has had the highest absorption of deposited carbon. Therefore, it could be argued that decision making on the selection of species for urban green spaces should take the above-mentioned parameters into account.Keywords: dust, leaves, uptake total carbon, Tehran, tree species
Procedia PDF Downloads 1418492 Use of Fault Tree Analysis for Technical Assessment of Waste-to-Energy Plants
Authors: Ying-Chu Chen
Abstract:
Waste to energy (WTE) technology is becoming increasingly important throughout the world. There are 24 WTE plants in operation in Taiwan that might be ranked the top in density (number of MSW incinerators/area) in the world. Many problems exist in WTE plants, such as low-quality construction, leakage of pipelines, irregular feedings, and lack of maintenance. These problems should be identified and analyzed for effective implementation and efficient operation of WTE plants. This research applies a fault tree analysis (FTA) to identify failures and evaluate their effects on the operation of WTE plants from a technical point of view. Five subsystems of a WTE plant were defined, including loading system, incineration system, effluent disposal system, structural components, and control system. This research results proved that FTA is suitable for WTE evaluation and is an effective analysis tool for technical evaluation in the field of WTE technology.Keywords: delphi method, fault tree approach, municipal solid waste, waste to energy, WTE
Procedia PDF Downloads 5678491 Cost Effective Real-Time Image Processing Based Optical Mark Reader
Authors: Amit Kumar, Himanshu Singal, Arnav Bhavsar
Abstract:
In this modern era of automation, most of the academic exams and competitive exams are Multiple Choice Questions (MCQ). The responses of these MCQ based exams are recorded in the Optical Mark Reader (OMR) sheet. Evaluation of the OMR sheet requires separate specialized machines for scanning and marking. The sheets used by these machines are special and costs more than a normal sheet. Available process is non-economical and dependent on paper thickness, scanning quality, paper orientation, special hardware and customized software. This study tries to tackle the problem of evaluating the OMR sheet without any special hardware and making the whole process economical. We propose an image processing based algorithm which can be used to read and evaluate the scanned OMR sheets with no special hardware required. It will eliminate the use of special OMR sheet. Responses recorded in normal sheet is enough for evaluation. The proposed system takes care of color, brightness, rotation, little imperfections in the OMR sheet images.Keywords: OMR, image processing, hough circle trans-form, interpolation, detection, binary thresholding
Procedia PDF Downloads 1758490 MhAGCN: Multi-Head Attention Graph Convolutional Network for Web Services Classification
Authors: Bing Li, Zhi Li, Yilong Yang
Abstract:
Web classification can promote the quality of service discovery and management in the service repository. It is widely used to locate developers desired services. Although traditional classification methods based on supervised learning models can achieve classification tasks, developers need to manually mark web services, and the quality of these tags may not be enough to establish an accurate classifier for service classification. With the doubling of the number of web services, the manual tagging method has become unrealistic. In recent years, the attention mechanism has made remarkable progress in the field of deep learning, and its huge potential has been fully demonstrated in various fields. This paper designs a multi-head attention graph convolutional network (MHAGCN) service classification method, which can assign different weights to the neighborhood nodes without complicated matrix operations or relying on understanding the entire graph structure. The framework combines the advantages of the attention mechanism and graph convolutional neural network. It can classify web services through automatic feature extraction. The comprehensive experimental results on a real dataset not only show the superior performance of the proposed model over the existing models but also demonstrate its potentially good interpretability for graph analysis.Keywords: attention mechanism, graph convolutional network, interpretability, service classification, service discovery
Procedia PDF Downloads 1378489 Musical Instruments Classification Using Machine Learning Techniques
Authors: Bhalke D. G., Bormane D. S., Kharate G. K.
Abstract:
This paper presents classification of musical instrument using machine learning techniques. The classification has been carried out using temporal, spectral, cepstral and wavelet features. Detail feature analysis is carried out using separate and combined features. Further, instrument model has been developed using K-Nearest Neighbor and Support Vector Machine (SVM). Benchmarked McGill university database has been used to test the performance of the system. Experimental result shows that SVM performs better as compared to KNN classifier.Keywords: feature extraction, SVM, KNN, musical instruments
Procedia PDF Downloads 480