Search results for: cleaning machine
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3073

Search results for: cleaning machine

2263 Internet of Things Networks: Denial of Service Detection in Constrained Application Protocol Using Machine Learning Algorithm

Authors: Adamu Abdullahi, On Francisca, Saidu Isah Rambo, G. N. Obunadike, D. T. Chinyio

Abstract:

The paper discusses the potential threat of Denial of Service (DoS) attacks in the Internet of Things (IoT) networks on constrained application protocols (CoAP). As billions of IoT devices are expected to be connected to the internet in the coming years, the security of these devices is vulnerable to attacks, disrupting their functioning. This research aims to tackle this issue by applying mixed methods of qualitative and quantitative for feature selection, extraction, and cluster algorithms to detect DoS attacks in the Constrained Application Protocol (CoAP) using the Machine Learning Algorithm (MLA). The main objective of the research is to enhance the security scheme for CoAP in the IoT environment by analyzing the nature of DoS attacks and identifying a new set of features for detecting them in the IoT network environment. The aim is to demonstrate the effectiveness of the MLA in detecting DoS attacks and compare it with conventional intrusion detection systems for securing the CoAP in the IoT environment. Findings: The research identifies the appropriate node to detect DoS attacks in the IoT network environment and demonstrates how to detect the attacks through the MLA. The accuracy detection in both classification and network simulation environments shows that the k-means algorithm scored the highest percentage in the training and testing of the evaluation. The network simulation platform also achieved the highest percentage of 99.93% in overall accuracy. This work reviews conventional intrusion detection systems for securing the CoAP in the IoT environment. The DoS security issues associated with the CoAP are discussed.

Keywords: algorithm, CoAP, DoS, IoT, machine learning

Procedia PDF Downloads 58
2262 Basic Characteristics and Prospects of Synchronized Stir Welding

Authors: Shoji Matsumoto

Abstract:

Friction Stir Welding (FSW) has been widely used in the automotive, aerospace, and high-tech industries due to its superior mechanical properties after welding. However, when it becomes a matter to perform a high-quality joint using FSW, it is necessary to secure an advanced tilt angle (usually 1 to 5 degrees) using a dedicated FSW machine and to use a joint structure and a restraining jig that can withstand the tool pressure applied during the jointing process using a highly rigid processing machine. One issue that has become a challenge in this process is ‘productivity and versatility’. To solve this problem, we have conducted research and development of multi-functioning machines and robotics with FSW tools, which combine cutting/milling and FSW functions as one in recent years. However, the narrow process window makes it prone to welding defects and lacks repeatability, which makes a limitation for FSW its use in the fields where precisions required. Another reason why FSW machines are not widely used in the world is because of the matter of very high cost of ownership.

Keywords: synchronized, stir, welding, friction, traveling speed, synchronized stir welding, friction stir welding

Procedia PDF Downloads 32
2261 Does sustainability disclosure improve analysts’ forecast accuracy Evidence from European banks

Authors: Albert Acheampong, Tamer Elshandidy

Abstract:

We investigate the extent to which sustainability disclosure from the narrative section of European banks’ annual reports improves analyst forecast accuracy. We capture sustainability disclosure using a machine learning approach and use forecast error to proxy analyst forecast accuracy. Our results suggest that sustainability disclosure significantly improves analyst forecast accuracy by reducing the forecast error. In a further analysis, we also find that the induction of Directive 2014/95/European Union (EU) is associated with increased disclosure content, which then reduces forecast error. Collectively, our results suggest that sustainability disclosure improves forecast accuracy, and the induction of the new EU directive strengthens this improvement. These results hold after several further and robustness analyses. Our findings have implications for market participants and policymakers.

Keywords: sustainability disclosure, machine learning, analyst forecast accuracy, forecast error, European banks, EU directive

Procedia PDF Downloads 56
2260 Unveiling Comorbidities in Irritable Bowel Syndrome: A UK BioBank Study utilizing Supervised Machine Learning

Authors: Uswah Ahmad Khan, Muhammad Moazam Fraz, Humayoon Shafique Satti, Qasim Aziz

Abstract:

Approximately 10-14% of the global population experiences a functional disorder known as irritable bowel syndrome (IBS). The disorder is defined by persistent abdominal pain and an irregular bowel pattern. IBS significantly impairs work productivity and disrupts patients' daily lives and activities. Although IBS is widespread, there is still an incomplete understanding of its underlying pathophysiology. This study aims to help characterize the phenotype of IBS patients by differentiating the comorbidities found in IBS patients from those in non-IBS patients using machine learning algorithms. In this study, we extracted samples coding for IBS from the UK BioBank cohort and randomly selected patients without a code for IBS to create a total sample size of 18,000. We selected the codes for comorbidities of these cases from 2 years before and after their IBS diagnosis and compared them to the comorbidities in the non-IBS cohort. Machine learning models, including Decision Trees, Gradient Boosting, Support Vector Machine (SVM), AdaBoost, Logistic Regression, and XGBoost, were employed to assess their accuracy in predicting IBS. The most accurate model was then chosen to identify the features associated with IBS. In our case, we used XGBoost feature importance as a feature selection method. We applied different models to the top 10% of features, which numbered 50. Gradient Boosting, Logistic Regression and XGBoost algorithms yielded a diagnosis of IBS with an optimal accuracy of 71.08%, 71.427%, and 71.53%, respectively. Among the comorbidities most closely associated with IBS included gut diseases (Haemorrhoids, diverticular diseases), atopic conditions(asthma), and psychiatric comorbidities (depressive episodes or disorder, anxiety). This finding emphasizes the need for a comprehensive approach when evaluating the phenotype of IBS, suggesting the possibility of identifying new subsets of IBS rather than relying solely on the conventional classification based on stool type. Additionally, our study demonstrates the potential of machine learning algorithms in predicting the development of IBS based on comorbidities, which may enhance diagnosis and facilitate better management of modifiable risk factors for IBS. Further research is necessary to confirm our findings and establish cause and effect. Alternative feature selection methods and even larger and more diverse datasets may lead to more accurate classification models. Despite these limitations, our findings highlight the effectiveness of Logistic Regression and XGBoost in predicting IBS diagnosis.

Keywords: comorbidities, disease association, irritable bowel syndrome (IBS), predictive analytics

Procedia PDF Downloads 100
2259 A Comparative Study for Various Techniques Using WEKA for Red Blood Cells Classification

Authors: Jameela Ali, Hamid A. Jalab, Loay E. George, Abdul Rahim Ahmad, Azizah Suliman, Karim Al-Jashamy

Abstract:

Red blood cells (RBC) are the most common types of blood cells and are the most intensively studied in cell biology. The lack of RBCs is a condition in which the amount of hemoglobin level is lower than normal and is referred to as “anemia”. Abnormalities in RBCs will affect the exchange of oxygen. This paper presents a comparative study for various techniques for classifyig the red blood cells as normal, or abnormal (anemic) using WEKA. WEKA is an open source consists of different machine learning algorithms for data mining applications. The algorithm tested are Radial Basis Function neural network, Support vector machine, and K-Nearest Neighbors algorithm. Two sets of combined features were utilized for classification of blood cells images. The first set, exclusively consist of geometrical features, was used to identify whether the tested blood cell has a spherical shape or non-spherical cells. While the second set, consist mainly of textural features was used to recognize the types of the spherical cells. We have provided an evaluation based on applying these classification methods to our RBCs image dataset which were obtained from Serdang Hospital-Malaysia, and measuring the accuracy of test results. The best achieved classification rates are 97%, 98%, and 79% for Support vector machines, Radial Basis Function neural network, and K-Nearest Neighbors algorithm respectively

Keywords: red blood cells, classification, radial basis function neural networks, suport vector machine, k-nearest neighbors algorithm

Procedia PDF Downloads 468
2258 Microbial and SARS-CoV-2 Efficiency Analysis of Froumann HEPA Filter Air Cleaner Brand

Authors: Serap Gedikli, Hakan Çakmak, M. Buğra Güldiken, Duygu Yalnızoğlu

Abstract:

Air, which is necessary for living things to survive; while it carries some useful substances in it, it can also carry foreign particles of different sizes that may be harmful to the health. All airborne organic substances of biological origin, including bacteria, fungi, fungal spores, viruses, pollen, and their components, are called "bioaerosols". Nowadays, everyone spends most of their time in closed areas such as home, workplace, school, etc. Although it is known that outdoor air pollution affects health, it is not known that indoor air pollution has harmful effects in terms of health. In this study, indoor air microbial load and SARS-CoV-2 virus cleaning efficiency of Froumann brand air cleaners were studied. This work in 300 m³, 600 m³, and 1000 m³ completely closed areas without any air circulation with Froumann N80, N90, and N100 air-cleaning devices. Analyzes were performed for both areas at 60 minutes before and after the device was operated using a particle measuring device (Particles Plus 7302) and an air sampler (Mas-100 ECO). The measurements were taken by placing the test equipment 1.5-2 m away from the air cleaner. At the same time, the efficiency of the HEPA filter was evaluated by taking samples from the air outlet point of the HEPA filter using the air sampling device (Mas-100 ECO) after the device was started. Nutrient agar and malt agar are used as total mesophilic bacteria and total fungi. The number of colony-forming units per m³ (cfu/m³) was calculated by counting colonies in Petri dishes after incubation for 48 hours at 37°C for bacteria and 72 hours at 30°C for fungi. The change in the number of colonies and the decrease in the microbial load was calculated as a percentage value. SARS-CoV-2 activity analysis studies were carried out by İnönü University Microbiology Department in accordance with the World Health Organization regulations. Finally, the HEPA filter in the devices used was taken and kept under a certain temperature and humidity, and the change in the microbial load on it was monitored over a 6-month period. At the end of the studies, a 91%-94% reduction was determined in the total mesophilic bacteria count of Frouman brand N80, N90, and N100 model air cleaners. A decrease of 94%-96% was detected in the total number of yeast/molds. HEPA filter efficiency was evaluated, and at the end of the analysis, 98% of the bacterial load and approximately 100% of yeast/mold load at the HEPA filter air outlet point were decreased. According to the SARS- CoV-2 analysis results, when the device is operating at the medium airflow level 3, it can filter virus-carrying aerosols by 99%. As a result, it was determined that the Froumann model air cleaner was effective in controlling and reducing the microbial load in the indoor air.

Keywords: HEPA filter, indoor air quality, microbial load, SARS-CoV-2

Procedia PDF Downloads 191
2257 Unlocking Green Hydrogen Potential: A Machine Learning-Based Assessment

Authors: Said Alshukri, Mazhar Hussain Malik

Abstract:

Green hydrogen is hydrogen produced using renewable energy sources. In the last few years, Oman aimed to reduce its dependency on fossil fuels. Recently, the hydrogen economy has become a global trend, and many countries have started to investigate the feasibility of implementing this sector. Oman created an alliance to establish the policy and rules for this sector. With motivation coming from both global and local interest in green hydrogen, this paper investigates the potential of producing hydrogen from wind and solar energies in three different locations in Oman, namely Duqm, Salalah, and Sohar. By using machine learning-based software “WEKA” and local metrological data, the project was designed to figure out which location has the highest wind and solar energy potential. First, various supervised models were tested to obtain their prediction accuracy, and it was found that the Random Forest (RF) model has the best prediction performance. The RF model was applied to 2021 metrological data for each location, and the results indicated that Duqm has the highest wind and solar energy potential. The system of one wind turbine in Duqm can produce 8335 MWh/year, which could be utilized in the water electrolysis process to produce 88847 kg of hydrogen mass, while a solar system consisting of 2820 solar cells is estimated to produce 1666.223 MWh/ year which is capable of producing 177591 kg of hydrogen mass.

Keywords: green hydrogen, machine learning, wind and solar energies, WEKA, supervised models, random forest

Procedia PDF Downloads 62
2256 Wireless Sensor Anomaly Detection Using Soft Computing

Authors: Mouhammd Alkasassbeh, Alaa Lasasmeh

Abstract:

We live in an era of rapid development as a result of significant scientific growth. Like other technologies, wireless sensor networks (WSNs) are playing one of the main roles. Based on WSNs, ZigBee adds many features to devices, such as minimum cost and power consumption, and increasing the range and connect ability of sensor nodes. ZigBee technology has come to be used in various fields, including science, engineering, and networks, and even in medicinal aspects of intelligence building. In this work, we generated two main datasets, the first being based on tree topology and the second on star topology. The datasets were evaluated by three machine learning (ML) algorithms: J48, meta.j48 and multilayer perceptron (MLP). Each topology was classified into normal and abnormal (attack) network traffic. The dataset used in our work contained simulated data from network simulation 2 (NS2). In each database, the Bayesian network meta.j48 classifier achieved the highest accuracy level among other classifiers, of 99.7% and 99.2% respectively.

Keywords: IDS, Machine learning, WSN, ZigBee technology

Procedia PDF Downloads 532
2255 Naïve Bayes: A Classical Approach for the Epileptic Seizures Recognition

Authors: Bhaveek Maini, Sanjay Dhanka, Surita Maini

Abstract:

Electroencephalography (EEG) is used to classify several epileptic seizures worldwide. It is a very crucial task for the neurologist to identify the epileptic seizure with manual EEG analysis, as it takes lots of effort and time. Human error is always at high risk in EEG, as acquiring signals needs manual intervention. Disease diagnosis using machine learning (ML) has continuously been explored since its inception. Moreover, where a large number of datasets have to be analyzed, ML is acting as a boon for doctors. In this research paper, authors proposed two different ML models, i.e., logistic regression (LR) and Naïve Bayes (NB), to predict epileptic seizures based on general parameters. These two techniques are applied to the epileptic seizures recognition dataset, available on the UCI ML repository. The algorithms are implemented on an 80:20 train test ratio (80% for training and 20% for testing), and the performance of the model was validated by 10-fold cross-validation. The proposed study has claimed accuracy of 81.87% and 95.49% for LR and NB, respectively.

Keywords: epileptic seizure recognition, logistic regression, Naïve Bayes, machine learning

Procedia PDF Downloads 46
2254 Early Stage Suicide Ideation Detection Using Supervised Machine Learning and Neural Network Classifier

Authors: Devendra Kr Tayal, Vrinda Gupta, Aastha Bansal, Khushi Singh, Sristi Sharma, Hunny Gaur

Abstract:

In today's world, suicide is a serious problem. In order to save lives, early suicide attempt detection and prevention should be addressed. A good number of at-risk people utilize social media platforms to talk about their issues or find knowledge on related chores. Twitter and Reddit are two of the most common platforms that are used for expressing oneself. Extensive research has already been done in this field. Through supervised classification techniques like Nave Bayes, Bernoulli Nave Bayes, and Multiple Layer Perceptron on a Reddit dataset, we demonstrate the early recognition of suicidal ideation. We also performed comparative analysis on these approaches and used accuracy, recall score, F1 score, and precision score for analysis.

Keywords: machine learning, suicide ideation detection, supervised classification, natural language processing

Procedia PDF Downloads 81
2253 Self-Directed-Car on GT Road: Grand Trunk Road

Authors: Rameez Ahmad, Aqib Mehmood, Imran Khan

Abstract:

Self-directed car (SDC) that can drive itself from one fact to another without support from a driver. Certain trust that self-directed car obligate the probable to transform the transportation manufacturing while essentially removing coincidences, and cleaning up the environment. This study realizes the effects that SDC (also called a self-driving, driver or robotic) vehicle travel demands and ride scheme is likely to have. Without the typical obstacles that allows detection of a audio vision based hardware and software construction (It (SDC) and cost benefits, the vehicle technologies, Gold (Generic Obstacle and Lane Detection) to a knowledge-based system to predict their potential and consider the shape, color, or balance) and an organized environment with colored lane patterns, lane position ban. Discovery the problematic consequence of (SDC) on GT (grand trunk road) road and brand the car further effectual.

Keywords: SDC, gold, GT, knowledge-based system

Procedia PDF Downloads 355
2252 Theoretical and Experimental Analysis of End Milling Process with Multiple Finger Inserted Cutters

Authors: G. Krishna Mohana Rao, P. Ravi Kumar

Abstract:

Milling is the process of removing unwanted material with suitable tool. Even though the milling process is having wider application, the vibration of machine tool and work piece during the process produces chatter on the products. Various methods of preventing the chatter have been incorporated into machine tool systems. Damper is cut into equal number of parts. Each part is called as finger. Multiple fingers were inserted in the hollow portion of the shank to reduce tool vibrations. In the present work, nonlinear static and dynamic analysis of the damper inserted end milling cutter used to reduce the chatter was done. A comparison is made for the milling cutter with multiple dampers. Surface roughness was determined by machining with multiple finger inserted milling cutters.

Keywords: damping inserts, end milling, vibrations, nonlinear dynamic analysis, number of fingers

Procedia PDF Downloads 510
2251 Land Suitability Prediction Modelling for Agricultural Crops Using Machine Learning Approach: A Case Study of Khuzestan Province, Iran

Authors: Saba Gachpaz, Hamid Reza Heidari

Abstract:

The sharp increase in population growth leads to more pressure on agricultural areas to satisfy the food supply. To achieve this, more resources should be consumed and, besides other environmental concerns, highlight sustainable agricultural development. Land-use management is a crucial factor in obtaining optimum productivity. Machine learning is a widely used technique in the agricultural sector, from yield prediction to customer behavior. This method focuses on learning and provides patterns and correlations from our data set. In this study, nine physical control factors, namely, soil classification, electrical conductivity, normalized difference water index (NDWI), groundwater level, elevation, annual precipitation, pH of water, annual mean temperature, and slope in the alluvial plain in Khuzestan (an agricultural hotspot in Iran) are used to decide the best agricultural land use for both rainfed and irrigated agriculture for ten different crops. For this purpose, each variable was imported into Arc GIS, and a raster layer was obtained. In the next level, by using training samples, all layers were imported into the python environment. A random forest model was applied, and the weight of each variable was specified. In the final step, results were visualized using a digital elevation model, and the importance of all factors for each one of the crops was obtained. Our results show that despite 62% of the study area being allocated to agricultural purposes, only 42.9% of these areas can be defined as a suitable class for cultivation purposes.

Keywords: land suitability, machine learning, random forest, sustainable agriculture

Procedia PDF Downloads 68
2250 Using Geo-Statistical Techniques and Machine Learning Algorithms to Model the Spatiotemporal Heterogeneity of Land Surface Temperature and its Relationship with Land Use Land Cover

Authors: Javed Mallick

Abstract:

In metropolitan areas, rapid changes in land use and land cover (LULC) have ecological and environmental consequences. Saudi Arabia's cities have experienced tremendous urban growth since the 1990s, resulting in urban heat islands, groundwater depletion, air pollution, loss of ecosystem services, and so on. From 1990 to 2020, this study examines the variance and heterogeneity in land surface temperature (LST) caused by LULC changes in Abha-Khamis Mushyet, Saudi Arabia. LULC was mapped using the support vector machine (SVM). The mono-window algorithm was used to calculate the land surface temperature (LST). To identify LST clusters, the local indicator of spatial associations (LISA) model was applied to spatiotemporal LST maps. In addition, the parallel coordinate (PCP) method was used to investigate the relationship between LST clusters and urban biophysical variables as a proxy for LULC. According to LULC maps, urban areas increased by more than 330% between 1990 and 2018. Between 1990 and 2018, built-up areas had an 83.6% transitional probability. Furthermore, between 1990 and 2020, vegetation and agricultural land were converted into built-up areas at a rate of 17.9% and 21.8%, respectively. Uneven LULC changes in built-up areas result in more LST hotspots. LST hotspots were associated with high NDBI but not NDWI or NDVI. This study could assist policymakers in developing mitigation strategies for urban heat islands

Keywords: land use land cover mapping, land surface temperature, support vector machine, LISA model, parallel coordinate plot

Procedia PDF Downloads 62
2249 Power Angle Control Strategy of Virtual Synchronous Machine: A Novel Approach to Control Virtual Synchronous Machine

Authors: Shishir Lamichhane, Saurav Dulal, Bibek Gautam, Madan Thapa Magar, Indraman Tamrakar

Abstract:

Renewable energies such as wind turbines and solar photovoltaic have gained significance as a result of global environmental pollution and energy crises. These sources of energy are converted into electrical energy and delivered to end-users through the utility system. As a result of the widespread use of power electronics-based grid-interfacing technologies to accommodate renewable sources of energy, the prevalence of converters has expanded as well. As a result, the power system's rotating inertia is decreasing, endangering the utility grid's stability. The use of Virtual Synchronous Machine (VSM) technology has been proposed to overcome the grid stability problem due to low rotating inertia. The grid-connected inverter used in VSM can be controlled to emulate inertia, which replicates the external features of a synchronous generator. As a result, the rotating inertia is increased to support the power system's stability. A power angle control strategy is proposed in this paper and its model is simulated in MATLAB/Simulink to study the effects of parameter disturbances on the active power and frequency for a VSM. The system consists of a synchronous generator, which is modeled in such a way that the frequency drops to an unacceptable region during transient conditions due to a lack of inertia when VSM is not used. Then, the suggested model incorporating VSM emulates rotating inertia, injecting a controllable amount of energy into the grid during frequency transients to enhance transient stability.

Keywords: damping constant, inertia–constant, ROCOF, transient stability, distributed sources

Procedia PDF Downloads 198
2248 Theoretical and ML-Driven Identification of a Mispriced Credit Risk

Authors: Yuri Katz, Kun Liu, Arunram Atmacharan

Abstract:

Due to illiquidity, mispricing on Credit Markets is inevitable. This creates huge challenges to banks and investors as they seek to find new ways of risk valuation and portfolio management in a post-credit crisis world. Here, we analyze the difference in behavior of the spread-to-maturity in investment and high-yield categories of US corporate bonds between 2014 and 2023. Deviation from the theoretical dependency of this measure in the universe under study allows to identify multiple cases of mispriced credit risk. Remarkably, we observe mispriced bonds in both categories of credit ratings. This identification is supported by the application of the state-of-the-art machine learning model in more than 90% of cases. Noticeably, the ML-driven model-based forecasting of a category of bond’s credit ratings demonstrate an excellent out-of-sample accuracy (AUC = 98%). We believe that these results can augment conventional valuations of credit portfolios.

Keywords: credit risk, credit ratings, bond pricing, spread-to-maturity, machine learning

Procedia PDF Downloads 66
2247 A Risk Assessment Tool for the Contamination of Aflatoxins on Dried Figs Based on Machine Learning Algorithms

Authors: Kottaridi Klimentia, Demopoulos Vasilis, Sidiropoulos Anastasios, Ihara Diego, Nikolaidis Vasileios, Antonopoulos Dimitrios

Abstract:

Aflatoxins are highly poisonous and carcinogenic compounds produced by species of the genus Aspergillus spp. that can infect a variety of agricultural foods, including dried figs. Biological and environmental factors, such as population, pathogenicity, and aflatoxinogenic capacity of the strains, topography, soil, and climate parameters of the fig orchards, are believed to have a strong effect on aflatoxin levels. Existing methods for aflatoxin detection and measurement, such as high performance liquid chromatography (HPLC), and enzyme-linked immunosorbent assay (ELISA), can provide accurate results, but the procedures are usually time-consuming, sample-destructive, and expensive. Predicting aflatoxin levels prior to crop harvest is useful for minimizing the health and financial impact of a contaminated crop. Consequently, there is interest in developing a tool that predicts aflatoxin levels based on topography and soil analysis data of fig orchards. This paper describes the development of a risk assessment tool for the contamination of aflatoxin on dried figs, based on the location and altitude of the fig orchards, the population of the fungus Aspergillus spp. in the soil, and soil parameters such as pH, saturation percentage (SP), electrical conductivity (EC), organic matter, particle size analysis (sand, silt, clay), the concentration of the exchangeable cations (Ca, Mg, K, Na), extractable P, and trace of elements (B, Fe, Mn, Zn and Cu), by employing machine learning methods. In particular, our proposed method integrates three machine learning techniques, i.e., dimensionality reduction on the original dataset (principal component analysis), metric learning (Mahalanobis metric for clustering), and k-nearest neighbors learning algorithm (KNN), into an enhanced model, with mean performance equal to 85% by terms of the Pearson correlation coefficient (PCC) between observed and predicted values.

Keywords: aflatoxins, Aspergillus spp., dried figs, k-nearest neighbors, machine learning, prediction

Procedia PDF Downloads 172
2246 Machine learning Assisted Selective Emitter design for Solar Thermophotovoltaic System

Authors: Ambali Alade Odebowale, Andargachew Mekonnen Berhe, Haroldo T. Hattori, Andrey E. Miroshnichenko

Abstract:

Solar thermophotovoltaic systems (STPV) have emerged as a promising solution to overcome the Shockley-Queisser limit, a significant impediment in the direct conversion of solar radiation into electricity using conventional solar cells. The STPV system comprises essential components such as an optical concentrator, selective emitter, and a thermophotovoltaic (TPV) cell. The pivotal element in achieving high efficiency in an STPV system lies in the design of a spectrally selective emitter or absorber. Traditional methods for designing and optimizing selective emitters are often time-consuming and may not yield highly selective emitters, posing a challenge to the overall system performance. In recent years, the application of machine learning techniques in various scientific disciplines has demonstrated significant advantages. This paper proposes a novel nanostructure composed of four-layered materials (SiC/W/SiO2/W) to function as a selective emitter in the energy conversion process of an STPV system. Unlike conventional approaches widely adopted by researchers, this study employs a machine learning-based approach for the design and optimization of the selective emitter. Specifically, a random forest algorithm (RFA) is employed for the design of the selective emitter, while the optimization process is executed using genetic algorithms. This innovative methodology holds promise in addressing the challenges posed by traditional methods, offering a more efficient and streamlined approach to selective emitter design. The utilization of a machine learning approach brings several advantages to the design and optimization of a selective emitter within the STPV system. Machine learning algorithms, such as the random forest algorithm, have the capability to analyze complex datasets and identify intricate patterns that may not be apparent through traditional methods. This allows for a more comprehensive exploration of the design space, potentially leading to highly efficient emitter configurations. Moreover, the application of genetic algorithms in the optimization process enhances the adaptability and efficiency of the overall system. Genetic algorithms mimic the principles of natural selection, enabling the exploration of a diverse range of emitter configurations and facilitating the identification of optimal solutions. This not only accelerates the design and optimization process but also increases the likelihood of discovering configurations that exhibit superior performance compared to traditional methods. In conclusion, the integration of machine learning techniques in the design and optimization of a selective emitter for solar thermophotovoltaic systems represents a groundbreaking approach. This innovative methodology not only addresses the limitations of traditional methods but also holds the potential to significantly improve the overall performance of STPV systems, paving the way for enhanced solar energy conversion efficiency.

Keywords: emitter, genetic algorithm, radiation, random forest, thermophotovoltaic

Procedia PDF Downloads 51
2245 Efficient Manageability and Intelligent Classification of Web Browsing History Using Machine Learning

Authors: Suraj Gururaj, Sumantha Udupa U.

Abstract:

Browsing the Web has emerged as the de facto activity performed on the Internet. Although browsing gets tracked, the manageability aspect of Web browsing history is very poor. In this paper, we have a workable solution implemented by using machine learning and natural language processing techniques for efficient manageability of user’s browsing history. The significance of adding such a capability to a Web browser is that it ensures efficient and quick information retrieval from browsing history, which currently is very challenging. Our solution guarantees that any important websites visited in the past can be easily accessible because of the intelligent and automatic classification. In a nutshell, our solution-based paper provides an implementation as a browser extension by intelligently classifying the browsing history into most relevant category automatically without any user’s intervention. This guarantees no information is lost and increases productivity by saving time spent revisiting websites that were of much importance.

Keywords: adhoc retrieval, Chrome extension, supervised learning, tile, Web personalization

Procedia PDF Downloads 360
2244 Machine Learning Facing Behavioral Noise Problem in an Imbalanced Data Using One Side Behavioral Noise Reduction: Application to a Fraud Detection

Authors: Salma El Hajjami, Jamal Malki, Alain Bouju, Mohammed Berrada

Abstract:

With the expansion of machine learning and data mining in the context of Big Data analytics, the common problem that affects data is class imbalance. It refers to an imbalanced distribution of instances belonging to each class. This problem is present in many real world applications such as fraud detection, network intrusion detection, medical diagnostics, etc. In these cases, data instances labeled negatively are significantly more numerous than the instances labeled positively. When this difference is too large, the learning system may face difficulty when tackling this problem, since it is initially designed to work in relatively balanced class distribution scenarios. Another important problem, which usually accompanies these imbalanced data, is the overlapping instances between the two classes. It is commonly referred to as noise or overlapping data. In this article, we propose an approach called: One Side Behavioral Noise Reduction (OSBNR). This approach presents a way to deal with the problem of class imbalance in the presence of a high noise level. OSBNR is based on two steps. Firstly, a cluster analysis is applied to groups similar instances from the minority class into several behavior clusters. Secondly, we select and eliminate the instances of the majority class, considered as behavioral noise, which overlap with behavior clusters of the minority class. The results of experiments carried out on a representative public dataset confirm that the proposed approach is efficient for the treatment of class imbalances in the presence of noise.

Keywords: machine learning, imbalanced data, data mining, big data

Procedia PDF Downloads 116
2243 Study of the Use of Artificial Neural Networks in Islamic Finance

Authors: Kaoutar Abbahaddou, Mohammed Salah Chiadmi

Abstract:

The need to find a relevant way to predict the next-day price of a stock index is a real concern for many financial stakeholders and researchers. We have known across years the proliferation of several methods. Nevertheless, among all these methods, the most controversial one is a machine learning algorithm that claims to be reliable, namely neural networks. Thus, the purpose of this article is to study the prediction power of neural networks in the particular case of Islamic finance as it is an under-looked area. In this article, we will first briefly present a review of the literature regarding neural networks and Islamic finance. Next, we present the architecture and principles of artificial neural networks most commonly used in finance. Then, we will show its empirical application on two Islamic stock indexes. The accuracy rate would be used to measure the performance of the algorithm in predicting the right price the next day. As a result, we can conclude that artificial neural networks are a reliable method to predict the next-day price for Islamic indices as it is claimed for conventional ones.

Keywords: Islamic finance, stock price prediction, artificial neural networks, machine learning

Procedia PDF Downloads 214
2242 Sentiment Analysis on the East Timor Accession Process to the ASEAN

Authors: Marcelino Caetano Noronha, Vosco Pereira, Jose Soares Pinto, Ferdinando Da C. Saores

Abstract:

One particularly popular social media platform is Youtube. It’s a video-sharing platform where users can submit videos, and other users can like, dislike or comment on the videos. In this study, we conduct a binary classification task on YouTube’s video comments and review from the users regarding the accession process of Timor Leste to become the eleventh member of the Association of South East Asian Nations (ASEAN). We scrape the data directly from the public YouTube video and apply several pre-processing and weighting techniques. Before conducting the classification, we categorized the data into two classes, namely positive and negative. In the classification part, we apply Support Vector Machine (SVM) algorithm. By comparing with Naïve Bayes Algorithm, the experiment showed SVM achieved 84.1% of Accuracy, 94.5% of Precision, and Recall 73.8% simultaneously.

Keywords: classification, YouTube, sentiment analysis, support sector machine

Procedia PDF Downloads 90
2241 A Generalized Framework for Adaptive Machine Learning Deployments in Algorithmic Trading

Authors: Robert Caulk

Abstract:

A generalized framework for adaptive machine learning deployments in algorithmic trading is introduced, tested, and released as open-source code. The presented software aims to test the hypothesis that recent data contains enough information to form a probabilistically favorable short-term price prediction. Further, the framework contains various adaptive machine learning techniques that are geared toward generating profit during strong trends and minimizing losses during trend changes. Results demonstrate that this adaptive machine learning approach is capable of capturing trends and generating profit. The presentation also discusses the importance of defining the parameter space associated with the dynamic training data-set and using the parameter space to identify and remove outliers from prediction data points. Meanwhile, the generalized architecture enables common users to exploit the powerful machinery while focusing on high-level feature engineering and model testing. The presentation also highlights common strengths and weaknesses associated with the presented technique and presents a broad range of well-tested starting points for feature set construction, target setting, and statistical methods for enforcing risk management and maintaining probabilistically favorable entry and exit points. The presentation also describes the end-to-end data processing tools associated with FreqAI, including automatic data fetching, data aggregation, feature engineering, safe and robust data pre-processing, outlier detection, custom machine learning and statistical tools, data post-processing, and adaptive training backtest emulation, and deployment of adaptive training in live environments. Finally, the generalized user interface is also discussed in the presentation. Feature engineering is simplified so that users can seed their feature sets with common indicator libraries (e.g. TA-lib, pandas-ta). The user also feeds data expansion parameters to fill out a large feature set for the model, which can contain as many as 10,000+ features. The presentation describes the various object-oriented programming techniques employed to make FreqAI agnostic to third-party libraries and external data sources. In other words, the back-end is constructed in such a way that users can leverage a broad range of common regression libraries (Catboost, LightGBM, Sklearn, etc) as well as common Neural Network libraries (TensorFlow, PyTorch) without worrying about the logistical complexities associated with data handling and API interactions. The presentation finishes by drawing conclusions about the most important parameters associated with a live deployment of the adaptive learning framework and provides the road map for future development in FreqAI.

Keywords: machine learning, market trend detection, open-source, adaptive learning, parameter space exploration

Procedia PDF Downloads 77
2240 Combined Machine That Fertilizes Evenly under Plowing on Slopes and Planning an Experiment

Authors: Qurbanov Huseyn Nuraddin

Abstract:

The results of scientific research on a machine that pours an equal amount of mineral fertilizer under the soil to increase the productivity of grain in mountain farming and obtain quality grain are substantiated. The average yield of the crop depends on the nature of the distribution of fertilizers in the soil. Therefore, the study of effective energy-saving methods for the application of mineral fertilizers is the actual task of modern agriculture. Depending on the type and variety of plants in mountain farming, there is an optimal norm of mineral fertilizers. Applying an equal amount of fertilizer to the soil is one of the conditions that increase the efficiency of the field. One of the main agro-technical indicators of the work of mineral fertilizing machines is to ensure equal distribution of mineral fertilizers in the field. Taking into account the above-mentioned issues, a combined plough has been improved in our laboratory.

Keywords: combined plough, mineral fertilizers, sprinkle fluently, fertilizer rate, cereals

Procedia PDF Downloads 63
2239 A Review of Intelligent Fire Management Systems to Reduce Wildfires

Authors: Nomfundo Ngombane, Topside E. Mathonsi

Abstract:

Remote sensing and satellite imaging have been widely used to detect wildfires; nevertheless, the technologies present some limitations in terms of early wildfire detection as the technologies are greatly influenced by weather conditions and can miss small fires. The fires need to have spread a few kilometers for the technologies to provide accurate detection. The South African Advanced Fire Information System uses MODIS (Moderate Resolution Imaging Spectroradiometer) as satellite imaging. MODIS has limitations as it can exclude small fires and can fall short in validating fire vulnerability. Thus in the future, a Machine Learning algorithm will be designed and implemented for the early detection of wildfires. A simulator will be used to evaluate the effectiveness of the proposed solution, and the results of the simulation will be presented.

Keywords: moderate resolution imaging spectroradiometer, advanced fire information system, machine learning algorithm, detection of wildfires

Procedia PDF Downloads 70
2238 Integrating Natural Language Processing (NLP) and Machine Learning in Lung Cancer Diagnosis

Authors: Mehrnaz Mostafavi

Abstract:

The assessment and categorization of incidental lung nodules present a considerable challenge in healthcare, often necessitating resource-intensive multiple computed tomography (CT) scans for growth confirmation. This research addresses this issue by introducing a distinct computational approach leveraging radiomics and deep-learning methods. However, understanding local services is essential before implementing these advancements. With diverse tracking methods in place, there is a need for efficient and accurate identification approaches, especially in the context of managing lung nodules alongside pre-existing cancer scenarios. This study explores the integration of text-based algorithms in medical data curation, indicating their efficacy in conjunction with machine learning and deep-learning models for identifying lung nodules. Combining medical images with text data has demonstrated superior data retrieval compared to using each modality independently. While deep learning and text analysis show potential in detecting previously missed nodules, challenges persist, such as increased false positives. The presented research introduces a Structured-Query-Language (SQL) algorithm designed for identifying pulmonary nodules in a tertiary cancer center, externally validated at another hospital. Leveraging natural language processing (NLP) and machine learning, the algorithm categorizes lung nodule reports based on sentence features, aiming to facilitate research and assess clinical pathways. The hypothesis posits that the algorithm can accurately identify lung nodule CT scans and predict concerning nodule features using machine-learning classifiers. Through a retrospective observational study spanning a decade, CT scan reports were collected, and an algorithm was developed to extract and classify data. Results underscore the complexity of lung nodule cohorts in cancer centers, emphasizing the importance of careful evaluation before assuming a metastatic origin. The SQL and NLP algorithms demonstrated high accuracy in identifying lung nodule sentences, indicating potential for local service evaluation and research dataset creation. Machine-learning models exhibited strong accuracy in predicting concerning changes in lung nodule scan reports. While limitations include variability in disease group attribution, the potential for correlation rather than causality in clinical findings, and the need for further external validation, the algorithm's accuracy and potential to support clinical decision-making and healthcare automation represent a significant stride in lung nodule management and research.

Keywords: lung cancer diagnosis, structured-query-language (SQL), natural language processing (NLP), machine learning, CT scans

Procedia PDF Downloads 70
2237 Design and Modeling of Light Duty Trencher

Authors: Yegetaneh T. Dejenu, Delesa Kejela, Abdulak Alemu

Abstract:

From the earliest time of humankind, the trenches were used for water to flow along and for soldiers to hide in during enemy attacks. Now a day due to civilization, the needs of the human being become endless, and the living condition becomes sophisticated. The unbalance between the needs and resource obligates them to find the way to manage this condition. The attempt to use the scares resource in very efficient and effective way makes the trench an endeavor practice in the world in all countries. A trencher is a construction equipment used to dig trenches, especially for laying pipes or cables, installing drainage, irrigation, installing fencing, and in preparation for trench warfare. It is a machine used to make a ditch by cutting the soil ground and effectively used in agricultural irrigation. The most common types of trencher are wheel trencher, chain trencher, micro trencher, portable trencher. In Ethiopia people have been trenching the ditch for many purposes and the tools they are using are Pickaxe, Shovel and some are using Micro Excavators. The adverse effect of using traditional equipment is, time and energy consuming, less productive, difficult and more man power is required. Hence it is necessary to design and produce low price, and simple machine to narrow this gap. Our objective is to design and model a light duty trencher that is used for trenching the ground or soil for making ditch and used for agricultural, ground cabling, ground piping, and drainage system. The designed machine trenches, maximum of 1-meter depth, 30 cm width, and the required length. The working mechanism is fully hydraulic, and the engine with 12.7 hp will provide suitable power for the pump that delivers 23 l/min at 1500 rpm to drive hydraulic motors and actuators.

Keywords: hydraulics, modelling, trenching, ditch

Procedia PDF Downloads 203
2236 Syndromic Surveillance Framework Using Tweets Data Analytics

Authors: David Ming Liu, Benjamin Hirsch, Bashir Aden

Abstract:

Syndromic surveillance is to detect or predict disease outbreaks through the analysis of medical sources of data. Using social media data like tweets to do syndromic surveillance becomes more and more popular with the aid of open platform to collect data and the advantage of microblogging text and mobile geographic location features. In this paper, a Syndromic Surveillance Framework is presented with machine learning kernel using tweets data analytics. Influenza and the three cities Abu Dhabi, Al Ain and Dubai of United Arabic Emirates are used as the test disease and trial areas. Hospital cases data provided by the Health Authority of Abu Dhabi (HAAD) are used for the correlation purpose. In our model, Latent Dirichlet allocation (LDA) engine is adapted to do supervised learning classification and N-Fold cross validation confusion matrix are given as the simulation results with overall system recall 85.595% performance achieved.

Keywords: Syndromic surveillance, Tweets, Machine Learning, data mining, Latent Dirichlet allocation (LDA), Influenza

Procedia PDF Downloads 102
2235 A Data-Driven Platform for Studying the Liquid Plug Splitting Ratio

Authors: Ehsan Atefi, Michael Grigware

Abstract:

Respiratory failure secondary to surfactant deficiency resulting from respiratory distress syndrome is considered one major cause of morbidity in preterm infants. Surfactant replacement treatment (SRT) is considered an effective treatment for this disease. Here, we introduce an AI-mediated approach for estimating the distribution of surfactant in the lung airway of a newborn infant during SRT. Our approach implements machine learning to precisely estimate the splitting ratio of a liquid drop during bifurcation at different injection velocities and patient orientations. This technique can be used to calculate the surfactant residue remaining on the airway wall during the surfactant injection process. Our model works by minimizing the pressure drop difference between the two airway branches at each generation, subject to mass and momentum conservation. Our platform can be used to generate feedback for immediately adjusting the velocity of injection and patient orientation during SRT.

Keywords: respiratory failure, surfactant deficiency, surfactant replacement, machine learning

Procedia PDF Downloads 115
2234 Exploring Artificial Intelligence as a Transformative Tool for Urban Management

Authors: R. R. Govind

Abstract:

In the digital age, artificial intelligence (AI) is having a significant impact on the rapid changes that cities are experiencing. This study explores the profound impact of AI on urban morphology, especially with regard to promoting friendly design choices. It addresses a significant research gap by examining the real-world effects of integrating AI into urban design and management. The main objective is to outline a framework for integrating AI to transform urban settings. The study employs an urban design framework to effectively navigate complicated urban environments, emphasize the need for urban management, and provide efficient planning and design strategies. Taking Gangtok's informal settlements as a focal point, the study employs AI methodologies such as machine learning, predictive analytics, and generative AI to tackle issues of 'urban informality'. The insights garnered not only offer valuable perspectives but also unveil AI's transformative potential in addressing contemporary urban challenges.

Keywords: urban design, artificial intelligence, urban challenges, machine learning, urban informality

Procedia PDF Downloads 46