Search results for: cable splicing machine
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3017

Search results for: cable splicing machine

1817 A.T.O.M.- Artificial Intelligent Omnipresent Machine

Authors: R. Kanthavel, R. Yogesh Kumar, T. Narendrakumar, B. Santhosh, S. Surya Prakash

Abstract:

This paper primarily focuses on developing an affordable personal assistant and the implementation of it in the field of Artificial Intelligence (AI) to create a virtual assistant/friend. The problem in existing home automation techniques is that it requires the usage of exact command words present in the database to execute the corresponding task. Our proposed work is ATOM a.k.a ‘Artificial intelligence Talking Omnipresent Machine’. Our inspiration came from an unlikely source- the movie ‘Iron Man’ in which a character called J.A.R.V.I.S has omnipresence, and device controlling capability. This device can control household devices in real time and send the live information to the user. This device does not require the user to utter the exact commands specified in the database as it can capture the keywords from the uttered commands, correlates the obtained keywords and perform the specified task. This ability to compare and correlate the keywords gives the user the liberty to give commands which are not necessarily the exact words provided in the database. The proposed work has a higher flexibility (due to its keyword extracting ability from the user input) comparing to the existing work Intelligent Home automation System (IHAS), is more accurate, and is much more affordable as it makes use of WI-FI module and raspberry pi 2 instead of ZigBee and a computer respectively.

Keywords: home automation, speech recognition, voice control, personal assistant, artificial intelligence

Procedia PDF Downloads 336
1816 Synergizing Additive Manufacturing and Artificial Intelligence: Analyzing and Predicting the Mechanical Behavior of 3D-Printed CF-PETG Composites

Authors: Sirine Sayed, Mostapha Tarfaoui, Abdelmalek Toumi, Youssef Qarssis, Mohamed Daly, Chokri Bouraoui

Abstract:

This paper delves into the combination of additive manufacturing (AM) and artificial intelligence (AI) to solve challenges related to the mechanical behavior of AM-produced parts. The article highlights the fundamentals and benefits of additive manufacturing, including creating complex geometries, optimizing material use, and streamlining manufacturing processes. The paper also addresses the challenges associated with additive manufacturing, such as ensuring stable mechanical performance and material properties. The role of AI in improving the static behavior of AM-produced parts, including machine learning, especially the neural network, is to make regression models to analyze the large amounts of data generated during experimental tests. It investigates the potential synergies between AM and AI to achieve enhanced functions and personalized mechanical properties. The mechanical behavior of parts produced using additive manufacturing methods can be further improved using design optimization, structural analysis, and AI-based adaptive manufacturing. The article concludes by emphasizing the importance of integrating AM and AI to enhance mechanical operations, increase reliability, and perform advanced functions, paving the way for innovative applications in different fields.

Keywords: additive manufacturing, mechanical behavior, artificial intelligence, machine learning, neural networks, reliability, advanced functionalities

Procedia PDF Downloads 10
1815 Experimental Investigation of Seawater Thermophysical Properties: Understanding Climate Change Impacts on Marine Ecosystems Through Internal Pressure and Cohesion Energy Analysis

Authors: Nishaben Dholakiya, Anirban Roy, Ranjan Dey

Abstract:

The unprecedented rise in global temperatures has triggered complex changes in marine ecosystems, necessitating a deeper understanding of seawater's thermophysical properties by experimentally measuring ultrasonic velocity and density at varying temperatures and salinity. This study investigates the critical relationship between temperature variations and molecular-level interactions in Arabian Sea surface waters, specifically focusing on internal pressure (π) and cohesion energy density (CED) as key indicators of ecosystem disruption. Our experimental findings reveal that elevated temperatures significantly reduce internal pressure, weakening the intermolecular forces that maintain seawater's structural integrity. This reduction in π correlates directly with decreased habitat stability for marine organisms, particularly affecting pressure-sensitive species and their physiological processes. Similarly, the observed decline in cohesion energy density at higher temperatures indicates a fundamental shift in water molecule organization, impacting the dissolution and distribution of vital nutrients and gases. These molecular-level changes cascade through the ecosystem, affecting everything from planktonic organisms to complex food webs. By employing advanced machine learning techniques, including Stacked Ensemble Machine Learning (SEML) and AdaBoost (AB), we developed highly accurate predictive models (>99% accuracy) for these thermophysical parameters. The results provide crucial insights into the mechanistic relationship between climate warming and marine ecosystem degradation, offering valuable data for environmental policymaking and conservation strategies. The novelty of this research serves as no such thermodynamic investigation has been conducted before in literature, whereas this research establishes a quantitative framework for understanding how molecular-level changes in seawater properties directly influence marine ecosystem stability, emphasizing the urgent need for climate change mitigation efforts.

Keywords: thermophysical properties, Arabian Sea, internal pressure, cohesion energy density, machine learning

Procedia PDF Downloads 5
1814 Semantic Differences between Bug Labeling of Different Repositories via Machine Learning

Authors: Pooja Khanal, Huaming Zhang

Abstract:

Labeling of issues/bugs, also known as bug classification, plays a vital role in software engineering. Some known labels/classes of bugs are 'User Interface', 'Security', and 'API'. Most of the time, when a reporter reports a bug, they try to assign some predefined label to it. Those issues are reported for a project, and each project is a repository in GitHub/GitLab, which contains multiple issues. There are many software project repositories -ranging from individual projects to commercial projects. The labels assigned for different repositories may be dependent on various factors like human instinct, generalization of labels, label assignment policy followed by the reporter, etc. While the reporter of the issue may instinctively give that issue a label, another person reporting the same issue may label it differently. This way, it is not known mathematically if a label in one repository is similar or different to the label in another repository. Hence, the primary goal of this research is to find the semantic differences between bug labeling of different repositories via machine learning. Independent optimal classifiers for individual repositories are built first using the text features from the reported issues. The optimal classifiers may include a combination of multiple classifiers stacked together. Then, those classifiers are used to cross-test other repositories which leads the result to be deduced mathematically. The produce of this ongoing research includes a formalized open-source GitHub issues database that is used to deduce the similarity of the labels pertaining to the different repositories.

Keywords: bug classification, bug labels, GitHub issues, semantic differences

Procedia PDF Downloads 200
1813 Machine Learning in Gravity Models: An Application to International Recycling Trade Flow

Authors: Shan Zhang, Peter Suechting

Abstract:

Predicting trade patterns is critical to decision-making in public and private domains, especially in the current context of trade disputes among major economies. In the past, U.S. recycling has relied heavily on strong demand for recyclable materials overseas. However, starting in 2017, a series of new recycling policies (bans and higher inspection standards) was enacted by multiple countries that were the primary importers of recyclables from the U.S. prior to that point. As the global trade flow of recycling shifts, some new importers, mostly developing countries in South and Southeast Asia, have been overwhelmed by the sheer quantities of scrap materials they have received. As the leading exporter of recyclable materials, the U.S. now has a pressing need to build its recycling industry domestically. With respect to the global trade in scrap materials used for recycling, the interest in this paper is (1) predicting how the export of recyclable materials from the U.S. might vary over time, and (2) predicting how international trade flows for recyclables might change in the future. Focusing on three major recyclable materials with a history of trade, this study uses data-driven and machine learning (ML) algorithms---supervised (shrinkage and tree methods) and unsupervised (neural network method)---to decipher the international trade pattern of recycling. Forecasting the potential trade values of recyclables in the future could help importing countries, to which those materials will shift next, to prepare related trade policies. Such policies can assist policymakers in minimizing negative environmental externalities and in finding the optimal amount of recyclables needed by each country. Such forecasts can also help exporting countries, like the U.S understand the importance of healthy domestic recycling industry. The preliminary result suggests that gravity models---in addition to particular selection macroeconomic predictor variables--are appropriate predictors of the total export value of recyclables. With the inclusion of variables measuring aspects of the political conditions (trade tariffs and bans), predictions show that recyclable materials are shifting from more policy-restricted countries to less policy-restricted countries in international recycling trade. Those countries also tend to have high manufacturing activities as a percentage of their GDP.

Keywords: environmental economics, machine learning, recycling, international trade

Procedia PDF Downloads 168
1812 Advancing the Analysis of Physical Activity Behaviour in Diverse, Rapidly Evolving Populations: Using Unsupervised Machine Learning to Segment and Cluster Accelerometer Data

Authors: Christopher Thornton, Niina Kolehmainen, Kianoush Nazarpour

Abstract:

Background: Accelerometers are widely used to measure physical activity behavior, including in children. The traditional method for processing acceleration data uses cut points, relying on calibration studies that relate the quantity of acceleration to energy expenditure. As these relationships do not generalise across diverse populations, they must be parametrised for each subpopulation, including different age groups, which is costly and makes studies across diverse populations difficult. A data-driven approach that allows physical activity intensity states to emerge from the data under study without relying on parameters derived from external populations offers a new perspective on this problem and potentially improved results. We evaluated the data-driven approach in a diverse population with a range of rapidly evolving physical and mental capabilities, namely very young children (9-38 months old), where this new approach may be particularly appropriate. Methods: We applied an unsupervised machine learning approach (a hidden semi-Markov model - HSMM) to segment and cluster the accelerometer data recorded from 275 children with a diverse range of physical and cognitive abilities. The HSMM was configured to identify a maximum of six physical activity intensity states and the output of the model was the time spent by each child in each of the states. For comparison, we also processed the accelerometer data using published cut points with available thresholds for the population. This provided us with time estimates for each child’s sedentary (SED), light physical activity (LPA), and moderate-to-vigorous physical activity (MVPA). Data on the children’s physical and cognitive abilities were collected using the Paediatric Evaluation of Disability Inventory (PEDI-CAT). Results: The HSMM identified two inactive states (INS, comparable to SED), two lightly active long duration states (LAS, comparable to LPA), and two short-duration high-intensity states (HIS, comparable to MVPA). Overall, the children spent on average 237/392 minutes per day in INS/SED, 211/129 minutes per day in LAS/LPA, and 178/168 minutes in HIS/MVPA. We found that INS overlapped with 53% of SED, LAS overlapped with 37% of LPA and HIS overlapped with 60% of MVPA. We also looked at the correlation between the time spent by a child in either HIS or MVPA and their physical and cognitive abilities. We found that HIS was more strongly correlated with physical mobility (R²HIS =0.5, R²MVPA= 0.28), cognitive ability (R²HIS =0.31, R²MVPA= 0.15), and age (R²HIS =0.15, R²MVPA= 0.09), indicating increased sensitivity to key attributes associated with a child’s mobility. Conclusion: An unsupervised machine learning technique can segment and cluster accelerometer data according to the intensity of movement at a given time. It provides a potentially more sensitive, appropriate, and cost-effective approach to analysing physical activity behavior in diverse populations, compared to the current cut points approach. This, in turn, supports research that is more inclusive across diverse populations.

Keywords: physical activity, machine learning, under 5s, disability, accelerometer

Procedia PDF Downloads 210
1811 Study of the Physical Aging of Polyvinyl Chloride (PVC)

Authors: Mohamed Ouazene

Abstract:

The insulating properties of the polymers are widely used in electrical engineering for the production of insulators and various supports, as well as for the insulation of electric cables for medium and high voltage, etc. These polymeric materials have significant advantages both technically and economically. However, although the insulation with polymeric materials has advantages, there are also certain disadvantages such as the influence of the heat which can have a detrimental effect on these materials. Polyvinyl chloride (PVC) is one of the polymers used in a plasticized state in the cable insulation to medium and high voltage. The studied material is polyvinyl chloride (PVC 4000 M) from the Algerian national oil company whose formula is: Industrial PVC 4000 M is in the form of white powder. The test sample is a pastille of 1 mm thick and 1 cm in diameter. The consequences of increasing the temperature of a polymer are modifications; some of them are reversible and others irreversible [1]. The reversible changes do not affect the chemical composition of the polymer, or its structure. They are characterized by transitions and relaxations. The glass transition temperature is an important feature of a polymer. Physical aging of PVC is to maintain the material for a longer or shorter time to its glass transition temperature. The aim of this paper is to study this phenomenon by the method of thermally stimulated depolarization currents. Relaxations within the polymer have been recorded in the form of current peaks. We have found that the intensity decreases for more residence time in the polymer along its glass transition temperature. Furthermore, it is inferred from this work that the phenomenon of physical aging can have important consequences on the properties of the polymer. It leads to a more compact rearrangement of the material and a reconstruction or reinforcement of structural connections.

Keywords: depolarization currents, glass transition temperature, physical aging, polyvinyl chloride (PVC)

Procedia PDF Downloads 388
1810 Application of Machine Learning on Google Earth Engine for Forest Fire Severity, Burned Area Mapping and Land Surface Temperature Analysis: Rajasthan, India

Authors: Alisha Sinha, Laxmi Kant Sharma

Abstract:

Forest fires are a recurring issue in many parts of the world, including India. These fires can have various causes, including human activities (such as agricultural burning, campfires, or discarded cigarettes) and natural factors (such as lightning). This study presents a comprehensive and advanced methodology for assessing wildfire susceptibility by integrating diverse environmental variables and leveraging cutting-edge machine learning techniques across Rajasthan, India. The primary goal of the study is to utilize Google Earth Engine to compare locations in Sariska National Park, Rajasthan (India), before and after forest fires. High-resolution satellite data were used to assess the amount and types of changes caused by forest fires. The present study meticulously analyzes various environmental variables, i.e., slope orientation, elevation, normalized difference vegetation index (NDVI), drainage density, precipitation, and temperature, to understand landscape characteristics and assess wildfire susceptibility. In addition, a sophisticated random forest regression model is used to predict land surface temperature based on a set of environmental parameters.

Keywords: wildfire susceptibility mapping, LST, random forest, GEE, MODIS, climatic parameters

Procedia PDF Downloads 22
1809 Development of a Turbulent Boundary Layer Wall-pressure Fluctuations Power Spectrum Model Using a Stepwise Regression Algorithm

Authors: Zachary Huffman, Joana Rocha

Abstract:

Wall-pressure fluctuations induced by the turbulent boundary layer (TBL) developed over aircraft are a significant source of aircraft cabin noise. Since the power spectral density (PSD) of these pressure fluctuations is directly correlated with the amount of sound radiated into the cabin, the development of accurate empirical models that predict the PSD has been an important ongoing research topic. The sound emitted can be represented from the pressure fluctuations term in the Reynoldsaveraged Navier-Stokes equations (RANS). Therefore, early TBL empirical models (including those from Lowson, Robertson, Chase, and Howe) were primarily derived by simplifying and solving the RANS for pressure fluctuation and adding appropriate scales. Most subsequent models (including Goody, Efimtsov, Laganelli, Smol’yakov, and Rackl and Weston models) were derived by making modifications to these early models or by physical principles. Overall, these models have had varying levels of accuracy, but, in general, they are most accurate under the specific Reynolds and Mach numbers they were developed for, while being less accurate under other flow conditions. Despite this, recent research into the possibility of using alternative methods for deriving the models has been rather limited. More recent studies have demonstrated that an artificial neural network model was more accurate than traditional models and could be applied more generally, but the accuracy of other machine learning techniques has not been explored. In the current study, an original model is derived using a stepwise regression algorithm in the statistical programming language R, and TBL wall-pressure fluctuations PSD data gathered at the Carleton University wind tunnel. The theoretical advantage of a stepwise regression approach is that it will automatically filter out redundant or uncorrelated input variables (through the process of feature selection), and it is computationally faster than machine learning. The main disadvantage is the potential risk of overfitting. The accuracy of the developed model is assessed by comparing it to independently sourced datasets.

Keywords: aircraft noise, machine learning, power spectral density models, regression models, turbulent boundary layer wall-pressure fluctuations

Procedia PDF Downloads 135
1808 AI-Driven Forecasting Models for Anticipating Oil Market Trends and Demand

Authors: Gaurav Kumar Sinha

Abstract:

The volatility of the oil market, influenced by geopolitical, economic, and environmental factors, presents significant challenges for stakeholders in predicting trends and demand. This article explores the application of artificial intelligence (AI) in developing robust forecasting models to anticipate changes in the oil market more accurately. We delve into various AI techniques, including machine learning, deep learning, and time series analysis, that have been adapted to analyze historical data and current market conditions to forecast future trends. The study evaluates the effectiveness of these models in capturing complex patterns and dependencies in market data, which traditional forecasting methods often miss. Additionally, the paper discusses the integration of external variables such as political events, economic policies, and technological advancements that influence oil prices and demand. By leveraging AI, stakeholders can achieve a more nuanced understanding of market dynamics, enabling better strategic planning and risk management. The article concludes with a discussion on the potential of AI-driven models in enhancing the predictive accuracy of oil market forecasts and their implications for global economic planning and strategic resource allocation.

Keywords: AI forecasting, oil market trends, machine learning, deep learning, time series analysis, predictive analytics, economic factors, geopolitical influence, technological advancements, strategic planning

Procedia PDF Downloads 35
1807 Development of an Advanced Power Ultrasonic-Assisted Drilling System

Authors: M. A. Moghaddas, M. Short, N. Wiley, A. Y. Yi, K. F. Graff

Abstract:

The application of ultrasonic vibrations to machining processes has a long history, ranging from slurry-based systems able to drill brittle materials, to more recent developments involving low power ultrasonics for high precision machining, with many of these at the research and laboratory stages. The focus of this development is the application of high levels of ultrasonic power (1,000’s of watts) to standard, heavy duty machine tools – drilling being the immediate focus, with developments in milling in progress – with the objective of dramatically increasing system productivity through faster feed rates, this benefit arising from the thrust force reductions obtained by power ultrasonic vibrations. The presentation will describe development of an advanced drilling system based on a special, acoustically designed, rugged drill module capable of functioning under heavy duty production conditions, and making use of standard tool holder means, and able to obtain thrust force reductions while maintaining or improving surface finish and drilling accuracy. The characterization of the system performance will be described, and results obtained in drilling several materials (Aluminum, Stainless steel, Titanium) presented.

Keywords: dimensional accuracy, machine tool, productivity, surface roughness, thrust force, ultrasonic vibrations, ultrasonic-assisted drilling

Procedia PDF Downloads 277
1806 Medium-Scale Multi-Juice Extractor for Food Processing

Authors: Flordeliza L. Mercado, Teresito G. Aguinaldo, Helen F. Gavino, Victorino T. Taylan

Abstract:

Most fruits and vegetables are available in large quantities during peak season which are oftentimes marketed at low price and left to rot or fed to farm animals. The lack of efficient storage facilities, and the additional cost and unavailability of small machinery for food processing, results to low price and wastage. Incidentally, processed fresh fruits and vegetables are gaining importance nowadays and health conscious people are also into ‘juicing’. One way to reduce wastage and ensure an all-season availability of crop juices at reasonable costs is to develop equipment for effective extraction of juice. The study was conducted to design, fabricate and evaluate a multi-juice extractor using locally available materials, making it relatively cheaper and affordable for medium-scale enterprises. The study was also conducted to formulate juice blends using extracted juices and calamansi juice at different blending percentage, and evaluate its chemical properties and sensory attributes. Furthermore, the chemical properties of extracted meals were evaluated for future applications. The multi-juice extractor has an overall dimension of 963mm x 300mm x 995mm, a gross weight of 82kg and 5 major components namely; feeding hopper, extracting chamber, juice and meal outlet, transmission assembly, and frame. The machine performance was evaluated based on juice recovery, extraction efficiency, extraction rate, extraction recovery, and extraction loss considering type of crop as apple and carrot with three replications each and was analyzed using T-test. The formulated juice blends were subjected to sensory evaluation and data gathered were analyzed using Analysis of Variance appropriate for Complete Randomized Design. Results showed that the machine’s juice recovery (73.39%), extraction rate (16.40li/hr), and extraction efficiency (88.11%) for apple were significantly higher than for carrot while extraction recovery (99.88%) was higher for apple than for carrot. Extraction loss (0.12%) was lower for apple than for carrot, but was not significantly affected by crop. Based on adding percentage mark-up on extraction cost (Php 2.75/kg), the breakeven weight and payback period for a 35% mark-up is 4,710.69kg and 1.22 years, respectively and for a 50% mark-up, the breakeven weight is 3,492.41kg and the payback period is 0.86 year (10.32 months). Results on the sensory evaluation of juice blends showed that the type of juice significantly influenced all the sensory parameters while the blending percentage including their respective interaction, had no significant effect on all sensory parameters, making the apple-calamansi juice blend more preferred than the carrot-calamansi juice blend in terms of all the sensory parameter. The machine’s performance is higher for apple than for carrot and the cost analysis on the use of the machine revealed that it is financially viable with a payback period of 1.22 years (35% mark-up) and 0.86 year (50% mark-up) for machine cost, generating an income of Php 23,961.60 and Php 34,444.80 per year using 35% and 50% mark-up, respectively. The juice blends were of good qualities based on the values obtained in the chemical analysis and the extracted meal could also be used to produce another product based on the values obtained from proximate analysis.

Keywords: food processing, fruits and vegetables, juice extraction, multi-juice extractor

Procedia PDF Downloads 324
1805 Integrated Machine Learning Framework for At-Home Patients Personalized Risk Prediction Using Activities, Biometric, and Demographic Features

Authors: Claire Xu, Welton Wang, Manasvi Pinnaka, Anqi Pan, Michael Han

Abstract:

Hospitalizations account for one-third of the total health care spending in the US. Early risk detection and intervention can reduce this high cost and increase the satisfaction of both patients and physicians. Due to the lack of awareness of the potential arising risks in home environment, the opportunities for patients to seek early actions of clinical visits are dramatically reduced. This research aims to offer a highly personalized remote patients monitoring and risk assessment AI framework to identify the potentially preventable hospitalization for both acute as well as chronic diseases. A hybrid-AI framework is trained with data from clinical setting, patients surveys, as well as online databases. 20+ risk factors are analyzed ranging from activities, biometric info, demographic info, socio-economic info, hospitalization history, medication info, lifestyle info, etc. The AI model yields high performance of 87% accuracy and 88 sensitivity with 20+ features. This hybrid-AI framework is proven to be effective in identifying the potentially preventable hospitalization. Further, the high indicative features are identified by the models which guide us to a healthy lifestyle and early intervention suggestions.

Keywords: hospitalization prevention, machine learning, remote patient monitoring, risk prediction

Procedia PDF Downloads 230
1804 Noise Mitigation Techniques to Minimize Electromagnetic Interference/Electrostatic Discharge Effects for the Lunar Mission Spacecraft

Authors: Vabya Kumar Pandit, Mudit Mittal, N. Prahlad Rao, Ramnath Babu

Abstract:

TeamIndus is the only Indian team competing for the Google Lunar XPRIZE(GLXP). The GLXP is a global competition to challenge the private entities to soft land a rover on the moon, travel minimum 500 meters and transmit high definition images and videos to Earth. Towards this goal, the TeamIndus strategy is to design and developed lunar lander that will deliver a rover onto the surface of the moon which will accomplish GLXP mission objectives. This paper showcases the various system level noise control techniques adopted by Electrical Distribution System (EDS), to achieve the required Electromagnetic Compatibility (EMC) of the spacecraft. The design guidelines followed to control Electromagnetic Interference by proper electronic package design, grounding, shielding, filtering, and cable routing within the stipulated mass budget, are explained. The paper also deals with the challenges of achieving Electromagnetic Cleanliness in presence of various Commercial Off-The-Shelf (COTS) and In-House developed components. The methods of minimizing Electrostatic Discharge (ESD) by identifying the potential noise sources, susceptible areas for charge accumulation and the methodology to prevent arcing inside spacecraft are explained. The paper then provides the EMC requirements matrix derived from the mission requirements to meet the overall Electromagnetic compatibility of the Spacecraft.

Keywords: electromagnetic compatibility, electrostatic discharge, electrical distribution systems, grounding schemes, light weight harnessing

Procedia PDF Downloads 293
1803 Intelligent Recognition of Diabetes Disease via FCM Based Attribute Weighting

Authors: Kemal Polat

Abstract:

In this paper, an attribute weighting method called fuzzy C-means clustering based attribute weighting (FCMAW) for classification of Diabetes disease dataset has been used. The aims of this study are to reduce the variance within attributes of diabetes dataset and to improve the classification accuracy of classifier algorithm transforming from non-linear separable datasets to linearly separable datasets. Pima Indians Diabetes dataset has two classes including normal subjects (500 instances) and diabetes subjects (268 instances). Fuzzy C-means clustering is an improved version of K-means clustering method and is one of most used clustering methods in data mining and machine learning applications. In this study, as the first stage, fuzzy C-means clustering process has been used for finding the centers of attributes in Pima Indians diabetes dataset and then weighted the dataset according to the ratios of the means of attributes to centers of theirs. Secondly, after weighting process, the classifier algorithms including support vector machine (SVM) and k-NN (k- nearest neighbor) classifiers have been used for classifying weighted Pima Indians diabetes dataset. Experimental results show that the proposed attribute weighting method (FCMAW) has obtained very promising results in the classification of Pima Indians diabetes dataset.

Keywords: fuzzy C-means clustering, fuzzy C-means clustering based attribute weighting, Pima Indians diabetes, SVM

Procedia PDF Downloads 413
1802 A Machine Learning Model for Predicting Students’ Academic Performance in Higher Institutions

Authors: Emmanuel Osaze Oshoiribhor, Adetokunbo MacGregor John-Otumu

Abstract:

There has been a need in recent years to predict student academic achievement prior to graduation. This is to assist them in improving their grades, especially for those who have struggled in the past. The purpose of this research is to use supervised learning techniques to create a model that predicts student academic progress. Many scholars have developed models that predict student academic achievement based on characteristics including smoking, demography, culture, social media, parent educational background, parent finances, and family background, to mention a few. This element, as well as the model used, could have misclassified the kids in terms of their academic achievement. As a prerequisite to predicting if the student will perform well in the future on related courses, this model is built using a logistic regression classifier with basic features such as the previous semester's course score, attendance to class, class participation, and the total number of course materials or resources the student is able to cover per semester. With a 96.7 percent accuracy, the model outperformed other classifiers such as Naive bayes, Support vector machine (SVM), Decision Tree, Random forest, and Adaboost. This model is offered as a desktop application with user-friendly interfaces for forecasting student academic progress for both teachers and students. As a result, both students and professors are encouraged to use this technique to predict outcomes better.

Keywords: artificial intelligence, ML, logistic regression, performance, prediction

Procedia PDF Downloads 109
1801 Developing a Machine Learning-based Cost Prediction Model for Construction Projects using Particle Swarm Optimization

Authors: Soheila Sadeghi

Abstract:

Accurate cost prediction is essential for effective project management and decision-making in the construction industry. This study aims to develop a cost prediction model for construction projects using Machine Learning techniques and Particle Swarm Optimization (PSO). The research utilizes a comprehensive dataset containing project cost estimates, actual costs, resource details, and project performance metrics from a road reconstruction project. The methodology involves data preprocessing, feature selection, and the development of an Artificial Neural Network (ANN) model optimized using PSO. The study investigates the impact of various input features, including cost estimates, resource allocation, and project progress, on the accuracy of cost predictions. The performance of the optimized ANN model is evaluated using metrics such as Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and R-squared. The results demonstrate the effectiveness of the proposed approach in predicting project costs, outperforming traditional benchmark models. The feature selection process identifies the most influential variables contributing to cost variations, providing valuable insights for project managers. However, this study has several limitations. Firstly, the model's performance may be influenced by the quality and quantity of the dataset used. A larger and more diverse dataset covering different types of construction projects would enhance the model's generalizability. Secondly, the study focuses on a specific optimization technique (PSO) and a single Machine Learning algorithm (ANN). Exploring other optimization methods and comparing the performance of various ML algorithms could provide a more comprehensive understanding of the cost prediction problem. Future research should focus on several key areas. Firstly, expanding the dataset to include a wider range of construction projects, such as residential buildings, commercial complexes, and infrastructure projects, would improve the model's applicability. Secondly, investigating the integration of additional data sources, such as economic indicators, weather data, and supplier information, could enhance the predictive power of the model. Thirdly, exploring the potential of ensemble learning techniques, which combine multiple ML algorithms, may further improve cost prediction accuracy. Additionally, developing user-friendly interfaces and tools to facilitate the adoption of the proposed cost prediction model in real-world construction projects would be a valuable contribution to the industry. The findings of this study have significant implications for construction project management, enabling proactive cost estimation, resource allocation, budget planning, and risk assessment, ultimately leading to improved project performance and cost control. This research contributes to the advancement of cost prediction techniques in the construction industry and highlights the potential of Machine Learning and PSO in addressing this critical challenge. However, further research is needed to address the limitations and explore the identified future research directions to fully realize the potential of ML-based cost prediction models in the construction domain.

Keywords: cost prediction, construction projects, machine learning, artificial neural networks, particle swarm optimization, project management, feature selection, road reconstruction

Procedia PDF Downloads 59
1800 Web Development in Information Technology with Javascript, Machine Learning and Artificial Intelligence

Authors: Abdul Basit Kiani, Maryam Kiani

Abstract:

Online developers now have the tools necessary to create online apps that are not only reliable but also highly interactive, thanks to the introduction of JavaScript frameworks and APIs. The objective is to give a broad overview of the recent advances in the area. The fusion of machine learning (ML) and artificial intelligence (AI) has expanded the possibilities for web development. Modern websites now include chatbots, clever recommendation systems, and customization algorithms built in. In the rapidly evolving landscape of modern websites, it has become increasingly apparent that user engagement and personalization are key factors for success. To meet these demands, websites now incorporate a range of innovative technologies. One such technology is chatbots, which provide users with instant assistance and support, enhancing their overall browsing experience. These intelligent bots are capable of understanding natural language and can answer frequently asked questions, offer product recommendations, and even help with troubleshooting. Moreover, clever recommendation systems have emerged as a powerful tool on modern websites. By analyzing user behavior, preferences, and historical data, these systems can intelligently suggest relevant products, articles, or services tailored to each user's unique interests. This not only saves users valuable time but also increases the chances of conversions and customer satisfaction. Additionally, customization algorithms have revolutionized the way websites interact with users. By leveraging user preferences, browsing history, and demographic information, these algorithms can dynamically adjust the website's layout, content, and functionalities to suit individual user needs. This level of personalization enhances user engagement, boosts conversion rates, and ultimately leads to a more satisfying online experience. In summary, the integration of chatbots, clever recommendation systems, and customization algorithms into modern websites is transforming the way users interact with online platforms. These advanced technologies not only streamline user experiences but also contribute to increased customer satisfaction, improved conversions, and overall website success.

Keywords: Javascript, machine learning, artificial intelligence, web development

Procedia PDF Downloads 80
1799 talk2all: A Revolutionary Tool for International Medical Tourism

Authors: Madhukar Kasarla, Sumit Fogla, Kiran Panuganti, Gaurav Jain, Abhijit Ramanujam, Astha Jain, Shashank Kraleti, Sharat Musham, Arun Chaudhury

Abstract:

Patients have often chosen to travel for care — making pilgrimages to academic meccas and state-of-the-art hospitals for sophisticated surgery. This culture is still persistent in the landscape of US healthcare, with hundred thousand of visitors coming to the shores of United States to seek the high quality of medical care. One of the major challenges in this form of medical tourism has been the language barrier. Thus, an Iraqi patient, with immediate needs of communicating the healthcare needs to the treating team in the hospital, may face huge barrier in effective patient-doctor communication, delaying care and even at times reducing the quality. To circumvent these challenges, we are proposing the use of a state-of-the-art tool, Talk2All, which can translate nearly one hundred international languages (and even sign language) in real time. The tool is an easy to download app and highly user friendly. It builds on machine learning principles to decode different languages in real time. We suggest that the use of Talk2All will tremendously enhance communication in the hospital setting, effectively breaking the language barrier. We propose that vigorous incorporation of Talk2All shall overcome practical challenges in international medical and surgical tourism.

Keywords: language translation, communication, machine learning, medical tourism

Procedia PDF Downloads 214
1798 Modeling Aeration of Sharp Crested Weirs by Using Support Vector Machines

Authors: Arun Goel

Abstract:

The present paper attempts to investigate the prediction of air entrainment rate and aeration efficiency of a free over-fall jets issuing from a triangular sharp crested weir by using regression based modelling. The empirical equations, support vector machine (polynomial and radial basis function) models and the linear regression techniques were applied on the triangular sharp crested weirs relating the air entrainment rate and the aeration efficiency to the input parameters namely drop height, discharge, and vertex angle. It was observed that there exists a good agreement between the measured values and the values obtained using empirical equations, support vector machine (Polynomial and rbf) models, and the linear regression techniques. The test results demonstrated that the SVM based (Poly & rbf) model also provided acceptable prediction of the measured values with reasonable accuracy along with empirical equations and linear regression techniques in modelling the air entrainment rate and the aeration efficiency of a free over-fall jets issuing from triangular sharp crested weir. Further sensitivity analysis has also been performed to study the impact of input parameter on the output in terms of air entrainment rate and aeration efficiency.

Keywords: air entrainment rate, dissolved oxygen, weir, SVM, regression

Procedia PDF Downloads 436
1797 Alphabet Recognition Using Pixel Probability Distribution

Authors: Vaidehi Murarka, Sneha Mehta, Dishant Upadhyay

Abstract:

Our project topic is “Alphabet Recognition using pixel probability distribution”. The project uses techniques of Image Processing and Machine Learning in Computer Vision. Alphabet recognition is the mechanical or electronic translation of scanned images of handwritten, typewritten or printed text into machine-encoded text. It is widely used to convert books and documents into electronic files etc. Alphabet Recognition based OCR application is sometimes used in signature recognition which is used in bank and other high security buildings. One of the popular mobile applications includes reading a visiting card and directly storing it to the contacts. OCR's are known to be used in radar systems for reading speeders license plates and lots of other things. The implementation of our project has been done using Visual Studio and Open CV (Open Source Computer Vision). Our algorithm is based on Neural Networks (machine learning). The project was implemented in three modules: (1) Training: This module aims “Database Generation”. Database was generated using two methods: (a) Run-time generation included database generation at compilation time using inbuilt fonts of OpenCV library. Human intervention is not necessary for generating this database. (b) Contour–detection: ‘jpeg’ template containing different fonts of an alphabet is converted to the weighted matrix using specialized functions (contour detection and blob detection) of OpenCV. The main advantage of this type of database generation is that the algorithm becomes self-learning and the final database requires little memory to be stored (119kb precisely). (2) Preprocessing: Input image is pre-processed using image processing concepts such as adaptive thresholding, binarizing, dilating etc. and is made ready for segmentation. “Segmentation” includes extraction of lines, words, and letters from the processed text image. (3) Testing and prediction: The extracted letters are classified and predicted using the neural networks algorithm. The algorithm recognizes an alphabet based on certain mathematical parameters calculated using the database and weight matrix of the segmented image.

Keywords: contour-detection, neural networks, pre-processing, recognition coefficient, runtime-template generation, segmentation, weight matrix

Procedia PDF Downloads 389
1796 Hardware in the Loop Platform for Virtual Commissioning: Case Study of a Hydraulic-Press Model Simulated in Real-Time

Authors: Jorge Rodriguez-Guerra, Carlos Calleja, Aron Pujana, Ana Maria Macarulla

Abstract:

Hydraulic-press commissioning consumes a great amount of man-hours, due to the fact that it takes place several miles away from where it has been designed. This factor became exacerbated due to control designers’ lack of knowledge about which will be the final controller gains before they start working with it. Virtual commissioning has been postulated as an optimal solution to deal with this lack of knowledge. Here, a case study is presented in which a controller is set up against a real-time model based on a hydraulic-press. The press model is designed following manufacturer specifications and it is embedded in a real-time simulator. This methodology ensures that the model achieves similar responses as the real machine that would be placed on the industry. A deterministic communication protocol is in charge of the bidirectional information transmission between the real-time model and the controller. This platform allows the engineer to test and verify the final control responses with exactly the same hardware that is going to be installed in the hydraulic-press, in other words, realize a virtual commissioning of the electro-hydraulic actuator. The Hardware in the Loop (HiL) platform validates in laboratory conditions and harmless for the machine the control algorithms designed, which allows embedding them afterwards in the industrial environment without further modifications.

Keywords: deterministic communication protocol, electro-hydraulic actuator, hardware in the loop, real-time, virtual commissioning

Procedia PDF Downloads 143
1795 Identifying Autism Spectrum Disorder Using Optimization-Based Clustering

Authors: Sharifah Mousli, Sona Taheri, Jiayuan He

Abstract:

Autism spectrum disorder (ASD) is a complex developmental condition involving persistent difficulties with social communication, restricted interests, and repetitive behavior. The challenges associated with ASD can interfere with an affected individual’s ability to function in social, academic, and employment settings. Although there is no effective medication known to treat ASD, to our best knowledge, early intervention can significantly improve an affected individual’s overall development. Hence, an accurate diagnosis of ASD at an early phase is essential. The use of machine learning approaches improves and speeds up the diagnosis of ASD. In this paper, we focus on the application of unsupervised clustering methods in ASD as a large volume of ASD data generated through hospitals, therapy centers, and mobile applications has no pre-existing labels. We conduct a comparative analysis using seven clustering approaches such as K-means, agglomerative hierarchical, model-based, fuzzy-C-means, affinity propagation, self organizing maps, linear vector quantisation – as well as the recently developed optimization-based clustering (COMSEP-Clust) approach. We evaluate the performances of the clustering methods extensively on real-world ASD datasets encompassing different age groups: toddlers, children, adolescents, and adults. Our experimental results suggest that the COMSEP-Clust approach outperforms the other seven methods in recognizing ASD with well-separated clusters.

Keywords: autism spectrum disorder, clustering, optimization, unsupervised machine learning

Procedia PDF Downloads 115
1794 Charting Sentiments with Naive Bayes and Logistic Regression

Authors: Jummalla Aashrith, N. L. Shiva Sai, K. Bhavya Sri

Abstract:

The swift progress of web technology has not only amassed a vast reservoir of internet data but also triggered a substantial surge in data generation. The internet has metamorphosed into one of the dynamic hubs for online education, idea dissemination, as well as opinion-sharing. Notably, the widely utilized social networking platform Twitter is experiencing considerable expansion, providing users with the ability to share viewpoints, participate in discussions spanning diverse communities, and broadcast messages on a global scale. The upswing in online engagement has sparked a significant curiosity in subjective analysis, particularly when it comes to Twitter data. This research is committed to delving into sentiment analysis, focusing specifically on the realm of Twitter. It aims to offer valuable insights into deciphering information within tweets, where opinions manifest in a highly unstructured and diverse manner, spanning a spectrum from positivity to negativity, occasionally punctuated by neutrality expressions. Within this document, we offer a comprehensive exploration and comparative assessment of modern approaches to opinion mining. Employing a range of machine learning algorithms such as Naive Bayes and Logistic Regression, our investigation plunges into the domain of Twitter data streams. We delve into overarching challenges and applications inherent in the realm of subjectivity analysis over Twitter.

Keywords: machine learning, sentiment analysis, visualisation, python

Procedia PDF Downloads 56
1793 Dynamic Cellular Remanufacturing System (DCRS) Design

Authors: Tariq Aljuneidi, Akif Asil Bulgak

Abstract:

Remanufacturing may be defined as the process of bringing used products to “like-new” functional state with warranty to match, and it is one of the most popular product end-of-life scenarios. An efficient remanufacturing network lead to an efficient design of sustainable manufacturing enterprise. In remanufacturing network, products are collected from the customer zone, disassembled and remanufactured at a suitable remanufacturing facility. In this respect, another issue to consider is how the returned product to be remanufactured, in other words, what is the best layout for such facility. In order to achieve a sustainable manufacturing system, Cellular Manufacturing System (CMS) designs are highly recommended, CMSs combine high throughput rates of line layouts with the flexibility offered by functional layouts (job shop). Introducing the CMS while designing a remanufacturing network will benefit the utilization of such a network. This paper presents and analyzes a comprehensive mathematical model for the design of Dynamic Cellular Remanufacturing Systems (DCRSs). In this paper, the proposed model is the first one to date that consider CMS and remanufacturing system simultaneously. The proposed DCRS model considers several manufacturing attributes such as multi-period production planning, dynamic system reconfiguration, duplicate machines, machine capacity, available time for workers, worker assignments, and machine procurement, where the demand is totally satisfied from a returned product. A numerical example is presented to illustrate the proposed model.

Keywords: cellular manufacturing system, remanufacturing, mathematical programming, sustainability

Procedia PDF Downloads 378
1792 Performance Analysis of 5G for Low Latency Transmission Based on Universal Filtered Multi-Carrier Technique and Interleave Division Multiple Access

Authors: A. Asgharzadeh, M. Maroufi

Abstract:

5G mobile communication system has drawn more and more attention. The 5G system needs to provide three different types of services, including enhanced Mobile BroadBand (eMBB), massive machine-type communication (mMTC), and ultra-reliable and low-latency communication (URLLC). Universal Filtered Multi-Carrier (UFMC), Filter Bank Multicarrier (FBMC), and Filtered Orthogonal Frequency Division Multiplexing (f-OFDM) are suggested as a well-known candidate waveform for the coming 5G system. Themachine-to-machine (M2M) communications are one of the essential applications in 5G, and it involves exchanging of concise messages with a very short latency. However, in UFMC systems, the subcarriers are grouped into subbands but f-OFDM only one subband covers the entire band. Furthermore, in FBMC, a subband includes only one subcarrier, and the number of subbands is the same as the number of subcarriers. This paper mainly discusses the performance of UFMC with different parameters for the UFMC system. Also, paper shows that UFMC is the best choice outperforming OFDM in any case and FBMC in case of very short packets while performing similarly for long sequences with channel estimation techniques for Interleave Division Multiple Access (IDMA) systems.

Keywords: universal filtered multi-carrier technique, UFMC, interleave division multiple access, IDMA, fifth-generation, subband

Procedia PDF Downloads 134
1791 Estimation of the Exergy-Aggregated Value Generated by a Manufacturing Process Using the Theory of the Exergetic Cost

Authors: German Osma, Gabriel Ordonez

Abstract:

The production of metal-rubber spares for vehicles is a sequential process that consists in the transformation of raw material through cutting activities and chemical and thermal treatments, which demand electricity and fossil fuels. The energy efficiency analysis for these cases is mostly focused on studying of each machine or production step, but is not common to study of the quality of the production process achieves from aggregated value viewpoint, which can be used as a quality measurement for determining of impact on the environment. In this paper, the theory of exergetic cost is used for determining of aggregated exergy to three metal-rubber spares, from an exergy analysis and thermoeconomic analysis. The manufacturing processing of these spares is based into batch production technique, and therefore is proposed the use of this theory for discontinuous flows from of single models of workstations; subsequently, the complete exergy model of each product is built using flowcharts. These models are a representation of exergy flows between components into the machines according to electrical, mechanical and/or thermal expressions; they determine the demanded exergy to produce the effective transformation in raw materials (aggregated exergy value), the exergy losses caused by equipment and irreversibilities. The energy resources of manufacturing process are electricity and natural gas. The workstations considered are lathes, punching presses, cutters, zinc machine, chemical treatment tanks, hydraulic vulcanizing presses and rubber mixer. The thermoeconomic analysis was done by workstation and by spare; first of them describes the operation of the components of each machine and where the exergy losses are; while the second of them estimates the exergy-aggregated value for finished product and wasted feedstock. Results indicate that exergy efficiency of a mechanical workstation is between 10% and 60% while this value in the thermal workstations is less than 5%; also that each effective exergy-aggregated value is one-thirtieth of total exergy required for operation of manufacturing process, which amounts approximately to 2 MJ. These troubles are caused mainly by technical limitations of machines, oversizing of metal feedstock that demands more mechanical transformation work, and low thermal insulation of chemical treatment tanks and hydraulic vulcanizing presses. From established information, in this case, it is possible to appreciate the usefulness of theory of exergetic cost for analyzing of aggregated value in manufacturing processes.

Keywords: exergy-aggregated value, exergy efficiency, thermoeconomics, exergy modeling

Procedia PDF Downloads 170
1790 Parametric Study for Optimal Design of Hybrid Bridge Joint

Authors: Bongsik Park, Jae Hyun Park, Jae-Yeol Cho

Abstract:

Mixed structure, which is a kind of hybrid system, is incorporating steel beam and prestressed concrete beam. Hybrid bridge adopting mixed structure have some merits. Main span length can be made longer by using steel as main span material. In case of cable-stayed bridge having asymmetric span length, negative reaction at side span can be restrained without extra restraining devices by using weight difference between main span material and side span material. However angle of refraction might happen because of rigidity difference between materials and stress concentration also might happen because of abnormal loading transmission at joint in the hybrid bridge. Therefore the joint might be a weak point of the structural system and it needs to pay attention to design of the joint. However, design codes and standards about the joint in the hybrid-bridge have not been established so the joint designs in most of construction cases have been very conservative or followed previous design without extra verification. In this study parametric study using finite element analysis for optimal design of hybrid bridge joint is conducted. Before parametric study, finite element analysis was conducted based on previous experimental data and it is verified that analysis result approximated experimental data. Based on the finite element analysis results, parametric study was conducted. The parameters were selected as those have influences on joint behavior. Based on the parametric study results, optimal design of hybrid bridge joint has been determined.

Keywords: parametric study, optimal design, hybrid bridge, finite element analysis

Procedia PDF Downloads 425
1789 Hazard Alert in Malaysia Related to Occupational Safety and Health

Authors: Atikah Binti Azudin, Nurin Nazlah Binti Muhamad Yani, Nur Alya Nadhirah Binti Naaidith, Nur Amylia Wahida Binti Mat Ayob, Nurshamimi Shakirah Binti Suboh, Nur Auni Batrisyia Binti Md. Zaini, Nur Aziemah Binti Mohamad, Nurul Suffiyah Binti Sa’Dun, Sabrina Sasha Izzati Binti Zubaile, Umi Huwaina Binti Ahmiruddin, Wan Nur Shafawati Binti Wan Ghazali

Abstract:

A hazard alert is intended to provide brief information about significant incidents or existing difficulties in Department workplaces. The alert gives guidelines for proper processes, practices, and controls to be applied. When operated in accordance with the manufacturer's instructions, any machine or tool utilized at work provides a safe and dependable platform for workers to accomplish job duties. However, when not utilized appropriately, the machine might pose a major hazard to employees. Employers have a duty to keep employees safe in this scenario. This Hazard Alert outlines specific occupational dangers and the controls that employers must apply to prevent injury or fatal accidents. There have been several cases of hazard alerts in Malaysia, which have had a negative impact on a few workers. Looking on the bright side, we can overcome every incident in a variety of ways. One of these is that only qualified individuals operate mobile machinery and equipment. In addition, employees may also perform frequent pre-use inspections of machinery to discover and fix flaws. Hazard alert is very important, and this study would cover a variety of subjects, including the methods employed.

Keywords: safe, hazard, impacts, duties.

Procedia PDF Downloads 92
1788 Ensemble Methods in Machine Learning: An Algorithmic Approach to Derive Distinctive Behaviors of Criminal Activity Applied to the Poaching Domain

Authors: Zachary Blanks, Solomon Sonya

Abstract:

Poaching presents a serious threat to endangered animal species, environment conservations, and human life. Additionally, some poaching activity has even been linked to supplying funds to support terrorist networks elsewhere around the world. Consequently, agencies dedicated to protecting wildlife habitats have a near intractable task of adequately patrolling an entire area (spanning several thousand kilometers) given limited resources, funds, and personnel at their disposal. Thus, agencies need predictive tools that are both high-performing and easily implementable by the user to help in learning how the significant features (e.g. animal population densities, topography, behavior patterns of the criminals within the area, etc) interact with each other in hopes of abating poaching. This research develops a classification model using machine learning algorithms to aid in forecasting future attacks that is both easy to train and performs well when compared to other models. In this research, we demonstrate how data imputation methods (specifically predictive mean matching, gradient boosting, and random forest multiple imputation) can be applied to analyze data and create significant predictions across a varied data set. Specifically, we apply these methods to improve the accuracy of adopted prediction models (Logistic Regression, Support Vector Machine, etc). Finally, we assess the performance of the model and the accuracy of our data imputation methods by learning on a real-world data set constituting four years of imputed data and testing on one year of non-imputed data. This paper provides three main contributions. First, we extend work done by the Teamcore and CREATE (Center for Risk and Economic Analysis of Terrorism Events) research group at the University of Southern California (USC) working in conjunction with the Department of Homeland Security to apply game theory and machine learning algorithms to develop more efficient ways of reducing poaching. This research introduces ensemble methods (Random Forests and Stochastic Gradient Boosting) and applies it to real-world poaching data gathered from the Ugandan rain forest park rangers. Next, we consider the effect of data imputation on both the performance of various algorithms and the general accuracy of the method itself when applied to a dependent variable where a large number of observations are missing. Third, we provide an alternate approach to predict the probability of observing poaching both by season and by month. The results from this research are very promising. We conclude that by using Stochastic Gradient Boosting to predict observations for non-commercial poaching by season, we are able to produce statistically equivalent results while being orders of magnitude faster in computation time and complexity. Additionally, when predicting potential poaching incidents by individual month vice entire seasons, boosting techniques produce a mean area under the curve increase of approximately 3% relative to previous prediction schedules by entire seasons.

Keywords: ensemble methods, imputation, machine learning, random forests, statistical analysis, stochastic gradient boosting, wildlife protection

Procedia PDF Downloads 292