Search results for: machine performance
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14555

Search results for: machine performance

14015 The Impact of Environmental Social and Governance (ESG) on Corporate Financial Performance (CFP): Evidence from New Zealand Companies

Authors: Muhammad Akhtaruzzaman

Abstract:

The impact of corporate environmental social and governance (ESG) on financial performance is often difficult to quantify despite the ESG related theories predict that ESG performance improves financial performance of a company. This research examines the link between corporate ESG performance and the financial performance of the NZX (New Zealand Stock Exchange) listed companies. For this purpose, this research utilizes mixed methods approaches to examine and understand this link. While quantitative results found no robust evidence of such a link, however, the qualitative analysis of content data suggests a strong cooccurrence exists between ESG performance and financial performance. The findings of this research have important implications for policymakers to support higher ESG-performing companies and for management practitioners to develop ESG-related strategies.

Keywords: ESG, financial performance, New Zealand firms, thematic analysis, mixed methods

Procedia PDF Downloads 51
14014 An Improvement of Flow Forming Process for Pressure Vessels by Four Rollers Machine

Authors: P. Sawitri, S. Cdr. Sittha, T. Kritsana

Abstract:

Flow forming is widely used in many industries, especially in defence technology industries. Pressure vessels requirements are high precision, light weight, seamless and optimum strength. For large pressure vessels, flow forming by 3 rollers machine were used. In case of long range rocket motor case flow forming and welding of pressure vessels have been used for manufacturing. Due to complication of welding process, researchers had developed 4 meters length pressure vessels without weldment by 4 rollers flow forming machine. Design and preparation of preform work pieces are performed. The optimization of flow forming parameter such as feed rate, spindle speed and depth of cut will be discussed. The experimental result shown relation of flow forming parameters to quality of flow formed tube and prototype pressure vessels have been made.

Keywords: flow forming, pressure vessel, four rollers, feed rate, spindle speed, cold work

Procedia PDF Downloads 319
14013 A Review of Physiological Measures for Cognitive Workload Assessment of Aircrew

Authors: Naveed Tahir, Adnan Maqsood

Abstract:

Cognitive workload is a significant factor affecting user performance, and it has been broadly investigated for its application in ergonomics as well as in designing and optimizing effective human-machine interactions. It is mentally challenging to maneuver an aircraft, and pilots must control the aircraft and adequately communicate to the verbal-auditory stimuli. Several physiological measures have long been researched and used to demonstrate the cognitive workload. In our current study, we have summarized recent findings of the effectiveness, accuracy, and applicability of commonly used physiological measures in evaluating cognitive workload. We have also highlighted on the advancements in physiological measures. The strength and limitations of physiological measures have also been discussed to assess the cognitive workload of people, especially the aircrews in laboratory settings and real-time situations. We have presented the research findings of the physiological measures to base suggestions on the proper applications of the measures and settings demanding the use of single measure or their combinations.

Keywords: aircrew, cognitive workload, subjective measure, physiological measure, performance measure

Procedia PDF Downloads 145
14012 Efficient Fake News Detection Using Machine Learning and Deep Learning Approaches

Authors: Chaima Babi, Said Gadri

Abstract:

The rapid increase in fake news continues to grow at a very fast rate; this requires implementing efficient techniques that allow testing the re-liability of online content. For that, the current research strives to illuminate the fake news problem using deep learning DL and machine learning ML ap-proaches. We have developed the traditional LSTM (Long short-term memory), and the bidirectional BiLSTM model. A such process is to perform a training task on almost of samples of the dataset, validate the model on a subset called the test set to provide an unbiased evaluation of the final model fit on the training dataset, then compute the accuracy of detecting classifica-tion and comparing the results. For the programming stage, we used Tensor-Flow and Keras libraries on Python to support Graphical Processing Units (GPUs) that are being used for developing deep learning applications.

Keywords: machine learning, deep learning, natural language, fake news, Bi-LSTM, LSTM, multiclass classification

Procedia PDF Downloads 77
14011 A Framework of Virtualized Software Controller for Smart Manufacturing

Authors: Pin Xiu Chen, Shang Liang Chen

Abstract:

A virtualized software controller is developed in this research to replace traditional hardware control units. This virtualized software controller transfers motion interpolation calculations from the motion control units of end devices to edge computing platforms, thereby reducing the end devices' computational load and hardware requirements and making maintenance and updates easier. The study also applies the concept of microservices, dividing the control system into several small functional modules and then deploy into a cloud data server. This reduces the interdependency among modules and enhances the overall system's flexibility and scalability. Finally, with containerization technology, the system can be deployed and started in a matter of seconds, which is more efficient than traditional virtual machine deployment methods. Furthermore, this virtualized software controller communicates with end control devices via wireless networks, making the placement of production equipment or the redesign of processes more flexible and no longer limited by physical wiring. To handle the large data flow and maintain low-latency transmission, this study integrates 5G technology, fully utilizing its high speed, wide bandwidth, and low latency features to achieve rapid and stable remote machine control. An experimental setup is designed to verify the feasibility and test the performance of this framework. This study designs a smart manufacturing site with a 5G communication architecture, serving as a field for experimental data collection and performance testing. The smart manufacturing site includes one robotic arm, three Computer Numerical Control machine tools, several Input/Output ports, and an edge computing architecture. All machinery information is uploaded to edge computing servers and cloud servers via 5G communication and the Internet of Things framework. After analysis and computation, this information is converted into motion control commands, which are transmitted back to the relevant machinery for motion control through 5G communication. The communication time intervals at each stage are calculated using the C++ chrono library to measure the time difference for each command transmission. The relevant test results will be organized and displayed in the full-text.

Keywords: 5G, MEC, microservices, virtualized software controller, smart manufacturing

Procedia PDF Downloads 70
14010 Human Digital Twin for Personal Conversation Automation Using Supervised Machine Learning Approaches

Authors: Aya Salama

Abstract:

Digital Twin is an emerging research topic that attracted researchers in the last decade. It is used in many fields, such as smart manufacturing and smart healthcare because it saves time and money. It is usually related to other technologies such as Data Mining, Artificial Intelligence, and Machine Learning. However, Human digital twin (HDT), in specific, is still a novel idea that still needs to prove its feasibility. HDT expands the idea of Digital Twin to human beings, which are living beings and different from the inanimate physical entities. The goal of this research was to create a Human digital twin that is responsible for real-time human replies automation by simulating human behavior. For this reason, clustering, supervised classification, topic extraction, and sentiment analysis were studied in this paper. The feasibility of the HDT for personal replies generation on social messaging applications was proved in this work. The overall accuracy of the proposed approach in this paper was 63% which is a very promising result that can open the way for researchers to expand the idea of HDT. This was achieved by using Random Forest for clustering the question data base and matching new questions. K-nearest neighbor was also applied for sentiment analysis.

Keywords: human digital twin, sentiment analysis, topic extraction, supervised machine learning, unsupervised machine learning, classification, clustering

Procedia PDF Downloads 82
14009 Machine Learning Approach for Mutation Testing

Authors: Michael Stewart

Abstract:

Mutation testing is a type of software testing proposed in the 1970s where program statements are deliberately changed to introduce simple errors so that test cases can be validated to determine if they can detect the errors. Test cases are executed against the mutant code to determine if one fails, detects the error and ensures the program is correct. One major issue with this type of testing was it became intensive computationally to generate and test all possible mutations for complex programs. This paper used reinforcement learning and parallel processing within the context of mutation testing for the selection of mutation operators and test cases that reduced the computational cost of testing and improved test suite effectiveness. Experiments were conducted using sample programs to determine how well the reinforcement learning-based algorithm performed with one live mutation, multiple live mutations and no live mutations. The experiments, measured by mutation score, were used to update the algorithm and improved accuracy for predictions. The performance was then evaluated on multiple processor computers. With reinforcement learning, the mutation operators utilized were reduced by 50 – 100%.

Keywords: automated-testing, machine learning, mutation testing, parallel processing, reinforcement learning, software engineering, software testing

Procedia PDF Downloads 187
14008 Effect of Communication Pattern on Agricultural Employees' Job Performance

Authors: B. G. Abiona, E. O. Fakoya, S. O. Adeogun, J. O. Blessed

Abstract:

This study assessed the influence of communication pattern on agricultural employees’ job performance. Data were collected from 61 randomly selected respondents using a structured questionnaire. Perceived communication pattern that influence job performance include: the attitude of the administrators (x̅ = 3.41, physical barriers to communication flow among employees (x̅ = 3.21). Major challenges to respondents’ job performance were different language among employees (x̅ = 3.12), employees perception on organizational issues (x̅ = 3.09), networking (x̅ = 2.88), and unclear definition of work (x̅ = 2.74). A significant relationship was found between employees’ perceived communication pattern (r = 0.423, p < 0.00) and job performance. Information must be well designed in such a way that would positively influence employees’ job performance as this is essential in any agricultural organizations.

Keywords: communication pattern, job performance, agricultural employees, constraint, administrators, attitude

Procedia PDF Downloads 344
14007 Smoker Recognition from Lung X-Ray Images Using Convolutional Neural Network

Authors: Moumita Chanda, Md. Fazlul Karim Patwary

Abstract:

Smoking is one of the most popular recreational drug use behaviors, and it contributes to birth defects, COPD, heart attacks, and erectile dysfunction. To completely eradicate this disease, it is imperative that it be identified and treated. Numerous smoking cessation programs have been created, and they demonstrate how beneficial it may be to help someone stop smoking at the ideal time. A tomography meter is an effective smoking detector. Other wearables, such as RF-based proximity sensors worn on the collar and wrist to detect when the hand is close to the mouth, have been proposed in the past, but they are not impervious to deceptive variables. In this study, we create a machine that can discriminate between smokers and non-smokers in real-time with high sensitivity and specificity by watching and collecting the human lung and analyzing the X-ray data using machine learning. If it has the highest accuracy, this machine could be utilized in a hospital, in the selection of candidates for the army or police, or in university entrance.

Keywords: CNN, smoker detection, non-smoker detection, OpenCV, artificial Intelligence, X-ray Image detection

Procedia PDF Downloads 73
14006 Algorithms Minimizing Total Tardiness

Authors: Harun Aydilek, Asiye Aydilek, Ali Allahverdi

Abstract:

The total tardiness is a widely used performance measure in the scheduling literature. This performance measure is particularly important in situations where there is a cost to complete a job beyond its due date. The cost of scheduling increases as the gap between a job's due date and its completion time increases. Such costs may also be penalty costs in contracts, loss of goodwill. This performance measure is important as the fulfillment of due dates of customers has to be taken into account while making scheduling decisions. The problem is addressed in the literature, however, it has been assumed zero setup times. Even though this assumption may be valid for some environments, it is not valid for some other scheduling environments. When setup times are treated as separate from processing times, it is possible to increase machine utilization and to reduce total tardiness. Therefore, non-zero setup times need to be considered as separate. A dominance relation is developed and several algorithms are proposed. The developed dominance relation is utilized in the proposed algorithms. Extensive computational experiments are conducted for the evaluation of the algorithms. The experiments indicated that the developed algorithms perform much better than the existing algorithms in the literature. More specifically, one of the newly proposed algorithms reduces the error of the best existing algorithm in the literature by 40 percent.

Keywords: algorithm, assembly flowshop, dominance relation, total tardiness

Procedia PDF Downloads 343
14005 Mental Health Diagnosis through Machine Learning Approaches

Authors: Md Rafiqul Islam, Ashir Ahmed, Anwaar Ulhaq, Abu Raihan M. Kamal, Yuan Miao, Hua Wang

Abstract:

Mental health of people is equally important as of their physical health. Mental health and well-being are influenced not only by individual attributes but also by the social circumstances in which people find themselves and the environment in which they live. Like physical health, there is a number of internal and external factors such as biological, social and occupational factors that could influence the mental health of people. People living in poverty, suffering from chronic health conditions, minority groups, and those who exposed to/or displaced by war or conflict are generally more likely to develop mental health conditions. However, to authors’ best knowledge, there is dearth of knowledge on the impact of workplace (especially the highly stressed IT/Tech workplace) on the mental health of its workers. This study attempts to examine the factors influencing the mental health of tech workers. A publicly available dataset containing more than 65,000 cells and 100 attributes is examined for this purpose. Number of machine learning techniques such as ‘Decision Tree’, ‘K nearest neighbor’ ‘Support Vector Machine’ and ‘Ensemble’, are then applied to the selected dataset to draw the findings. It is anticipated that the analysis reported in this study would contribute in presenting useful insights on the attributes contributing in the mental health of tech workers using relevant machine learning techniques.

Keywords: mental disorder, diagnosis, occupational stress, IT workplace

Procedia PDF Downloads 281
14004 Quantum Statistical Machine Learning and Quantum Time Series

Authors: Omar Alzeley, Sergey Utev

Abstract:

Minimizing a constrained multivariate function is the fundamental of Machine learning, and these algorithms are at the core of data mining and data visualization techniques. The decision function that maps input points to output points is based on the result of optimization. This optimization is the central of learning theory. One approach to complex systems where the dynamics of the system is inferred by a statistical analysis of the fluctuations in time of some associated observable is time series analysis. The purpose of this paper is a mathematical transition from the autoregressive model of classical time series to the matrix formalization of quantum theory. Firstly, we have proposed a quantum time series model (QTS). Although Hamiltonian technique becomes an established tool to detect a deterministic chaos, other approaches emerge. The quantum probabilistic technique is used to motivate the construction of our QTS model. The QTS model resembles the quantum dynamic model which was applied to financial data. Secondly, various statistical methods, including machine learning algorithms such as the Kalman filter algorithm, are applied to estimate and analyses the unknown parameters of the model. Finally, simulation techniques such as Markov chain Monte Carlo have been used to support our investigations. The proposed model has been examined by using real and simulated data. We establish the relation between quantum statistical machine and quantum time series via random matrix theory. It is interesting to note that the primary focus of the application of QTS in the field of quantum chaos was to find a model that explain chaotic behaviour. Maybe this model will reveal another insight into quantum chaos.

Keywords: machine learning, simulation techniques, quantum probability, tensor product, time series

Procedia PDF Downloads 462
14003 Response Surface Methodology for the Optimization of Paddy Husker by Medium Brown Rice Peeling Machine 6 Rubber Type

Authors: S. Bangphan, P. Bangphan, C. Ketsombun, T. Sammana

Abstract:

Optimization of response surface methodology (RSM) was employed to study the effects of three factor (rubber of clearance, spindle of speed, and rice of moisture) in brown rice peeling machine of the optimal good rice yield (99.67, average of three repeats). The optimized composition derived from RSM regression was analyzed using Regression analysis and Analysis of Variance (ANOVA). At a significant level α=0.05, the values of Regression coefficient, R2 adjust were 96.55% and standard deviation were 1.05056. The independent variables are initial rubber of clearance, spindle of speed and rice of moisture parameters namely. The investigating responses are final rubber clearance, spindle of speed and moisture of rice.

Keywords: brown rice, response surface methodology (RSM), peeling machine, optimization, paddy husker

Procedia PDF Downloads 564
14002 Cutting Performance of BDD Coating on WC-Co Tools

Authors: Feng Xu, Zhaozhi Liu, Junhua Xu, Xiaolong Tang, Dunwen Zuo

Abstract:

Chemical vapor deposition (CVD) diamond coated cutting tool has excellent cutting performance, it is the most ideal tool for the processing of nonferrous metals and alloys, composites, nonmetallic materials and other difficult-to-machine materials efficiently and accurately. Depositing CVD diamond coating on the cemented carbide with high cobalt content can improve its toughness and strength, therefore, it is very important to research on the preparation technology and cutting properties of CVD diamond coated cemented carbide cutting tool with high cobalt content. The preparation technology of boron-doped diamond (BDD) coating has been studied and the coated drills were prepared. BDD coating were deposited on the drills by using the optimized parameters and the SEM results show that there are no cracks or collapses in the coating. Cutting tests with the prepared drills against the silumin and aluminum base printed circuit board (PCB) have been studied. The results show that the wear amount of the coated drill is small and the machined surface has a better precision. The coating does not come off during the test, which shows good adhesion and cutting performance of the drill.

Keywords: cemented carbide with high cobalt content, CVD boron-doped diamond, cutting test, drill

Procedia PDF Downloads 429
14001 Comparison of Two Home Sleep Monitors Designed for Self-Use

Authors: Emily Wood, James K. Westphal, Itamar Lerner

Abstract:

Background: Polysomnography (PSG) recordings are regularly used in research and clinical settings to study sleep and sleep-related disorders. Typical PSG studies are conducted in professional laboratories and performed by qualified researchers. However, the number of sleep labs worldwide is disproportionate to the increasing number of individuals with sleep disorders like sleep apnea and insomnia. Consequently, there is a growing need to supply cheaper yet reliable means to measure sleep, preferably autonomously by subjects in their own home. Over the last decade, a variety of devices for self-monitoring of sleep became available in the market; however, very few have been directly validated against PSG to demonstrate their ability to perform reliable automatic sleep scoring. Two popular mobile EEG-based systems that have published validation results, the DREEM 3 headband and the Z-Machine, have never been directly compared one to the other by independent researchers. The current study aimed to compare the performance of DREEM 3 and the Z-Machine to help investigators and clinicians decide which of these devices may be more suitable for their studies. Methods: 26 participants have completed the study for credit or monetary compensation. Exclusion criteria included any history of sleep, neurological or psychiatric disorders. Eligible participants arrived at the lab in the afternoon and received the two devices. They then spent two consecutive nights monitoring their sleep at home. Participants were also asked to keep a sleep log, indicating the time they fell asleep, woke up, and the number of awakenings occurring during the night. Data from both devices, including detailed sleep hypnograms in 30-second epochs (differentiating Wake, combined N1/N2, N3; and Rapid Eye Movement sleep), were extracted and aligned upon retrieval. For analysis, the number of awakenings each night was defined as four or more consecutive wake epochs between sleep onset and termination. Total sleep time (TST) and the number of awakenings were compared to subjects’ sleep logs to measure consistency with the subjective reports. In addition, the sleep scores from each device were compared epoch-by-epoch to calculate the agreement between the two devices using Cohen’s Kappa. All analysis was performed using Matlab 2021b and SPSS 27. Results/Conclusion: Subjects consistently reported longer times spent asleep than the time reported by each device (M= 448 minutes for sleep logs compared to M= 406 and M= 345 minutes for the DREEM and Z-Machine, respectively; both ps<0.05). Linear correlations between the sleep log and each device were higher for the DREEM than the Z-Machine for both TST and the number of awakenings, and, likewise, the mean absolute bias between the sleep logs and each device was higher for the Z-Machine for both TST (p<0.001) and awakenings (p<0.04). There was some indication that these effects were stronger for the second night compared to the first night. Epoch-by-epoch comparisons showed that the main discrepancies between the devices were for detecting N2 and REM sleep, while N3 had a high agreement. Overall, the DREEM headband seems superior for reliably scoring sleep at home.

Keywords: DREEM, EEG, seep monitoring, Z-machine

Procedia PDF Downloads 99
14000 Optimization of Three Phase Squirrel Cage Induction Motor

Authors: Tunahan Sapmaz, Harun Etçi, İbrahim Şenol, Yasemin Öner

Abstract:

Rotor bar dimensions have a great influence on the air-gap magnetic flux density. Therefore, poor selection of this parameter during the machine design phase causes the air-gap magnetic flux density to be distorted. Thus, it causes noise, torque fluctuation, and losses in the induction motor. On the other hand, the change in rotor bar dimensions will change the resistance of the conductor, so the current will be affected. Therefore, the increase and decrease of rotor bar current affect operation, starting torque, and efficiency. The aim of this study is to examine the effect of rotor bar dimensions on the electromagnetic performance criteria of the induction motor. Modeling of the induction motor is done by the finite element method (FEM), which is a very powerful tool. In FEM, the results generally focus on performance criteria such as torque, torque fluctuation, efficiency, and current.

Keywords: induction motor, finite element method, optimization, rotor bar

Procedia PDF Downloads 118
13999 Increasing a Computer Performance by Overclocking Central Processing Unit (CPU)

Authors: Witthaya Mekhum, Wutthikorn Malikong

Abstract:

The objective of this study is to investigate the increasing desktop computer performance after overclocking central processing unit or CPU by running a computer component at a higher clock rate (more clock cycles per second) than it was designed at the rate of 0.1 GHz for each level or 100 MHz starting at 4000 GHz-4500 GHz. The computer performance is tested for each level with 4 programs, i.e. Hyper PI ver. 0.99b, Cinebench R15, LinX ver.0.6.4 and WinRAR . After the CPU overclock, the computer performance increased. When overclocking CPU at 29% the computer performance tested by Hyper PI ver. 0.99b increased by 10.03% and when tested by Cinebench R15 the performance increased by 20.05% and when tested by LinX Program the performance increased by 16.61%. However, the performance increased only 8.14% when tested with Winrar program. The computer performance did not increase according to the overclock rate because the computer consists of many components such as Random Access Memory or RAM, Hard disk Drive, Motherboard and Display Card, etc.

Keywords: overclock, performance, central processing unit, computer

Procedia PDF Downloads 275
13998 Permanent Magnet Machine Can Be a Vibration Sensor for Itself

Authors: M. Barański

Abstract:

The article presents a new vibration diagnostic method designed to (PM) machines with permanent magnets. Those devices are commonly used in small wind and water systems or vehicles drives. The author’s method is very innovative and unique. Specific structural properties of PM machines are used in this method - electromotive force (EMF) generated due to vibrations. There was analysed number of publications which describe vibration diagnostic methods and tests of electrical PM machines and there was no method found to determine the technical condition of such machine basing on their own signals. In this article, the method genesis, the similarity of machines with permanent magnet to vibration sensor and simulation and laboratory tests results will be discussed. The method of determination the technical condition of electrical machine with permanent magnets basing on its own signals is the subject of patent application No P.405669, and it is the main thesis of author’s doctoral dissertation.

Keywords: vibrations, generator, permanent magnet, traction drive, electrical vehicle

Procedia PDF Downloads 360
13997 Predicting Timely Delivery of Humanitarian Supplies Using Machine Learning Techniques

Authors: Mohammad Alshehri, Fahd Alfarsi

Abstract:

Efficient supply chains play an essential role in delivering humanitarian supplies and directly impact the success of public aid initiatives globally. Predicting the delivery status of these essential supplies in a timely manner is crucial. Therefore, this study explores the application of various machine learning approaches to predict whether humanitarian deliveries will be made on time, using a comprehensive case-study dataset provided by one of the largest international supplying organisations. We employed several machine learning methods such as regression-based (e.g., logistics regression), tree-based (e.g., decision trees) and ensemble-based (e.g., AdaBoost, XGBoost, and Gradient Boosting) to develop our predictive model. Our findings demonstrate that ensemble algorithms achieved promising results, with F1 scores ranging from 0.90 to 0.98. These high accuracy levels indicate the robustness of ensemble-learning techniques in forecasting delivery status, potentially enabling more proactive and efficient supply chain management in global aid initiatives. The implications of this study suggest that integrating advanced predictive analytics can significantly enhance the reliability of supply chains, ensuring the timely delivery of critical commodities to those in need.

Keywords: humanitarian aids, supply chains, machine learning, delivery status

Procedia PDF Downloads 12
13996 A General Framework for Knowledge Discovery Using High Performance Machine Learning Algorithms

Authors: S. Nandagopalan, N. Pradeep

Abstract:

The aim of this paper is to propose a general framework for storing, analyzing, and extracting knowledge from two-dimensional echocardiographic images, color Doppler images, non-medical images, and general data sets. A number of high performance data mining algorithms have been used to carry out this task. Our framework encompasses four layers namely physical storage, object identification, knowledge discovery, user level. Techniques such as active contour model to identify the cardiac chambers, pixel classification to segment the color Doppler echo image, universal model for image retrieval, Bayesian method for classification, parallel algorithms for image segmentation, etc., were employed. Using the feature vector database that have been efficiently constructed, one can perform various data mining tasks like clustering, classification, etc. with efficient algorithms along with image mining given a query image. All these facilities are included in the framework that is supported by state-of-the-art user interface (UI). The algorithms were tested with actual patient data and Coral image database and the results show that their performance is better than the results reported already.

Keywords: active contour, bayesian, echocardiographic image, feature vector

Procedia PDF Downloads 410
13995 A Method for False Alarm Recognition Based on Multi-Classification Support Vector Machine

Authors: Weiwei Cui, Dejian Lin, Leigang Zhang, Yao Wang, Zheng Sun, Lianfeng Li

Abstract:

Built-in test (BIT) is an important technology in testability field, and it is widely used in state monitoring and fault diagnosis. With the improvement of modern equipment performance and complexity, the scope of BIT becomes larger, and it leads to the emergence of false alarm problem. The false alarm makes the health assessment unstable, and it reduces the effectiveness of BIT. The conventional false alarm suppression methods such as repeated test and majority voting cannot meet the requirement for a complicated system, and the intelligence algorithms such as artificial neural networks (ANN) are widely studied and used. However, false alarm has a very low frequency and small sample, yet a method based on ANN requires a large size of training sample. To recognize the false alarm, we propose a method based on multi-classification support vector machine (SVM) in this paper. Firstly, we divide the state of a system into three states: healthy, false-alarm, and faulty. Then we use multi-classification with '1 vs 1' policy to train and recognize the state of a system. Finally, an example of fault injection system is taken to verify the effectiveness of the proposed method by comparing ANN. The result shows that the method is reasonable and effective.

Keywords: false alarm, fault diagnosis, SVM, k-means, BIT

Procedia PDF Downloads 150
13994 A Machine Learning Model for Dynamic Prediction of Chronic Kidney Disease Risk Using Laboratory Data, Non-Laboratory Data, and Metabolic Indices

Authors: Amadou Wurry Jallow, Adama N. S. Bah, Karamo Bah, Shih-Ye Wang, Kuo-Chung Chu, Chien-Yeh Hsu

Abstract:

Chronic kidney disease (CKD) is a major public health challenge with high prevalence, rising incidence, and serious adverse consequences. Developing effective risk prediction models is a cost-effective approach to predicting and preventing complications of chronic kidney disease (CKD). This study aimed to develop an accurate machine learning model that can dynamically identify individuals at risk of CKD using various kinds of diagnostic data, with or without laboratory data, at different follow-up points. Creatinine is a key component used to predict CKD. These models will enable affordable and effective screening for CKD even with incomplete patient data, such as the absence of creatinine testing. This retrospective cohort study included data on 19,429 adults provided by a private research institute and screening laboratory in Taiwan, gathered between 2001 and 2015. Univariate Cox proportional hazard regression analyses were performed to determine the variables with high prognostic values for predicting CKD. We then identified interacting variables and grouped them according to diagnostic data categories. Our models used three types of data gathered at three points in time: non-laboratory, laboratory, and metabolic indices data. Next, we used subgroups of variables within each category to train two machine learning models (Random Forest and XGBoost). Our machine learning models can dynamically discriminate individuals at risk for developing CKD. All the models performed well using all three kinds of data, with or without laboratory data. Using only non-laboratory-based data (such as age, sex, body mass index (BMI), and waist circumference), both models predict chronic kidney disease as accurately as models using laboratory and metabolic indices data. Our machine learning models have demonstrated the use of different categories of diagnostic data for CKD prediction, with or without laboratory data. The machine learning models are simple to use and flexible because they work even with incomplete data and can be applied in any clinical setting, including settings where laboratory data is difficult to obtain.

Keywords: chronic kidney disease, glomerular filtration rate, creatinine, novel metabolic indices, machine learning, risk prediction

Procedia PDF Downloads 96
13993 Development of a Decision-Making Method by Using Machine Learning Algorithms in the Early Stage of School Building Design

Authors: Pegah Eshraghi, Zahra Sadat Zomorodian, Mohammad Tahsildoost

Abstract:

Over the past decade, energy consumption in educational buildings has steadily increased. The purpose of this research is to provide a method to quickly predict the energy consumption of buildings using separate evaluation of zones and decomposing the building to eliminate the complexity of geometry at the early design stage. To produce this framework, machine learning algorithms such as Support vector regression (SVR) and Artificial neural network (ANN) are used to predict energy consumption and thermal comfort metrics in a school as a case. The database consists of more than 55000 samples in three climates of Iran. Cross-validation evaluation and unseen data have been used for validation. In a specific label, cooling energy, it can be said the accuracy of prediction is at least 84% and 89% in SVR and ANN, respectively. The results show that the SVR performed much better than the ANN.

Keywords: early stage of design, energy, thermal comfort, validation, machine learning

Procedia PDF Downloads 76
13992 Value Addition of Quinoa (Chenopodium Quinoa Willd.) Using an Indigenously Developed Saponin Removal Machine

Authors: M.A. Ali, M. Matloob, A. Sahar, M. Yamin, M. Imran, Y.A. Yusof

Abstract:

Quinoa (Chenopodium quinoa Willd.) is known as pseudocereal was originated in South America's Andes. Quinoa is a good source of protein, amino acids, micronutrients and bioactive components. The lack of gluten makes it suitable for celiac patients. Saponins, the leading ant-nutrient, are found in the pericarp, which adheres to the seed and transmits the bitter flavor to the quinoa grain. It is found in varying amounts in quinoa from 0.1% to 5%. This study was planned to design an indigenous machine to remove saponin from quinoa grains at the farm level to promote entrepreneurship. The machine consisted of a feeding hopper, rotating shaft, grooved stone, perforated steel cylinder, V-belts, pulleys, electric motor and mild steel angle iron and sheets. The motor transmitted power to the shaft with a belt drive. The shaft on which the grooved stone was attached rotated inside the perforated cylinder having a clearance of 2 mm and was removed saponin by an abrasion mechanism. The saponin-removed quinoa was then dipped in water to determine the presence of saponin as it produced foam in water and data were statistically analyzed. The results showed that the raw seed feeding rate of 25 g/s and milling time of 135 s completely removed saponin from seeds with minimum grain losses of 2.85% as compared to the economic analysis of the machine showed that its break-even point was achieved after one and half months with 18,000 s and a production capacity of 33 g/s.

Keywords: quinoa seeds, saponin, abrasion mechanism, stone polishing, indigenous machine

Procedia PDF Downloads 64
13991 The Effect of Integrated Reporting on Corporate Financial Performance: A Bibliometric Analysis

Authors: Adhila Sandra Devy, Evangeline Syalomita Silitonga

Abstract:

The landscape of corporate governance and accountability has led to the emergence of Integrated Reporting (IR) in response to the shortcomings of traditional reporting frameworks. Developed by The International Integrated Reporting Council (IIRC), IR aims to offer stakeholders a comprehensive view of a company’s performance by integrating financial and non-financial disclosures. This study analyzes literature on Integrated Reporting and Corporate Financial Performance from 2013 to 2024, employing a descriptive analysis methodology. 31 relevant articles were gathered from various sources, indicating a positive correlation between integrated reporting and financial performance, albeit without conclusive evidence of long-term impact.

Keywords: integrated reporting, corporate financial performance, corporate performance, firm performance, bibliometric analysis

Procedia PDF Downloads 26
13990 Machine Learning Approach for Anomaly Detection in the Simulated Iec-60870-5-104 Traffic

Authors: Stepan Grebeniuk, Ersi Hodo, Henri Ruotsalainen, Paul Tavolato

Abstract:

Substation security plays an important role in the power delivery system. During the past years, there has been an increase in number of attacks on automation networks of the substations. In spite of that, there hasn’t been enough focus dedicated to the protection of such networks. Aiming to design a specialized anomaly detection system based on machine learning, in this paper we will discuss the IEC 60870-5-104 protocol that is used for communication between substation and control station and focus on the simulation of the substation traffic. Firstly, we will simulate the communication between substation slave and server. Secondly, we will compare the system's normal behavior and its behavior under the attack, in order to extract the right features which will be needed for building an anomaly detection system. Lastly, based on the features we will suggest the anomaly detection system for the asynchronous protocol IEC 60870-5-104.

Keywords: Anomaly detection, IEC-60870-5-104, Machine learning, Man-in-the-Middle attacks, Substation security

Procedia PDF Downloads 356
13989 Challenges for Interface Designers in Designing Sensor Dashboards in the Context of Industry 4.0

Authors: Naveen Kumar, Shyambihari Prajapati

Abstract:

Industry 4.0 is the fourth industrial revolution that focuses on interconnectivity of machine to machine, human to machine and human to human via Internet of Things (IoT). Technologies of industry 4.0 facilitate communication between human and machine through IoT and forms Cyber-Physical Production System (CPPS). In CPPS, multiple shop floors sensor data are connected through IoT and displayed through sensor dashboard to the operator. These sensor dashboards have enormous amount of information to be presented which becomes complex for operators to perform monitoring, controlling and interpretation tasks. Designing handheld sensor dashboards for supervision task will become a challenge for the interface designers. This paper reports emerging technologies of industry 4.0, changing context of increasing information complexity in consecutive industrial revolutions and upcoming design challenges for interface designers in context of Industry 4.0. Authors conclude that information complexity of sensor dashboards design has increased with consecutive industrial revolutions and designs of sensor dashboard causes cognitive load on users. Designing such complex dashboards interfaces in Industry 4.0 context will become main challenges for the interface designers.

Keywords: Industry4.0, sensor dashboard design, cyber-physical production system, Interface designer

Procedia PDF Downloads 122
13988 An Approximation Technique to Automate Tron

Authors: P. Jayashree, S. Rajkumar

Abstract:

With the trend of virtual and augmented reality environments booming to provide a life like experience, gaming is a major tool in supporting such learning environments. In this work, a variant of Voronoi heuristics, employing supervised learning for the TRON game is proposed. The paper discusses the features that would be really useful when a machine learning bot is to be used as an opponent against a human player. Various game scenarios, nature of the bot and the experimental results are provided for the proposed variant to prove that the approach is better than those that are currently followed.

Keywords: artificial Intelligence, automation, machine learning, TRON game, Voronoi heuristics

Procedia PDF Downloads 458
13987 Managing Data from One Hundred Thousand Internet of Things Devices Globally for Mining Insights

Authors: Julian Wise

Abstract:

Newcrest Mining is one of the world’s top five gold and rare earth mining organizations by production, reserves and market capitalization in the world. This paper elaborates on the data acquisition processes employed by Newcrest in collaboration with Fortune 500 listed organization, Insight Enterprises, to standardize machine learning solutions which process data from over a hundred thousand distributed Internet of Things (IoT) devices located at mine sites globally. Through the utilization of software architecture cloud technologies and edge computing, the technological developments enable for standardized processes of machine learning applications to influence the strategic optimization of mineral processing. Target objectives of the machine learning optimizations include time savings on mineral processing, production efficiencies, risk identification, and increased production throughput. The data acquired and utilized for predictive modelling is processed through edge computing by resources collectively stored within a data lake. Being involved in the digital transformation has necessitated the standardization software architecture to manage the machine learning models submitted by vendors, to ensure effective automation and continuous improvements to the mineral process models. Operating at scale, the system processes hundreds of gigabytes of data per day from distributed mine sites across the globe, for the purposes of increased improved worker safety, and production efficiency through big data applications.

Keywords: mineral technology, big data, machine learning operations, data lake

Procedia PDF Downloads 103
13986 Firm Performance and Evolving Corporate Governance: An Empirical Study from Pakistan

Authors: Mohammed Nishat, Ahmad Ghazali

Abstract:

This study empirically examines the corporate governance and firm performance, and tries to evaluate the governance, ownership and control related variables which are hypothesized to affect on firms performance. This study tries to evaluate the effectiveness of corporate governance mechanism to achieve high level performance among companies listed on the Karachi Stock Exchange (KSE) over the period from 2005 to 2008. To measure the firm performance level this research uses three measures of performance; Return on assets (ROA), Return on Equity (ROE) and Tobin’s Q. To link the performance of firms with the corporate governance three categories of corporate governance variables are tested which includes governance, ownership and control related variables. Fixed effect regression model is used to test the link between corporate governance and firm performance for 267 KSE listed Pakistani firms. The result shows that corporate governance variables such as percentage block holding by individuals have positive impact on firm performance. When CEO is also the chairperson of board then it is found that firm performance is adversely affected. Also negative relationship is found between share held by insiders and performance of firm. Leverage has negative impact on the performance of the firm and firm size is positively related with the firms performance.

Keywords: corporate governance, performance, agency cost, Karachi stock market

Procedia PDF Downloads 347