Search results for: improvement of model accuracy and reliability
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 23634

Search results for: improvement of model accuracy and reliability

19764 Development of Risk Assessment and Occupational Safety Management Model for Building Construction Projects

Authors: Preeda Sansakorn, Min An

Abstract:

In order to be capable of dealing with uncertainties, subjectivities, including vagueness arising in building construction projects, the application of fuzzy reasoning technique based on fuzzy set theory is proposed. This study contributes significantly to the development of a fuzzy reasoning safety risk assessment model for building construction projects that could be employed to assess the risk magnitude of each hazardous event identified during construction, and a third parameter of probability of consequence is incorporated in the model. By using the proposed safety risk analysis methodology, more reliable and less ambiguities, which provide the safety risk management project team for decision-making purposes.

Keywords: safety risk assessment, building construction safety, fuzzy reasoning, construction risk assessment model, building construction projects

Procedia PDF Downloads 492
19763 The Supply Chain Operation Reference Model Adaptation in the Developing Countries: An Empirical Study on the Egyptian Automotive Sector

Authors: Alaa Osman, Sara Elgazzar, Breksal Elmiligy

Abstract:

The Supply Chain Operation Reference (SCOR) model is considered one of the most widely implemented supply chain performance measurement systems (SCPMSs). Several studies have been proposed on the SCOR model adaptation in developed countries context; while there is a limited availability of previous work on the SCPMSs application generally and the SCOR model specifically in developing nations. This paper presents a research agenda on the SCOR model adaptation in the developing countries. It aims at investigating the challenges of adapting the SCOR model to manage and measure supply chain performance in developing countries. The research will exemplify the system in the Egyptian automotive sector to gain a comprehensive understanding of how the application of the SCOR model can affect the performance of automotive companies in Egypt, with a necessary understanding of challenges and obstacles faced the adaptation of the model in the Egyptian supply chain context. An empirical study was conducted on the Egyptian automotive sector in three companies considering three different classes: BMW, Hyundai and Brilliance. First, in-depth interviews were carried out to gain an insight into the implementation and the relevance of the concepts of supply chain management and performance measurement in the Egyptian automotive industry. Then, a formal survey was designed based on the SCOR model five main processes (plan, source, make, deliver and return) and best practices to investigate the challenges and obstacles faced the adaptation of the SCOR model in the Egyptian automotive supply chain. Finally, based on the survey results, the appropriate best practices for each process were identified in order to overcome the SCOR model adaptation challenges. The results showed that the implementation of the SCOR model faced different challenges and unavailability of the required enablers. The survey highlighted the low integration of end-to-end supply chain, lacks commitment for the innovative ideas and technologies, financial constraints and lack of practical training and support as the main challenges faced the adaptation of the SCOR model in the Egyptian automotive supply chain. The research provides an original contribution to knowledge by proposing a procedure to identify challenges encountered during the process of SCOR model adoption which can pave a way for further research in the area of SCPMSs adaptation, particularly in the developing countries. The research can help managers and organizations to identify obstacles and difficulties of the SCOR model adaptation, subsequently this can facilitate measuring the improved performance or changes in the organizational performance.

Keywords: automotive sector, developing countries, SCOR model, supply chain performance

Procedia PDF Downloads 374
19762 Kou Jump Diffusion Model: An Application to the SP 500; Nasdaq 100 and Russell 2000 Index Options

Authors: Wajih Abbassi, Zouhaier Ben Khelifa

Abstract:

The present research points towards the empirical validation of three options valuation models, the ad-hoc Black-Scholes model as proposed by Berkowitz (2001), the constant elasticity of variance model of Cox and Ross (1976) and the Kou jump-diffusion model (2002). Our empirical analysis has been conducted on a sample of 26,974 options written on three indexes, the S&P 500, Nasdaq 100 and the Russell 2000 that were negotiated during the year 2007 just before the sub-prime crisis. We start by presenting the theoretical foundations of the models of interest. Then we use the technique of trust-region-reflective algorithm to estimate the structural parameters of these models from cross-section of option prices. The empirical analysis shows the superiority of the Kou jump-diffusion model. This superiority arises from the ability of this model to portray the behavior of market participants and to be closest to the true distribution that characterizes the evolution of these indices. Indeed the double-exponential distribution covers three interesting properties that are: the leptokurtic feature, the memory less property and the psychological aspect of market participants. Numerous empirical studies have shown that markets tend to have both overreaction and under reaction over good and bad news respectively. Despite of these advantages there are not many empirical studies based on this model partly because probability distribution and option valuation formula are rather complicated. This paper is the first to have used the technique of nonlinear curve-fitting through the trust-region-reflective algorithm and cross-section options to estimate the structural parameters of the Kou jump-diffusion model.

Keywords: jump-diffusion process, Kou model, Leptokurtic feature, trust-region-reflective algorithm, US index options

Procedia PDF Downloads 429
19761 Integrating Machine Learning and Rule-Based Decision Models for Enhanced B2B Sales Forecasting and Customer Prioritization

Authors: Wenqi Liu, Reginald Bailey

Abstract:

This study proposes a comprehensive and effective approach to business-to-business (B2B) sales forecasting by integrating advanced machine learning models with a rule-based decision-making framework. The methodology addresses the critical challenge of optimizing sales pipeline performance and improving conversion rates through predictive analytics and actionable insights. The first component involves developing a classification model to predict the likelihood of conversion, aiming to outperform traditional methods such as logistic regression in terms of accuracy, precision, recall, and F1 score. Feature importance analysis highlights key predictive factors, such as client revenue size and sales velocity, providing valuable insights into conversion dynamics. The second component focuses on forecasting sales value using a regression model, designed to achieve superior performance compared to linear regression by minimizing mean absolute error (MAE), mean squared error (MSE), and maximizing R-squared metrics. The regression analysis identifies primary drivers of sales value, further informing data-driven strategies. To bridge the gap between predictive modeling and actionable outcomes, a rule-based decision framework is introduced. This model categorizes leads into high, medium, and low priorities based on thresholds for conversion probability and predicted sales value. By combining classification and regression outputs, this framework enables sales teams to allocate resources effectively, focus on high-value opportunities, and streamline lead management processes. The integrated approach significantly enhances lead prioritization, increases conversion rates, and drives revenue generation, offering a robust solution to the declining pipeline conversion rates faced by many B2B organizations. Our findings demonstrate the practical benefits of blending machine learning with decision-making frameworks, providing a scalable, data-driven solution for strategic sales optimization. This study underscores the potential of predictive analytics to transform B2B sales operations, enabling more informed decision-making and improved organizational outcomes in competitive markets.

Keywords: machine learning, XGBoost, regression, decision making framework, system engineering

Procedia PDF Downloads 17
19760 Performance Assessment of Multi-Level Ensemble for Multi-Class Problems

Authors: Rodolfo Lorbieski, Silvia Modesto Nassar

Abstract:

Many supervised machine learning tasks require decision making across numerous different classes. Multi-class classification has several applications, such as face recognition, text recognition and medical diagnostics. The objective of this article is to analyze an adapted method of Stacking in multi-class problems, which combines ensembles within the ensemble itself. For this purpose, a training similar to Stacking was used, but with three levels, where the final decision-maker (level 2) performs its training by combining outputs from the tree-based pair of meta-classifiers (level 1) from Bayesian families. These are in turn trained by pairs of base classifiers (level 0) of the same family. This strategy seeks to promote diversity among the ensembles forming the meta-classifier level 2. Three performance measures were used: (1) accuracy, (2) area under the ROC curve, and (3) time for three factors: (a) datasets, (b) experiments and (c) levels. To compare the factors, ANOVA three-way test was executed for each performance measure, considering 5 datasets by 25 experiments by 3 levels. A triple interaction between factors was observed only in time. The accuracy and area under the ROC curve presented similar results, showing a double interaction between level and experiment, as well as for the dataset factor. It was concluded that level 2 had an average performance above the other levels and that the proposed method is especially efficient for multi-class problems when compared to binary problems.

Keywords: stacking, multi-layers, ensemble, multi-class

Procedia PDF Downloads 269
19759 Radiative Reactions Analysis at the Range of Astrophysical Energies

Authors: A. Amar

Abstract:

Analysis of the elastic scattering of protons on 10B nuclei has been done in the framework of the optical model and single folding model at the beam energies up to 17 MeV. We could enhance the optical potential parameters using Esis88 Code, as well as SPI GENOA Code. Linear relationship between volume real potential (V0) and proton energy (Ep) has been obtained. Also, surface imaginary potential WD is proportional to the proton energy (Ep) in the range 0.400 and 17 MeV. The radiative reaction 10B(p,γ)11C has been analyzed using potential model. A comparison between 10B(p,γ)11C and 6Li(p,γ)7Be has been made. Good agreement has been found between theoretical and experimental results in the whole range of energy. The radiative resonance reaction 7Li(p,γ)8Be has been studied.

Keywords: elastic scattering of protons on 10B nuclei, optical potential parameters, potential model, radiative reaction

Procedia PDF Downloads 211
19758 Statistical Scientific Investigation of Popular Cultural Heritage in the Relationship between Astronomy and Weather Conditions in the State of Kuwait

Authors: Ahmed M. AlHasem

Abstract:

The Kuwaiti society has long been aware of climatic changes and their annual dates and trying to link them to astronomy in an attempt to forecast the future weather conditions. The reason for this concern is that many of the economic, social and living activities of the society depend deeply on the nature of the weather conditions directly and indirectly. In other words, Kuwaiti society, like the case of many human societies, has in the past tried to predict climatic conditions by linking them to astronomy or popular statements to indicate the timing of climate changes. Accordingly, this study was devoted to scientific investigation based on the statistical analysis of climatic data to show the accuracy and compatibility of some of the most important elements of the cultural heritage in relation to climate change and to relate it scientifically to precise climatic measurements for decades. The research has been divided into 10 topics, each topic has been focused on one legacy, whether by linking climate changes to the appearance/disappearance of star or a popular statement inherited through generations, through explain the nature and timing and thereby statistical analysis to indicate the proportion of accuracy based on official climatic data since 1962. The study's conclusion is that the relationship is weak and, in some cases, non-existent between the popular heritage and the actual climatic data. Therefore, it does not have a dependable relationship and a reliable scientific prediction between both the popular heritage and the forecast of weather conditions.

Keywords: astronomy, cultural heritage, statistical analysis, weather prediction

Procedia PDF Downloads 123
19757 Gaussian Probability Density for Forest Fire Detection Using Satellite Imagery

Authors: S. Benkraouda, Z. Djelloul-Khedda, B. Yagoubi

Abstract:

we present a method for early detection of forest fires from a thermal infrared satellite image, using the image matrix of the probability of belonging. The principle of the method is to compare a theoretical mathematical model to an experimental model. We considered that each line of the image matrix, as an embodiment of a non-stationary random process. Since the distribution of pixels in the satellite image is statistically dependent, we divided these lines into small stationary and ergodic intervals to characterize the image by an adequate mathematical model. A standard deviation was chosen to generate random variables, so each interval behaves naturally like white Gaussian noise. The latter has been selected as the mathematical model that represents a set of very majority pixels, which we can be considered as the image background. Before modeling the image, we made a few pretreatments, then the parameters of the theoretical Gaussian model were extracted from the modeled image, these settings will be used to calculate the probability of each interval of the modeled image to belong to the theoretical Gaussian model. The high intensities pixels are regarded as foreign elements to it, so they will have a low probability, and the pixels that belong to the background image will have a high probability. Finally, we did present the reverse of the matrix of probabilities of these intervals for a better fire detection.

Keywords: forest fire, forest fire detection, satellite image, normal distribution, theoretical gaussian model, thermal infrared matrix image

Procedia PDF Downloads 142
19756 Harmonic Analysis to Improve Power Quality

Authors: Rumana Ali

Abstract:

The presence of nonlinear and power electronic switching devices produce distorted output and harmonics into the system. This paper presents a technique to analyze harmonics using digital series oscilloscope (DSO). In power distribution system further measurements are done by DSO, and the waveforms are analyzed using FFT program. The results of this proposed work are helpful for the investigator to install an appropriate compensating device to mitigate the harmonics, in turn, improve the power quality. This case study is carried out at AIT Chikmagalur. It is done as a starting step towards the improvement of energy efficiency at AIT Chikmagalur, and with an overall aim of reducing the electricity bill with a complete energy audit of the institution. Strategies were put forth to reach the above objective: The following strategies were proposed to be implemented to analyze the power quality in EEE department of the institution. Strategy 1: The power factor has to be measured using the energy meter. Power factor improvement may reduce the voltage drop in lines. This brings the voltages at the socket in the labs closer to the nominal voltage of 230V, and thus power quality improves. Strategy 2: The harmonics at the power inlet has to be measured by means of a DSO. The DSO waveform is analyzed using FFT to know the percentage harmonic up to the 13th harmonics of 50Hz. Reduction in the harmonics in the inlet of the EEE department may reduce line losses and therefore reduces energy bill to the institution.

Keywords: harmonic analysis, energy bill, power quality, electronic switching devices

Procedia PDF Downloads 309
19755 Numerical Approach for Characterization of Flow Field in Pump Intake Using Two Phase Model: Detached Eddy Simulation

Authors: Rahul Paliwal, Gulshan Maheshwari, Anant S. Jhaveri, Channamallikarjun S. Mathpati

Abstract:

Large pumping facility is the necessary requirement of the cooling water systems for power plants, process and manufacturing facilities, flood control and water or waste water treatment plant. With a large capacity of few hundred to 50,000 m3/hr, cares must be taken to ensure the uniform flow to the pump to limit vibration, flow induced cavitation and performance problems due to formation of air entrained vortex and swirl flow. Successful prediction of these phenomena requires numerical method and turbulence model to characterize the dynamics of these flows. In the past years, single phase shear stress transport (SST) Reynolds averaged Navier Stokes Models (like k-ε, k-ω and RSM) were used to predict the behavior of flow. Literature study showed that two phase model will be more accurate over single phase model. In this paper, a 3D geometries simulated using detached eddy simulation (LES) is used to predict the behavior of the fluid and the results are compared with experimental results. Effect of different grid structure and boundary condition is also studied. It is observed that two phase flow model can more accurately predict the mean flow and turbulence statistics compared to the steady SST model. These validate model will be used for further analysis of vortex structure in lab scale model to generate their frequency-plot and intensity at different location in the set-up. This study will help in minimizing the ill effect of vortex on pump performance.

Keywords: grid structure, pump intake, simulation, vibration, vortex

Procedia PDF Downloads 175
19754 Satellite Imagery Classification Based on Deep Convolution Network

Authors: Zhong Ma, Zhuping Wang, Congxin Liu, Xiangzeng Liu

Abstract:

Satellite imagery classification is a challenging problem with many practical applications. In this paper, we designed a deep convolution neural network (DCNN) to classify the satellite imagery. The contributions of this paper are twofold — First, to cope with the large-scale variance in the satellite image, we introduced the inception module, which has multiple filters with different size at the same level, as the building block to build our DCNN model. Second, we proposed a genetic algorithm based method to efficiently search the best hyper-parameters of the DCNN in a large search space. The proposed method is evaluated on the benchmark database. The results of the proposed hyper-parameters search method show it will guide the search towards better regions of the parameter space. Based on the found hyper-parameters, we built our DCNN models, and evaluated its performance on satellite imagery classification, the results show the classification accuracy of proposed models outperform the state of the art method.

Keywords: satellite imagery classification, deep convolution network, genetic algorithm, hyper-parameter optimization

Procedia PDF Downloads 301
19753 Numerical Simulation of the Production of Ceramic Pigments Using Microwave Radiation: An Energy Efficiency Study Towards the Decarbonization of the Pigment Sector

Authors: Pedro A. V. Ramos, Duarte M. S. Albuquerque, José C. F. Pereira

Abstract:

Global warming mitigation is one of the main challenges of this century, having the net balance of greenhouse gas (GHG) emissions to be null or negative in 2050. Industry electrification is one of the main paths to achieving carbon neutrality within the goals of the Paris Agreement. Microwave heating is becoming a popular industrial heating mechanism due to the absence of direct GHG emissions, but also the rapid, volumetric, and efficient heating. In the present study, a mathematical model is used to simulate the production using microwave heating of two ceramic pigments, at high temperatures (above 1200 Celsius degrees). The two pigments studied were the yellow (Pr, Zr)SiO₂ and the brown (Ti, Sb, Cr)O₂. The chemical conversion of reactants into products was included in the model by using the kinetic triplet obtained with the model-fitting method and experimental data present in the Literature. The coupling between the electromagnetic, thermal, and chemical interfaces was also included. The simulations were computed in COMSOL Multiphysics. The geometry includes a moving plunger to allow for the cavity impedance matching and thus maximize the electromagnetic efficiency. To accomplish this goal, a MATLAB controller was developed to automatically search the position of the moving plunger that guarantees the maximum efficiency. The power is automatically and permanently adjusted during the transient simulation to impose stationary regime and total conversion, the two requisites of every converged solution. Both 2D and 3D geometries were used and a parametric study regarding the axial bed velocity and the heat transfer coefficient at the boundaries was performed. Moreover, a Verification and Validation study was carried out by comparing the conversion profiles obtained numerically with the experimental data available in the Literature; the numerical uncertainty was also estimated to attest to the result's reliability. The results show that the model-fitting method employed in this work is a suitable tool to predict the chemical conversion of reactants into the pigment, showing excellent agreement between the numerical results and the experimental data. Moreover, it was demonstrated that higher velocities lead to higher thermal efficiencies and thus lower energy consumption during the process. This work concludes that the electromagnetic heating of materials having high loss tangent and low thermal conductivity, like ceramic materials, maybe a challenge due to the presence of hot spots, which may jeopardize the product quality or even the experimental apparatus. The MATLAB controller increased the electromagnetic efficiency by 25% and global efficiency of 54% was obtained for the titanate brown pigment. This work shows that electromagnetic heating will be a key technology in the decarbonization of the ceramic sector as reductions up to 98% in the specific GHG emissions were obtained when compared to the conventional process. Furthermore, numerical simulations appear as a suitable technique to be used in the design and optimization of microwave applicators, showing high agreement with experimental data.

Keywords: automatic impedance matching, ceramic pigments, efficiency maximization, high-temperature microwave heating, input power control, numerical simulation

Procedia PDF Downloads 138
19752 Quantification of Leachate Potential of the Quezon City Controlled Dumping Facility Using Help Model

Authors: Paul Kenneth D. Luzon, Maria Antonia N. Tanchuling

Abstract:

The Quezon City Controlled Dumping facility also known as Payatas produces leachate which can contaminate soil and water environment in the area. The goal of this study is to quantify the leachate produced by the QCCDF using the Hydrologic Evaluation of Landfill Performance (HELP) model. Results could be used as input for groundwater contaminant transport studies. The HELP model is based on a simple water budget and is an essential “model requirement” used by the US Environmental Protection Agency (EPA). Annual waste profile of the QCCDF was calculated. Based on topographical maps and estimation of settlement due to overburden pressure and degradation, a total of 10M m^3 of waste is contained in the landfill. The input necessary for the HELP model are weather data, soil properties, and landfill design. Results showed that from 1988 to 2011, an average of 50% of the total precipitation percolates through the bottom layer. Validation of the results is still needed due to the assumptions made in the study. The decrease in porosity of the top soil cover showed the best mitigation for minimizing percolation rate. This study concludes that there is a need for better leachate management system in the QCCDF.

Keywords: help model, landfill, payatas trash slide, quezon city controlled dumping facility

Procedia PDF Downloads 291
19751 An Experimental Machine Learning Analysis on Adaptive Thermal Comfort and Energy Management in Hospitals

Authors: Ibrahim Khan, Waqas Khalid

Abstract:

The Healthcare sector is known to consume a higher proportion of total energy consumption in the HVAC market owing to an excessive cooling and heating requirement in maintaining human thermal comfort in indoor conditions, catering to patients undergoing treatment in hospital wards, rooms, and intensive care units. The indoor thermal comfort conditions in selected hospitals of Islamabad, Pakistan, were measured on a real-time basis with the collection of first-hand experimental data using calibrated sensors measuring Ambient Temperature, Wet Bulb Globe Temperature, Relative Humidity, Air Velocity, Light Intensity and CO2 levels. The Experimental data recorded was analyzed in conjunction with the Thermal Comfort Questionnaire Surveys, where the participants, including patients, doctors, nurses, and hospital staff, were assessed based on their thermal sensation, acceptability, preference, and comfort responses. The Recorded Dataset, including experimental and survey-based responses, was further analyzed in the development of a correlation between operative temperature, operative relative humidity, and other measured operative parameters with the predicted mean vote and adaptive predicted mean vote, with the adaptive temperature and adaptive relative humidity estimated using the seasonal data set gathered for both summer – hot and dry, and hot and humid as well as winter – cold and dry, and cold and humid climate conditions. The Machine Learning Logistic Regression Algorithm was incorporated to train the operative experimental data parameters and develop a correlation between patient sensations and the thermal environmental parameters for which a new ML-based adaptive thermal comfort model was proposed and developed in our study. Finally, the accuracy of our model was determined using the K-fold cross-validation.

Keywords: predicted mean vote, thermal comfort, energy management, logistic regression, machine learning

Procedia PDF Downloads 63
19750 Effect of Self-Questioning Strategy on the Improvement of Reading Comprehension of ESL Learners

Authors: Muhammad Hamza

Abstract:

This research is based on the effect of self-questioning strategy on reading comprehension of second language learners at medium level. This research is conducted to find out the effects of self-questioning strategy and how self-questioning strategy helps English learners to improve their reading comprehension. In this research study the researcher has analyzed that how much self-questioning is effective in the field of learning second language and how much it helps second language learners to improve their reading comprehension. For this purpose, the researcher has studied different reading strategies, analyzed, collected data from certificate level class at NUML, Peshawar campus and then found out the effects of self-questioning strategy on reading comprehension of ESL learners. The researcher has randomly selected the participants from certificate class. The data was analyzed through pre-test and post-test and then in the final stage the results of both tests were compared. After the pre-test and post-test, the result of both pre-test and post-test indicated that if the learners start to use self-questioning strategy before reading a text, while reading a text and after reading a particular text there’ll be improvement in comprehension level of ESL learners. The present research has addressed the benefits of self-questioning strategy by taking two tests (pre and post-test).After the result of post-test it is revealed that the use of the self-questioning strategy has a significant effect on the readers’ comprehension thus, they can improve their reading comprehension by using self-questioning strategy.

Keywords: strategy, self-questioning, comprehension, intermediate level ESL learner

Procedia PDF Downloads 66
19749 Investigation of the Speckle Pattern Effect for Displacement Assessments by Digital Image Correlation

Authors: Salim Çalışkan, Hakan Akyüz

Abstract:

Digital image correlation has been accustomed as a versatile and efficient method for measuring displacements on the article surfaces by comparing reference subsets in undeformed images with the define target subset in the distorted image. The theoretical model points out that the accuracy of the digital image correlation displacement data can be exactly anticipated based on the divergence of the image noise and the sum of the squares of the subset intensity gradients. The digital image correlation procedure locates each subset of the original image in the distorted image. The software then determines the displacement values of the centers of the subassemblies, providing the complete displacement measures. In this paper, the effect of the speckle distribution and its effect on displacements measured out plane displacement data as a function of the size of the subset was investigated. Nine groups of speckle patterns were used in this study: samples are sprayed randomly by pre-manufactured patterns of three different hole diameters, each with three coverage ratios, on a computer numerical control punch press. The resulting displacement values, referenced at the center of the subset, are evaluated based on the average of the displacements of the pixel’s interior the subset.

Keywords: digital image correlation, speckle pattern, experimental mechanics, tensile test, aluminum alloy

Procedia PDF Downloads 74
19748 ANOVA-Based Feature Selection and Machine Learning System for IoT Anomaly Detection

Authors: Muhammad Ali

Abstract:

Cyber-attacks and anomaly detection on the Internet of Things (IoT) infrastructure is emerging concern in the domain of data-driven intrusion. Rapidly increasing IoT risk is now making headlines around the world. denial of service, malicious control, data type probing, malicious operation, DDos, scan, spying, and wrong setup are attacks and anomalies that can affect an IoT system failure. Everyone talks about cyber security, connectivity, smart devices, and real-time data extraction. IoT devices expose a wide variety of new cyber security attack vectors in network traffic. For further than IoT development, and mainly for smart and IoT applications, there is a necessity for intelligent processing and analysis of data. So, our approach is too secure. We train several machine learning models that have been compared to accurately predicting attacks and anomalies on IoT systems, considering IoT applications, with ANOVA-based feature selection with fewer prediction models to evaluate network traffic to help prevent IoT devices. The machine learning (ML) algorithms that have been used here are KNN, SVM, NB, D.T., and R.F., with the most satisfactory test accuracy with fast detection. The evaluation of ML metrics includes precision, recall, F1 score, FPR, NPV, G.M., MCC, and AUC & ROC. The Random Forest algorithm achieved the best results with less prediction time, with an accuracy of 99.98%.

Keywords: machine learning, analysis of variance, Internet of Thing, network security, intrusion detection

Procedia PDF Downloads 125
19747 A Practical Approach and Implementation of Digital Library Towards Best Practice in Malaysian Academic Library

Authors: Zainab Ajab Mohideen, Kiran Kaur, A. Basheer Ahamadhu, Noor Azlinda Wan Jan, Sukmawati Muhammad

Abstract:

The corpus in the digital library is to provide an overview and evidence from library automation that can be used to justify the needs of the digital library. This paper disperses the approach and implementation of the digital library as part of best practices by the Automation Division at Hamzah Sendut Library of the University Science Malaysia (USM). The implemented digital library model emphasizes on the entire library collections, technical perspective, and automation solution. This model served as a foundation for digital library services as part of information delivery in the USM digital library. The approach to digital library includes discussion on key factors, design, architecture, and pragmatic model that has been collected, captured, and identified during the implementation stages. At present, the USM digital library has achieved the status of an Institutional Repository (IR).

Keywords: academic digital library, digital information system, digital library best practice, digital library model

Procedia PDF Downloads 555
19746 Development of a Value Evaluation Model of Highway Box-Girder Bridge

Authors: Hao Hsi Tseng

Abstract:

Taiwan’s infrastructure is gradually deteriorating, while resources for maintenance and replacement are increasingly limited, raising the urgent need for methods for maintaining existing infrastructure within constrained budgets. Infrastructure value evaluation is used to enhance the efficiency of infrastructure maintenance work, allowing administrators to quickly assess the maintenance needs and performance by observing variation in infrastructure value. This research establishes a value evaluation model for Taiwan’s highway box girder bridges. The operating mechanism and process of the model are illustrated in a practical case.

Keywords: box girder bridge, deterioration, infrastructure, maintenance, value evaluation

Procedia PDF Downloads 190
19745 Developing an Information Model of Manufacturing Process for Sustainability

Authors: Jae Hyun Lee

Abstract:

Manufacturing companies use life-cycle inventory databases to analyze sustainability of their manufacturing processes. Life cycle inventory data provides reference data which may not be accurate for a specific company. Collecting accurate data of manufacturing processes for a specific company requires enormous time and efforts. An information model of typical manufacturing processes can reduce time and efforts to get appropriate reference data for a specific company. This paper shows an attempt to build an abstract information model which can be used to develop information models for specific manufacturing processes.

Keywords: process information model, sustainability, OWL, manufacturing

Procedia PDF Downloads 430
19744 Experimental and Numerical Investigation of Hardness and Compressive Strength of Hybrid Glass/Steel Fiber Reinforced Polymer Composites

Authors: Amar Patnaik, Pankaj Agarwal

Abstract:

This paper investigates the experimental study of hardness and compressive strength of hybrid glass/steel fiber reinforced polymer composites by varying the glass and steel fiber layer in the epoxy matrix. The hybrid composites with four stacking sequences HSG-1, HSG-2, HSG-3, and HSG-4 were fabricated by the VARTM process under the controlled environment. The experimentally evaluated results of Vicker’s hardness of the fabricated composites increases with an increase in the fiber layers sequence showing the high resistance. The improvement of micro-structure ability has been observed from the SEM study, which governs in the enhancement of compressive strength. The finite element model was developed on ANSYS to predict the above said properties and further compared with experimental results. The results predicted by the numerical simulation are in good agreement with the experimental results. The hybrid composites developed in this study was identified as the preferred materials due to their excellent mechanical properties to replace the conventional materialsused in the marine structures.

Keywords: finite element method, interfacial strength, polymer composites, VARTM

Procedia PDF Downloads 132
19743 Automatic Slider Design in Injection Moldings

Authors: Alan C. Lin, Tran Anh Son

Abstract:

This study proposes an approach to determine the undercut regions and their releasing directions for slider design of complex parts represented by the file format of STL (STereoLithography). In order to delineate the border of undercut regions, orthogonal cutting planes are firstly employed to automatically find the inner loops of a part model. To discover the facets belonging to undercut regions, attributes are then assigned to the facets of the part model based on the topological relationship of adjacent facets of each inner loop. After that, the undercut regions are separated from other facets in the model. Through the recognized facets of the undercut regions, the concept of 'visibility map (V-map)' is further applied to determine feasible releasing directions for each of the undercut regions. The undercut regions having the same releasing direction are finally grouped to form a slider in the injection mold.

Keywords: solid model, STL data, injection mold design, visibility map

Procedia PDF Downloads 395
19742 Autism Disease Detection Using Transfer Learning Techniques: Performance Comparison between Central Processing Unit vs. Graphics Processing Unit Functions for Neural Networks

Authors: Mst Shapna Akter, Hossain Shahriar

Abstract:

Neural network approaches are machine learning methods used in many domains, such as healthcare and cyber security. Neural networks are mostly known for dealing with image datasets. While training with the images, several fundamental mathematical operations are carried out in the Neural Network. The operation includes a number of algebraic and mathematical functions, including derivative, convolution, and matrix inversion and transposition. Such operations require higher processing power than is typically needed for computer usage. Central Processing Unit (CPU) is not appropriate for a large image size of the dataset as it is built with serial processing. While Graphics Processing Unit (GPU) has parallel processing capabilities and, therefore, has higher speed. This paper uses advanced Neural Network techniques such as VGG16, Resnet50, Densenet, Inceptionv3, Xception, Mobilenet, XGBOOST-VGG16, and our proposed models to compare CPU and GPU resources. A system for classifying autism disease using face images of an autistic and non-autistic child was used to compare performance during testing. We used evaluation matrices such as Accuracy, F1 score, Precision, Recall, and Execution time. It has been observed that GPU runs faster than the CPU in all tests performed. Moreover, the performance of the Neural Network models in terms of accuracy increases on GPU compared to CPU.

Keywords: autism disease, neural network, CPU, GPU, transfer learning

Procedia PDF Downloads 118
19741 Queuing Analysis and Optimization of Public Vehicle Transport Stations: A Case of South West Ethiopia Region Vehicle Stations

Authors: Mequanint Birhan

Abstract:

Modern urban environments present a dynamically growing field where, notwithstanding shared goals, several mutually conflicting interests frequently collide. However, it has a big impact on the city's socioeconomic standing, waiting lines and queues are common occurrences. This results in extremely long lines for both vehicles and people on incongruous routes, service coagulation, customer murmuring, unhappiness, complaints, and looking for other options sometimes illegally. The root cause of this is corruption, which leads to traffic jams, stopping, and packing vehicles beyond their safe carrying capacity, and violating the human rights and freedoms of passengers. This study focused on the optimizing time of passengers had to wait in public vehicle stations. This applied research employed both data gathering sources and mixed approaches, then 166 samples of key informants of transport station were taken by using the Slovin sampling formula. The length of time vehicles, including the drivers and auxiliary drivers ‘Weyala', had to wait was also studied. To maximize the service level at vehicle stations, a queuing model was subsequently devised ‘Menaharya’. Time, cost, and quality encompass performance, scope, and suitability for the intended purposes. The minimal response time for passengers and vehicles queuing to reach their final destination at the stations of the Tepi, Mizan, and Bonga towns was determined. A new bus station system was modeled and simulated by Arena simulation software in the chosen study area. 84% improvement on cost reduced by 56.25%, time 4hr to 1.5hr, quality, safety and designed load performance calculations employed. Stakeholders are asked to put the model into practice and monitor the results obtained.

Keywords: Arena 14 automatic rockwell, queue, transport services, vehicle stations

Procedia PDF Downloads 78
19740 Numerical Simulation of Kangimi Reservoir Sedimentation, Kaduna State, Nigeria

Authors: Abdurrasheed Sa'id, Abubakar Isma'il, Waheed Alayande

Abstract:

This study focused on carrying out numerical simulations of Kangimi reservoir sedimentation by reviewing different numerical sediment transport models, and GSTARS3 was selected. The model was developed using the 1977 data. It was calibrated by simulating the 2012 profile and sediment deposition and compared with 2012 hydrographic survey results of NWRI. The model was validated by simulating the 2016 deposition and compared the results with NWRI estimates. Also, the performance of the proposed model was tested using statistical parameters such as MSE (Mean Square Error), MAPE (Mean Average Percentage Error) and R2 (Coefficient of determination) with values of 1.32m, 0.17% and 0.914 respectively which shows strong agreement. After the calibration, validation and performance testing the model was used to simulate the 2032 and 2062 profiles and deposition. The results showed that by 2032 the reservoir will be silted by 25.34MCM or 43.3% of the design capacity and 60.7% of the capacity by the year 2062. A number of sedimentation mitigation measures were recommended.

Keywords: NWRI- national water resources institute, sedimentation, GSTARS3, model

Procedia PDF Downloads 220
19739 Use of Multistage Transition Regression Models for Credit Card Income Prediction

Authors: Denys Osipenko, Jonathan Crook

Abstract:

Because of the variety of the card holders’ behaviour types and income sources each consumer account can be transferred to a variety of states. Each consumer account can be inactive, transactor, revolver, delinquent, defaulted and requires an individual model for the income prediction. The estimation of transition probabilities between statuses at the account level helps to avoid the memorylessness of the Markov Chains approach. This paper investigates the transition probabilities estimation approaches to credit cards income prediction at the account level. The key question of empirical research is which approach gives more accurate results: multinomial logistic regression or multistage conditional logistic regression with binary target. Both models have shown moderate predictive power. Prediction accuracy for conditional logistic regression depends on the order of stages for the conditional binary logistic regression. On the other hand, multinomial logistic regression is easier for usage and gives integrate estimations for all states without priorities. Thus further investigations can be concentrated on alternative modeling approaches such as discrete choice models.

Keywords: multinomial regression, conditional logistic regression, credit account state, transition probability

Procedia PDF Downloads 487
19738 A Palmprint Identification System Based Multi-Layer Perceptron

Authors: David P. Tantua, Abdulkader Helwan

Abstract:

Biometrics has been recently used for the human identification systems using the biological traits such as the fingerprints and iris scanning. Identification systems based biometrics show great efficiency and accuracy in such human identification applications. However, these types of systems are so far based on some image processing techniques only, which may decrease the efficiency of such applications. Thus, this paper aims to develop a human palmprint identification system using multi-layer perceptron neural network which has the capability to learn using a backpropagation learning algorithms. The developed system uses images obtained from a public database available on the internet (CASIA). The processing system is as follows: image filtering using median filter, image adjustment, image skeletonizing, edge detection using canny operator to extract features, clear unwanted components of the image. The second phase is to feed those processed images into a neural network classifier which will adaptively learn and create a class for each different image. 100 different images are used for training the system. Since this is an identification system, it should be tested with the same images. Therefore, the same 100 images are used for testing it, and any image out of the training set should be unrecognized. The experimental results shows that this developed system has a great accuracy 100% and it can be implemented in real life applications.

Keywords: biometrics, biological traits, multi-layer perceptron neural network, image skeletonizing, edge detection using canny operator

Procedia PDF Downloads 371
19737 Autopsy-Based Study of Abdominal Traffic Trauma Death after Emergency Room Arrival

Authors: Satoshi Furukawa, Satomu Morita, Katsuji Nishi, Masahito Hitosugi

Abstract:

We experience the autopsy cases that the deceased was alive in emergency room on arrival. Bleeding is the leading cause of preventable death after injury. This retrospective study aimed to characterize opportunities for performance improvement identified in patients who died from traffic trauma and were considered by the quality improvement of education system. The Japan Advanced Trauma Evaluation and Care (JATEC) education program was introduced in 2002. We focused the abdominal traffic trauma injury. An autopsy-based cross-sectional study conducted. A purposive sampling technique was applied to select the study sample of 41 post-mortems of road traffic accident between April 1999 and March 2014 subjected to medico-legal autopsy at the department of Forensic Medicine, Shiga University of Medical Science. 16 patients (39.0%) were abdominal trauma injury. The mean period of survival after meet with accident was 13.5 hours, compared abdominal trauma death was 27.4 hours longer. In road traffic accidents, the most injured abdominal organs were liver followed by mesentery. We thought delayed treatment was associated with immediate diagnostic imaging, and so expected to expand trauma management examination.

Keywords: abdominal traffic trauma, preventable death, autopsy, emergency medicine

Procedia PDF Downloads 453
19736 Microencapsulated Boswellia serrata and Probiotic Bacteria Acted as Symbiotic in Metabolic Syndrome Rat Model

Authors: Moetazza M. Alshafei, Ahmed M. Mabrouk, Emtenan M. Hanafi, Manal M. Ramadan, Reda M. S. Korany, Seham S. Kassem, Dina Mostafa Mohammed

Abstract:

Metabolic syndrome (MeS) is a major health problem with a high incidence of obese individuals worldwide. Increased related morbidity of diabetes, hypertension and fatty liver disease, and complicated cardiovascular disease are inevitable. Boswellia serrata gum (Bos) is a promising traditional medicinal plant; it has several pharmacological properties, including anti-inflammatory, antioxidant, and antilipase activities. Probiotics (Bac) supplements have good benefits on health and MeS, whether it is supplemented in combination with prebiotics or alone. Microencapsulation helps to mask unpalatable taste and odor and deliver active ingredients to targeted organs. Methodology MeS rat model was produced by feeding rats with a high fat, high CHO diet (HFD). Bos was extracted, and both Bos and the probiotic were microencapsulated with a spray drier. Female rats were divided into 5 groups (N8). HFD control, control normal receiving basic diet, HFD treated, from the start of the experiment, either with encapsulated Bos, Bac and Bos or Bac only, all treatments were received for eight weeks (after approval from NRC animal ethical committee). Serum was collected to analyze lipid profile, blood sugar, liver and kidney functions, antioxidants, leptin, and progesterone. Rat's organs and body fat were weighed and collected for histopathology. Statistical analysis was done by use of one way Anova test in the SPSS program. Results showed control of elevated body weight, lipid profile, and glucose levels as well as decrease of body fat index and improvement of histopathology of liver and heart, especially in combination. Conclusion: We concluded that both microencapsulated Bos and probiotics have a controlling effect on MeS parameters.

Keywords: metabolic syndrome, Boswellia serata, probiotic, micro-encapsulation, histopathology, liver steatosis

Procedia PDF Downloads 102
19735 The Applications of Wire Print in Composite Material Research and Fabrication Process

Authors: Hsu Yi-Chia, Hoy June-Hao

Abstract:

FDM (Fused Deposition Modeling) is a rapid proofing method without mold, however, high material and time costs have always been a major disadvantage. Wire-printing is the next generation technology that can more flexible, and also easier to apply on a 3D printer and robotic arms printing. It can create its own construction methods. The research is mainly divided into three parts. The first is about the method of parameterizing the generated paths and the conversion of g-code to the wire-printing. The second is about material attempts and the application of effects. Third, is about the improvement of the operation of mechanical equipment and the design of robotic tool-head. The purpose of this study is to develop a new wire-print method that can efficiently generate line segments and paths in three- dimensions space. The parametric modeling software transforms the digital model into a 3D printer or robotic arms g-code, this article uses thermoplastics/ clay/composites materials for testing. The combination of materials and wire-print process makes architects and designers have the ability to research and develop works and construction in the future.

Keywords: parametric software, wire print, robotic arms fabrication, composite filament additive manufacturing

Procedia PDF Downloads 131