Search results for: impact of preprocessing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10496

Search results for: impact of preprocessing

10466 Deep Learning for Qualitative and Quantitative Grain Quality Analysis Using Hyperspectral Imaging

Authors: Ole-Christian Galbo Engstrøm, Erik Schou Dreier, Birthe Møller Jespersen, Kim Steenstrup Pedersen

Abstract:

Grain quality analysis is a multi-parameterized problem that includes a variety of qualitative and quantitative parameters such as grain type classification, damage type classification, and nutrient regression. Currently, these parameters require human inspection, a multitude of instruments employing a variety of sensor technologies, and predictive model types or destructive and slow chemical analysis. This paper investigates the feasibility of applying near-infrared hyperspectral imaging (NIR-HSI) to grain quality analysis. For this study two datasets of NIR hyperspectral images in the wavelength range of 900 nm - 1700 nm have been used. Both datasets contain images of sparsely and densely packed grain kernels. The first dataset contains ~87,000 image crops of bulk wheat samples from 63 harvests where protein value has been determined by the FOSS Infratec NOVA which is the golden industry standard for protein content estimation in bulk samples of cereal grain. The second dataset consists of ~28,000 image crops of bulk grain kernels from seven different wheat varieties and a single rye variety. In the first dataset, protein regression analysis is the problem to solve while variety classification analysis is the problem to solve in the second dataset. Deep convolutional neural networks (CNNs) have the potential to utilize spatio-spectral correlations within a hyperspectral image to simultaneously estimate the qualitative and quantitative parameters. CNNs can autonomously derive meaningful representations of the input data reducing the need for advanced preprocessing techniques required for classical chemometric model types such as artificial neural networks (ANNs) and partial least-squares regression (PLS-R). A comparison between different CNN architectures utilizing 2D and 3D convolution is conducted. These results are compared to the performance of ANNs and PLS-R. Additionally, a variety of preprocessing techniques from image analysis and chemometrics are tested. These include centering, scaling, standard normal variate (SNV), Savitzky-Golay (SG) filtering, and detrending. The results indicate that the combination of NIR-HSI and CNNs has the potential to be the foundation for an automatic system unifying qualitative and quantitative grain quality analysis within a single sensor technology and predictive model type.

Keywords: deep learning, grain analysis, hyperspectral imaging, preprocessing techniques

Procedia PDF Downloads 67
10465 Selecting the Best Sub-Region Indexing the Images in the Case of Weak Segmentation Based on Local Color Histograms

Authors: Mawloud Mosbah, Bachir Boucheham

Abstract:

Color Histogram is considered as the oldest method used by CBIR systems for indexing images. In turn, the global histograms do not include the spatial information; this is why the other techniques coming later have attempted to encounter this limitation by involving the segmentation task as a preprocessing step. The weak segmentation is employed by the local histograms while other methods as CCV (Color Coherent Vector) are based on strong segmentation. The indexation based on local histograms consists of splitting the image into N overlapping blocks or sub-regions, and then the histogram of each block is computed. The dissimilarity between two images is reduced, as consequence, to compute the distance between the N local histograms of the both images resulting then in N*N values; generally, the lowest value is taken into account to rank images, that means that the lowest value is that which helps to designate which sub-region utilized to index images of the collection being asked. In this paper, we make under light the local histogram indexation method in the hope to compare the results obtained against those given by the global histogram. We address also another noteworthy issue when Relying on local histograms namely which value, among N*N values, to trust on when comparing images, in other words, which sub-region among the N*N sub-regions on which we base to index images. Based on the results achieved here, it seems that relying on the local histograms, which needs to pose an extra overhead on the system by involving another preprocessing step naming segmentation, does not necessary mean that it produces better results. In addition to that, we have proposed here some ideas to select the local histogram on which we rely on to encode the image rather than relying on the local histogram having lowest distance with the query histograms.

Keywords: CBIR, color global histogram, color local histogram, weak segmentation, Euclidean distance

Procedia PDF Downloads 333
10464 Organizational Performance and Impact of Social Innovation

Authors: Alfonso Unceta, Javier Castro-Spila

Abstract:

This paper offers a conceptual and empirical exploration between the organizational performance and the impact of social innovation. The paper contributes on the social innovation field in three domains: a) It provides analytical and empirical evidence linking organizational performance to the impact of social innovation; b) it provides a first outline of impact assessment of social innovation when it is developed by a diversity of heterogeneous actors (systemic social innovation); c) it provides a first outline for the development of innovation policies to support social innovations according to a typology of organizations and a typology of impact.

Keywords: absorptive capacity, social innovation impact, organizational performance, RESINDEX, Basque Country

Procedia PDF Downloads 448
10463 Investigation of Single Particle Breakage inside an Impact Mill

Authors: E. Ghasemi Ardi, K. J. Dong, A. B. Yu, R. Y. Yang

Abstract:

In current work, a numerical model based on the discrete element method (DEM) was developed which provided information about particle dynamic and impact event condition inside a laboratory scale impact mill (Fritsch). It showed that each particle mostly experiences three impacts inside the mill. While the first impact frequently happens at front surface of the rotor’s rib, the frequent location of the second impact is side surfaces of the rotor’s rib. It was also showed that while the first impact happens at small impact angle mostly varying around 35º, the second impact happens at around 70º which is close to normal impact condition. Also analyzing impact energy revealed that varying mill speed from 6000 to 14000 rpm, the ratio of first impact’s average impact energy and minimum required energy to break particle (Wₘᵢₙ) increased from 0.30 to 0.85. Moreover, it was seen that second impact poses intense impact energy on particle which can be considered as the main cause of particle splitting. Finally, obtained information from DEM simulation along with obtained data from conducted experiments was implemented in semi-empirical equations in order to find selection and breakage functions. Then, using a back-calculation approach, those parameters were used to predict the PSDs of ground particles under different impact energies. Results were compared with experiment results and showed reasonable accuracy and prediction ability.

Keywords: single particle breakage, particle dynamic, population balance model, particle size distribution, discrete element method

Procedia PDF Downloads 259
10462 Analysis of the Environmental Impact of Selected Small Heat and Power Plants Operating in Poland

Authors: M. Stelmachowski, M. Wojtczak

Abstract:

The aim of the work was to assess the environmental impact of the selected small and medium-sized companies supplying heat and electricity to the cities with a population of about 50,000 inhabitants. Evaluation and comparison of the impact on the environment have been carried out for the three plants producing heat and two CHP plants with particular attention to emissions into the atmosphere and the impact of introducing a system of trading carbon emissions of these companies.

Keywords: CO2 emission, district heating, heat and power plant, impact on environment

Procedia PDF Downloads 445
10461 A New Instrumented Drop-Weight Test Machine for Studying the Impact Behaviour of Reinforced Concrete Beams

Authors: M. Al-Farttoosi, M. Y. Rafiq, J. Summerscales, C. Williams

Abstract:

Structures can be subjected to impact loading from various sources like earthquake, tsunami, missiles and explosions. The impact loading can cause different degrees of damage to concrete structures. The demand for strengthening and rehabilitation of damaged structures is increasing. In recent years, Car0bon Fibre Reinforced Polymer (CFRP) matrix composites has gain more attention for strengthening and repairing these structures. To study the impact behaviour of the reinforced concrete (RC) beams strengthened or repaired using CFRP, a heavy impact test machine was designed and manufactured .The machine included a newly designed support system for beams together with various instrumentation. This paper describes the support design configuration of the impact test machine, instrumentation and dynamic analysis of the concrete beams. To evaluate the efficiency of the new impact test machine, experimental impact tests were conducted on simple supported reinforced concrete beam. Different methods were used to determine the impact force and impact response of the RC beams in terms of inertia force, maximum deflection, reaction force and fracture energy. The manufactured impact test machine was successfully used in testing RC beams under impact loading and used successfully to test the reinforced concrete beams strengthened or repaired using CFRP under impact loading.

Keywords: beam, concrete, impact, machine

Procedia PDF Downloads 388
10460 An Adaptive Oversampling Technique for Imbalanced Datasets

Authors: Shaukat Ali Shahee, Usha Ananthakumar

Abstract:

A data set exhibits class imbalance problem when one class has very few examples compared to the other class, and this is also referred to as between class imbalance. The traditional classifiers fail to classify the minority class examples correctly due to its bias towards the majority class. Apart from between-class imbalance, imbalance within classes where classes are composed of a different number of sub-clusters with these sub-clusters containing different number of examples also deteriorates the performance of the classifier. Previously, many methods have been proposed for handling imbalanced dataset problem. These methods can be classified into four categories: data preprocessing, algorithmic based, cost-based methods and ensemble of classifier. Data preprocessing techniques have shown great potential as they attempt to improve data distribution rather than the classifier. Data preprocessing technique handles class imbalance either by increasing the minority class examples or by decreasing the majority class examples. Decreasing the majority class examples lead to loss of information and also when minority class has an absolute rarity, removing the majority class examples is generally not recommended. Existing methods available for handling class imbalance do not address both between-class imbalance and within-class imbalance simultaneously. In this paper, we propose a method that handles between class imbalance and within class imbalance simultaneously for binary classification problem. Removing between class imbalance and within class imbalance simultaneously eliminates the biases of the classifier towards bigger sub-clusters by minimizing the error domination of bigger sub-clusters in total error. The proposed method uses model-based clustering to find the presence of sub-clusters or sub-concepts in the dataset. The number of examples oversampled among the sub-clusters is determined based on the complexity of sub-clusters. The method also takes into consideration the scatter of the data in the feature space and also adaptively copes up with unseen test data using Lowner-John ellipsoid for increasing the accuracy of the classifier. In this study, neural network is being used as this is one such classifier where the total error is minimized and removing the between-class imbalance and within class imbalance simultaneously help the classifier in giving equal weight to all the sub-clusters irrespective of the classes. The proposed method is validated on 9 publicly available data sets and compared with three existing oversampling techniques that rely on the spatial location of minority class examples in the euclidean feature space. The experimental results show the proposed method to be statistically significantly superior to other methods in terms of various accuracy measures. Thus the proposed method can serve as a good alternative to handle various problem domains like credit scoring, customer churn prediction, financial distress, etc., that typically involve imbalanced data sets.

Keywords: classification, imbalanced dataset, Lowner-John ellipsoid, model based clustering, oversampling

Procedia PDF Downloads 387
10459 Investigation of Building Pounding during Earthquake and Calculation of Impact Force between Two Adjacent Structures

Authors: H. Naderpour, R. C. Barros, S. M. Khatami

Abstract:

Seismic excitation is naturally caused large horizontal relative displacements, which is able to provide collisions between two adjacent buildings due to insufficient separation distance and severe damages are occurred due to impact especially in tall buildings. In this paper, an impact is numerically simulated and two needed parameters are calculated, including impact force and energy absorption. In order to calculate mentioned parameters, mathematical study needs to model an unreal link element, which is logically assumed to be spring and dashpot to determine lateral displacement and damping ratio of impact. For the determination of dynamic response of impact, a new equation of motion is theoretically suggested to evaluate impact force and energy dissipation. In order to confirm the rendered equation, a series of parametric study are performed and the accuracy of formula is confirmed.

Keywords: pounding, impact, dissipated energy, coefficient of restitution

Procedia PDF Downloads 319
10458 Prediction of Heavy-Weight Impact Noise and Vibration of Floating Floor Using Modified Impact Spectrum

Authors: Ju-Hyung Kim, Dae-Ho Mun, Hong-Gun Park

Abstract:

When an impact is applied to a floating floor, noise and vibration response of high-frequency range is reduced effectively, while amplifies the response at low-frequency range. This means floating floor can make worse noise condition when heavy-weight impact is applied. The amplified response is the result of interaction between finishing layer (mortar plate) and concrete slab. Because an impact force is not directly delivered to concrete slab, the impact force waveform or spectrum can be changed. In this paper, the changed impact spectrum was derived from several floating floor vibration tests. Based on the measured data, numerical modeling can describe the floating floor response, especially at low-frequency range. As a result, heavy-weight impact noise can be predicted using modified impact spectrum.

Keywords: floating floor, heavy-weight impact, prediction, vibration

Procedia PDF Downloads 329
10457 Optimising GIS in Cushioning the Environmental Impact of Infrastructural Projects

Authors: Akerele Akintunde Hareef

Abstract:

GIS is an integrating tool for storing, retrieving, manipulating, and analyzing spatial data. It is a tool which defines an area with respect to features and other relevant thematic delineations. On the other hand, Environmental Impact Assessment in short is both positive and negative impact of an infrastructure on an environment. Impact of infrastructural projects on the environment is an aspect of development that barely get extensive portion of pre-project execution phase and when they do, the effects are most times not implemented to cushion the impact they have on human and the environment. In this research, infrastructural projects like road constructions, water reticulation projects, building constructions, bridge etc. have immense impact on the environment and the people that reside in location of construction. Hence, the need for this research tends to portray the relevance of Environmental Impact assessment in calculating the vulnerability of human and the environment to imbalance necessitated by this infrastructural development and how the use of GIS application can be optimally applied to annul or minimize the effect.

Keywords: environmental impact assessment (EIA), geographic information system (GIS), infrastructural projects, environment

Procedia PDF Downloads 514
10456 The Effect of Floor Impact Sound Insulation Performance Using Scrambled Thermoplastic Poly Urethane and Ethylene Vinyl Acetate

Authors: Bonsoo Koo, Seong Shin Hong, Byung Kwon Lee

Abstract:

Most of apartments in Korea have wall type structure that present poor performance regarding floor impact sound insulation. In order to minimize the transmission of floor impact sound, flooring structures are used in which an insulating material, 30 mm thickness pad of EPS or EVA, is sandwiched between a concrete slab and the finished mortar. Generally, a single-material pad used for insulation has a heavyweight impact sound level of 44~47 dB with 210 mm thickness slab. This study provides an analysis of the floor impact sound insulation performance using thermoplastic poly urethane (TPU), ethylene vinyl acetate (EVA), and expanded polystyrene (EPS) materials with buffering performance. Following mock-up tests the effect of lightweight impact sound turned out to be similar but heavyweight impact sound was decreased by 3 dB compared to conventional single material insulation pad.

Keywords: floor impact sound, thermoplastic poly urethane, ethylene vinyl acetate, heavyweight impact sound

Procedia PDF Downloads 373
10455 Determination of Physical Properties of Crude Oil Distillates by Near-Infrared Spectroscopy and Multivariate Calibration

Authors: Ayten Ekin Meşe, Selahattin Şentürk, Melike Duvanoğlu

Abstract:

Petroleum refineries are a highly complex process industry with continuous production and high operating costs. Physical separation of crude oil starts with the crude oil distillation unit, continues with various conversion and purification units, and passes through many stages until obtaining the final product. To meet the desired product specification, process parameters are strictly followed. To be able to ensure the quality of distillates, routine analyses are performed in quality control laboratories based on appropriate international standards such as American Society for Testing and Materials (ASTM) standard methods and European Standard (EN) methods. The cut point of distillates in the crude distillation unit is very crucial for the efficiency of the upcoming processes. In order to maximize the process efficiency, the determination of the quality of distillates should be as fast as possible, reliable, and cost-effective. In this sense, an alternative study was carried out on the crude oil distillation unit that serves the entire refinery process. In this work, studies were conducted with three different crude oil distillates which are Light Straight Run Naphtha (LSRN), Heavy Straight Run Naphtha (HSRN), and Kerosene. These products are named after separation by the number of carbons it contains. LSRN consists of five to six carbon-containing hydrocarbons, HSRN consist of six to ten, and kerosene consists of sixteen to twenty-two carbon-containing hydrocarbons. Physical properties of three different crude distillation unit products (LSRN, HSRN, and Kerosene) were determined using Near-Infrared Spectroscopy with multivariate calibration. The absorbance spectra of the petroleum samples were obtained in the range from 10000 cm⁻¹ to 4000 cm⁻¹, employing a quartz transmittance flow through cell with a 2 mm light path and a resolution of 2 cm⁻¹. A total of 400 samples were collected for each petroleum sample for almost four years. Several different crude oil grades were processed during sample collection times. Extended Multiplicative Signal Correction (EMSC) and Savitzky-Golay (SG) preprocessing techniques were applied to FT-NIR spectra of samples to eliminate baseline shifts and suppress unwanted variation. Two different multivariate calibration approaches (Partial Least Squares Regression, PLS and Genetic Inverse Least Squares, GILS) and an ensemble model were applied to preprocessed FT-NIR spectra. Predictive performance of each multivariate calibration technique and preprocessing techniques were compared, and the best models were chosen according to the reproducibility of ASTM reference methods. This work demonstrates the developed models can be used for routine analysis instead of conventional analytical methods with over 90% accuracy.

Keywords: crude distillation unit, multivariate calibration, near infrared spectroscopy, data preprocessing, refinery

Procedia PDF Downloads 83
10454 Agile Software Effort Estimation Using Regression Techniques

Authors: Mikiyas Adugna

Abstract:

Effort estimation is among the activities carried out in software development processes. An accurate model of estimation leads to project success. The method of agile effort estimation is a complex task because of the dynamic nature of software development. Researchers are still conducting studies on agile effort estimation to enhance prediction accuracy. Due to these reasons, we investigated and proposed a model on LASSO and Elastic Net regression to enhance estimation accuracy. The proposed model has major components: preprocessing, train-test split, training with default parameters, and cross-validation. During the preprocessing phase, the entire dataset is normalized. After normalization, a train-test split is performed on the dataset, setting training at 80% and testing set to 20%. We chose two different phases for training the two algorithms (Elastic Net and LASSO) regression following the train-test-split. In the first phase, the two algorithms are trained using their default parameters and evaluated on the testing data. In the second phase, the grid search technique (the grid is used to search for tuning and select optimum parameters) and 5-fold cross-validation to get the final trained model. Finally, the final trained model is evaluated using the testing set. The experimental work is applied to the agile story point dataset of 21 software projects collected from six firms. The results show that both Elastic Net and LASSO regression outperformed the compared ones. Compared to the proposed algorithms, LASSO regression achieved better predictive performance and has acquired PRED (8%) and PRED (25%) results of 100.0, MMRE of 0.0491, MMER of 0.0551, MdMRE of 0.0593, MdMER of 0.063, and MSE of 0.0007. The result implies LASSO regression algorithm trained model is the most acceptable, and higher estimation performance exists in the literature.

Keywords: agile software development, effort estimation, elastic net regression, LASSO

Procedia PDF Downloads 17
10453 Data Science-Based Key Factor Analysis and Risk Prediction of Diabetic

Authors: Fei Gao, Rodolfo C. Raga Jr.

Abstract:

This research proposal will ascertain the major risk factors for diabetes and to design a predictive model for risk assessment. The project aims to improve diabetes early detection and management by utilizing data science techniques, which may improve patient outcomes and healthcare efficiency. The phase relation values of each attribute were used to analyze and choose the attributes that might influence the examiner's survival probability using Diabetes Health Indicators Dataset from Kaggle’s data as the research data. We compare and evaluate eight machine learning algorithms. Our investigation begins with comprehensive data preprocessing, including feature engineering and dimensionality reduction, aimed at enhancing data quality. The dataset, comprising health indicators and medical data, serves as a foundation for training and testing these algorithms. A rigorous cross-validation process is applied, and we assess their performance using five key metrics like accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC-ROC). After analyzing the data characteristics, investigate their impact on the likelihood of diabetes and develop corresponding risk indicators.

Keywords: diabetes, risk factors, predictive model, risk assessment, data science techniques, early detection, data analysis, Kaggle

Procedia PDF Downloads 36
10452 Improve Student Performance Prediction Using Majority Vote Ensemble Model for Higher Education

Authors: Wade Ghribi, Abdelmoty M. Ahmed, Ahmed Said Badawy, Belgacem Bouallegue

Abstract:

In higher education institutions, the most pressing priority is to improve student performance and retention. Large volumes of student data are used in Educational Data Mining techniques to find new hidden information from students' learning behavior, particularly to uncover the early symptom of at-risk pupils. On the other hand, data with noise, outliers, and irrelevant information may provide incorrect conclusions. By identifying features of students' data that have the potential to improve performance prediction results, comparing and identifying the most appropriate ensemble learning technique after preprocessing the data, and optimizing the hyperparameters, this paper aims to develop a reliable students' performance prediction model for Higher Education Institutions. Data was gathered from two different systems: a student information system and an e-learning system for undergraduate students in the College of Computer Science of a Saudi Arabian State University. The cases of 4413 students were used in this article. The process includes data collection, data integration, data preprocessing (such as cleaning, normalization, and transformation), feature selection, pattern extraction, and, finally, model optimization and assessment. Random Forest, Bagging, Stacking, Majority Vote, and two types of Boosting techniques, AdaBoost and XGBoost, are ensemble learning approaches, whereas Decision Tree, Support Vector Machine, and Artificial Neural Network are supervised learning techniques. Hyperparameters for ensemble learning systems will be fine-tuned to provide enhanced performance and optimal output. The findings imply that combining features of students' behavior from e-learning and students' information systems using Majority Vote produced better outcomes than the other ensemble techniques.

Keywords: educational data mining, student performance prediction, e-learning, classification, ensemble learning, higher education

Procedia PDF Downloads 72
10451 Brain Tumor Detection and Classification Using Pre-Trained Deep Learning Models

Authors: Aditya Karade, Sharada Falane, Dhananjay Deshmukh, Vijaykumar Mantri

Abstract:

Brain tumors pose a significant challenge in healthcare due to their complex nature and impact on patient outcomes. The application of deep learning (DL) algorithms in medical imaging have shown promise in accurate and efficient brain tumour detection. This paper explores the performance of various pre-trained DL models ResNet50, Xception, InceptionV3, EfficientNetB0, DenseNet121, NASNetMobile, VGG19, VGG16, and MobileNet on a brain tumour dataset sourced from Figshare. The dataset consists of MRI scans categorizing different types of brain tumours, including meningioma, pituitary, glioma, and no tumour. The study involves a comprehensive evaluation of these models’ accuracy and effectiveness in classifying brain tumour images. Data preprocessing, augmentation, and finetuning techniques are employed to optimize model performance. Among the evaluated deep learning models for brain tumour detection, ResNet50 emerges as the top performer with an accuracy of 98.86%. Following closely is Xception, exhibiting a strong accuracy of 97.33%. These models showcase robust capabilities in accurately classifying brain tumour images. On the other end of the spectrum, VGG16 trails with the lowest accuracy at 89.02%.

Keywords: brain tumour, MRI image, detecting and classifying tumour, pre-trained models, transfer learning, image segmentation, data augmentation

Procedia PDF Downloads 34
10450 A Study on the Life Prediction Performance Degradation Analysis of the Hydraulic Breaker

Authors: Jong Won, Park, Sung Hyun, Kim

Abstract:

The kinetic energy to pass subjected to shock and chisel reciprocating piston hydraulic power supplied by the excavator using for the purpose of crushing the rock, and roads, buildings, etc., hydraulic breakers blow. Impact frequency, efficiency measurement of the impact energy, hydraulic breakers, to demonstrate the ability of hydraulic breaker manufacturers and users to a very important item. And difficult in order to confirm the initial performance degradation in the life of the hydraulic breaker has been thought to be a problem.In this study, we measure the efficiency of hydraulic breaker, Impact energy and Impact frequency, the degradation analysis of research to predict the life.

Keywords: impact energy, impact frequency, hydraulic breaker, life prediction

Procedia PDF Downloads 400
10449 Prediction of Pounding between Two SDOF Systems by Using Link Element Based On Mathematic Relations and Suggestion of New Equation for Impact Damping Ratio

Authors: Seyed M. Khatami, H. Naderpour, R. Vahdani, R. C. Barros

Abstract:

Many previous studies have been carried out to calculate the impact force and the dissipated energy between two neighboring buildings during seismic excitation, when they collide with each other. Numerical studies are an important part of impact, which several researchers have tried to simulate the impact by using different formulas. Estimation of the impact force and the dissipated energy depends significantly on some parameters of impact. Mass of bodies, stiffness of spring, coefficient of restitution, damping ratio of dashpot and impact velocity are some known and unknown parameters to simulate the impact and measure dissipated energy during collision. Collision is usually shown by force-displacement hysteresis curve. The enclosed area of the hysteresis loop explains the dissipated energy during impact. In this paper, the effect of using different types of impact models is investigated in order to calculate the impact force. To increase the accuracy of impact model and to optimize the results of simulations, a new damping equation is assumed and is validated to get the best results of impact force and dissipated energy, which can show the accuracy of suggested equation of motion in comparison with other formulas. This relation is called "n-m". Based on mathematical relation, an initial value is selected for the mentioned coefficients and kinetic energy loss is calculated. After each simulation, kinetic energy loss and energy dissipation are compared with each other. If they are equal, selected parameters are true and, if not, the constant of parameters are modified and a new analysis is performed. Finally, two unknown parameters are suggested to estimate the impact force and calculate the dissipated energy.

Keywords: impact force, dissipated energy, kinetic energy loss, damping relation

Procedia PDF Downloads 521
10448 Considering Partially Developed Artifacts in Change Impact Analysis Implementation

Authors: Nazri Kama, Sufyan Basri, Roslina Ibrahim

Abstract:

It is important to manage the changes in the software to meet the evolving needs of the customer. Accepting too many changes causes delay in the completion and it incurs additional cost. One type of information that helps to make the decision is through change impact analysis. Current impact analysis approaches assume that all classes in the class artifact are completely developed and the class artifact is used as a source of analysis. However, these assumptions are impractical for impact analysis in the software development phase as some classes in the class artifact are still under development or partially developed that leads to inaccuracy. This paper presents a novel impact analysis approach to be used in the software development phase. The significant achievements of the approach are demonstrated through an extensive experimental validation using three case studies.

Keywords: software development, impact analysis, traceability, static analysis.

Procedia PDF Downloads 576
10447 Finite Element Analysis of Low-Velocity Impact Damage on Stiffened Composite Panels

Authors: Xuan Sun, Mingbo Tong

Abstract:

To understand the factors which affect impact damage on composite structures, particularly the effects of impact position and ribs. In this paper, a finite element model (FEM) of low-velocity impact damage on the composite structure was established via the nonlinear finite element method, combined with the user-defined materials subroutine (VUMAT) of the ABAQUS software. The structural elements chosen for the investigation comprised a series of stiffened composite panels, representative of real aircraft structure. By impacting the panels at different positions relative to the ribs, the effect of relative position of ribs was found out. Then the simulation results and the experiments data were compared. Finally, the factors which affect impact damage on the structures were discussed. The paper was helpful for the design of stiffened composite structures.

Keywords: stiffened, low-velocity impact, Abaqus, impact energy

Procedia PDF Downloads 590
10446 Principal Component Analysis Combined Machine Learning Techniques on Pharmaceutical Samples by Laser Induced Breakdown Spectroscopy

Authors: Kemal Efe Eseller, Göktuğ Yazici

Abstract:

Laser-induced breakdown spectroscopy (LIBS) is a rapid optical atomic emission spectroscopy which is used for material identification and analysis with the advantages of in-situ analysis, elimination of intensive sample preparation, and micro-destructive properties for the material to be tested. LIBS delivers short pulses of laser beams onto the material in order to create plasma by excitation of the material to a certain threshold. The plasma characteristics, which consist of wavelength value and intensity amplitude, depends on the material and the experiment’s environment. In the present work, medicine samples’ spectrum profiles were obtained via LIBS. Medicine samples’ datasets include two different concentrations for both paracetamol based medicines, namely Aferin and Parafon. The spectrum data of the samples were preprocessed via filling outliers based on quartiles, smoothing spectra to eliminate noise and normalizing both wavelength and intensity axis. Statistical information was obtained and principal component analysis (PCA) was incorporated to both the preprocessed and raw datasets. The machine learning models were set based on two different train-test splits, which were 70% training – 30% test and 80% training – 20% test. Cross-validation was preferred to protect the models against overfitting; thus the sample amount is small. The machine learning results of preprocessed and raw datasets were subjected to comparison for both splits. This is the first time that all supervised machine learning classification algorithms; consisting of Decision Trees, Discriminant, naïve Bayes, Support Vector Machines (SVM), k-NN(k-Nearest Neighbor) Ensemble Learning and Neural Network algorithms; were incorporated to LIBS data of paracetamol based pharmaceutical samples, and their different concentrations on preprocessed and raw dataset in order to observe the effect of preprocessing.

Keywords: machine learning, laser-induced breakdown spectroscopy, medicines, principal component analysis, preprocessing

Procedia PDF Downloads 62
10445 Classification of Impact Damages with Respect of Damage Tolerance Design Approach and Airworthiness Requirements

Authors: T. Mrna, R. Doubrava

Abstract:

This paper describes airworthiness requirements with respect damage tolerance. Damage tolerance determines the amount and magnitude of damage on parts of the airplane. Airworthiness requirements determine the amount of damage that can still be in flight capable of the condition. Component damage can be defined as barely visible impact damage, visible impact damage or clear visible impact damage. Damage is also distributed it according to the velocity. It is divided into low or high velocity impact damage. The severity of damage to the part of airplane divides the airworthiness requirements into several categories according to severity. Airworthiness requirements are determined by type airplane. All types of airplane do not have the same conditions for airworthiness requirements. This knowledge is important for designing and operating an airplane.

Keywords: airworthiness requirements, composite, damage tolerance, low and high velocity impact

Procedia PDF Downloads 536
10444 A Life Cycle Assessment (LCA) of Aluminum Production Process

Authors: Alaa Al Hawari, Mohammad Khader, Wael El Hasan, Mahmoud Alijla, Ammar Manawi, Abdelbaki Benamour

Abstract:

The production of aluminium alloys and ingots -starting from the processing of alumina to aluminium, and the final cast product- was studied using a Life Cycle Assessment (LCA) approach. The studied aluminium supply chain consisted of a carbon plant, a reduction plant, a casting plant, and a power plant. In the LCA model, the environmental loads of the different plants for the production of 1 ton of aluminium metal were investigated. The impact of the aluminium production was assessed in eight impact categories. The results showed that for all of the impact categories the power plant had the highest impact only in the cases of Human Toxicity Potential (HTP) the reduction plant had the highest impact and in the Marine Aquatic Eco-Toxicity Potential (MAETP) the carbon plant had the highest impact. Furthermore, the impact of the carbon plant and the reduction plant combined was almost the same as the impact of the power plant in the case of the Acidification Potential (AP). The carbon plant had a positive impact on the environment when it comes to the Eutrophication Potential (EP) due to the production of clean water in the process. The natural gas based power plant used in the case study had 8.4 times less negative impact on the environment when compared to the heavy fuel based power plant and 10.7 times less negative impact when compared to the hard coal based power plant.

Keywords: life cycle assessment, aluminium production, supply chain, ecological impacts

Procedia PDF Downloads 488
10443 A Co-Relational Descriptive Study to Assess the Impact of Cancer Event on Self, Family, Coping Level of Cancer Clients and Quality of Life among Them

Authors: Padma Sree Potru

Abstract:

Abstract: A co-relational descriptive study was conducted to assess the impact of cancer event on self, on family, coping strategies of cancer clients and quality of life among them in G.G.H., Guntur, Andhra Pradesh, India. Aim: The aim of the study was to investigate the impact of cancer events on self, on family, coping of clients and quality of life among cancer patients. Methods: 50 cancer patients were selected through random sampling technique. The data were obtained by using impact of events scale, impact on family scale, coping health inventory and WHOQOL-BREF scale. Results: The results revealed that majority (32%) of them were in the age group of 36-45 years, 72% were females, 44% were having the income of Rs. 5001-10000/- per month, 40% were working for daily wage, and 15% were newly diagnosed of cancer. Among 50 cancer patients, 65% had extreme impact of events, 61% shows extreme impact on family, 46% possess minimal coping strategies and 68% had poor quality of life. This study focuses on that there is a strong positive correlation between quality of life and coping behavior r=0.603 and also between impact of event and impact on family r=0.610, but a negative correlation existed between quality of life and impact of events r= -0.201. ANOVA test reveals that there is a significant difference between subscales of impact on family and coping behavior with f values = 3.893, 3.957 respectively. Chi-square highlights that there is a significant association between impact of events with age, occupation and impact on family with duration of illness. Conclusion: Even though cancer is a dreadful disease still there are many emerging treatment modalities and innovative procedures which are focusing on improving the standards of life among cancer clients. But all this can happen only when the clients accepts the reality, increase their willpower and confidence, desire to live, focusing on coping mechanisms and good ongoing support from the family members.

Keywords: impact of event, impact on family, coping, quality of event

Procedia PDF Downloads 416
10442 Developing a Hybrid Method to Diagnose and Predict Sports Related Concussions with Machine Learning

Authors: Melody Yin

Abstract:

Concussions impact a large amount of adolescents; they make up as much as half of the diagnosed concussions in America. This research proposes a hybrid machine learning model based on the combination of human/knowledge-based domains and computer-generated feature rankings to improve the accuracy of diagnosing sports related concussion (SRC). Using a data set of symptoms collected on the sideline post-SRC events, the symptom selection criteria method has been developed by using Google AutoML's important score function to identify the top 10 symptom features. In addition, symptom domains have been introduced as another parameter, categorizing the symptoms into physical, cognitive, sleep, and emotional domains. The hybrid machine learning model has been trained with a combination of the top 10 symptoms and 4 domains. From the results, the hybrid model was the best performer for symptom resolution time prediction in 2 and 4-week thresholds. This research is a proof of concept study in the use of domains along with machine learning in order to improve concussion prediction accuracy. It is also possible that the use of domains can make the model more efficient due to reduced training time. This research examines the use of a hybrid method in predicting sports-related concussion. This achievement is based on data preprocessing, using a hybrid method to select criteria to achieve high performance.

Keywords: hybrid model, machine learning, sports related concussion, symptom resolution time

Procedia PDF Downloads 138
10441 Procedure for Impact Testing of Fused Recycled Glass

Authors: David Halley, Tyra Oseng-Rees, Luca Pagano, Juan A Ferriz-Papi

Abstract:

Recycled glass material is made from 100% recycled bottle glass and consumes less energy than re-melt technology. It also uses no additives in the manufacturing process allowing the recycled glass material, in principal, to go back to the recycling stream after end-of-use, contributing to the circular economy with a low ecological impact. The aim of this paper is to investigate the procedure for testing the recycled glass material for impact resistance, so it can be applied to pavements and other surfaces which are at risk of impact during service. A review of different impact test procedures for construction materials was undertaken, comparing methodologies and international standards applied to other materials such as natural stone, ceramics and glass. A drop weight impact testing machine was designed and manufactured in-house to perform these tests. As a case study, samples of the recycled glass material were manufactured with two different thicknesses and tested. The impact energy was calculated theoretically, obtaining results with 5 and 10 J. The results on the material were subsequently discussed. Improvements on the procedure can be made using high speed video technology to calculate velocity just before and immediately after the impact to know the absorbed energy. The initial results obtained in this procedure were positive although repeatability needs to be developed to obtain a correlation of results and finally be able to validate the procedure. The experiment with samples showed the practicality of this procedure and application to the recycled glass material impact testing although further research needs to be developed.

Keywords: construction materials, drop weight impact, impact testing, recycled glass

Procedia PDF Downloads 262
10440 Energy Absorption Characteristic of a Coupler Rubber Buffer Used in Rail Vehicles

Authors: Zhixiang Li, Shuguang Yao, Wen Ma

Abstract:

Coupler rubber buffer has been widely applied on the high-speed trains and the main function of the rubber buffer is dissipating the impact energy between vehicles. The rubber buffer consists of two groups of rubbers, which are both pre-compressed and then installed into the frame body. This paper focuses on the energy absorption characteristics of the rubber buffers particularly. Firstly, the quasi-static compression tests were carried out for 1 and 3 pairs of rubber sheets and some energy absorption responses relationship, i.e. Eabn = n×Eab1, Edissn = n×Ediss1, and Ean = Ea1, were obtained. Next, a series of quasi-static tests were performed for 1 pair of rubber sheet to investigate the energy absorption performance with different compression ratio of the rubber buffers. Then the impact tests with five impact velocities were conducted and the coupler knuckle was destroyed when the impact velocity was 10.807 km/h. The impact tests results showed that with the increase of impact velocity, the Eab, Ediss and Ea of rear buffer increased a lot, but the three responses of front buffer had not much increase. Finally, the results of impact tests and quasi-static tests were contrastively analysed and the results showed that with the increase of the stroke, the values of Eab, Ediss, and Ea were all increase. However, the increasing rates of impact tests were all larger than that of quasi-static tests. The maximum value of Ea was 68.76% in impact tests, it was a relatively high value for vehicle coupler buffer. The energy capacity of the rear buffer was determined for dynamic loading, it was 22.98 kJ.

Keywords: rubber buffer, coupler, energy absorption, impact tests

Procedia PDF Downloads 156
10439 The Impact of Trade Liberalization on Current Account Deficit: The Turkish Case

Authors: E. Selçuk, Z. Karaçor, P. Yardımcı

Abstract:

Trade liberalization and its effects on the economies of developing countries have been investigated by many different studies, and some of them have focused on its impact on the current account balance. Turkey, as being one of the countries, which has liberalized its foreign trade in the 1980s, also needs to be studied in terms of the impact of liberalization on current account deficits. Therefore, the aim of this study is to find out whether trade liberalization has affected Turkey’s trade and current account balances. In order to determine this, yearly data of Turkey from 1980 to 2013 is used. As liberalization dummy, the year 1989, which was set for Turkey, is selected. Structural break test and model estimation results show that trade liberalization has a negative impact on trade balance but do not have a significant impact on the current account balance.

Keywords: budget deficit, liberalization, Turkish economy, current account

Procedia PDF Downloads 349
10438 Impact Modified Oil Palm Empty Fruit Bunch Fiber/Poly(Lactic) Acid Composite

Authors: Mohammad D. H. Beg, John O. Akindoyo, Suriati Ghazali, Abdullah A. Mamun

Abstract:

In this study, composites were fabricated from oil palm empty fruit bunch fiber and poly(lactic) acid by extrusion followed by injection moulding. Surface of the fiber was pre-treated by ultrasound in an alkali medium and treatment efficiency was investigated by scanning electron microscopy (SEM) analysis and Fourier transforms infrared spectrometer (FTIR). Effect of fiber treatment on composite was characterized by tensile strength (TS), tensile modulus (TM) and impact strength (IS). Furthermore, biostrong impact modifier was incorporated into the treated fiber composite to improve its impact properties. Mechanical testing showed an improvement of up to 23.5% and 33.6% respectively for TS and TM of treated fiber composite above untreated fiber composite. On the other hand incorporation of impact modifier led to enhancement of about 20% above the initial IS of the treated fiber composite.

Keywords: fiber treatment, impact modifier, natural fibers, ultrasound

Procedia PDF Downloads 450
10437 Prediction of Alzheimer's Disease Based on Blood Biomarkers and Machine Learning Algorithms

Authors: Man-Yun Liu, Emily Chia-Yu Su

Abstract:

Alzheimer's disease (AD) is the public health crisis of the 21st century. AD is a degenerative brain disease and the most common cause of dementia, a costly disease on the healthcare system. Unfortunately, the cause of AD is poorly understood, furthermore; the treatments of AD so far can only alleviate symptoms rather cure or stop the progress of the disease. Currently, there are several ways to diagnose AD; medical imaging can be used to distinguish between AD, other dementias, and early onset AD, and cerebrospinal fluid (CSF). Compared with other diagnostic tools, blood (plasma) test has advantages as an approach to population-based disease screening because it is simpler, less invasive also cost effective. In our study, we used blood biomarkers dataset of The Alzheimer’s disease Neuroimaging Initiative (ADNI) which was funded by National Institutes of Health (NIH) to do data analysis and develop a prediction model. We used independent analysis of datasets to identify plasma protein biomarkers predicting early onset AD. Firstly, to compare the basic demographic statistics between the cohorts, we used SAS Enterprise Guide to do data preprocessing and statistical analysis. Secondly, we used logistic regression, neural network, decision tree to validate biomarkers by SAS Enterprise Miner. This study generated data from ADNI, contained 146 blood biomarkers from 566 participants. Participants include cognitive normal (healthy), mild cognitive impairment (MCI), and patient suffered Alzheimer’s disease (AD). Participants’ samples were separated into two groups, healthy and MCI, healthy and AD, respectively. We used the two groups to compare important biomarkers of AD and MCI. In preprocessing, we used a t-test to filter 41/47 features between the two groups (healthy and AD, healthy and MCI) before using machine learning algorithms. Then we have built model with 4 machine learning methods, the best AUC of two groups separately are 0.991/0.709. We want to stress the importance that the simple, less invasive, common blood (plasma) test may also early diagnose AD. As our opinion, the result will provide evidence that blood-based biomarkers might be an alternative diagnostics tool before further examination with CSF and medical imaging. A comprehensive study on the differences in blood-based biomarkers between AD patients and healthy subjects is warranted. Early detection of AD progression will allow physicians the opportunity for early intervention and treatment.

Keywords: Alzheimer's disease, blood-based biomarkers, diagnostics, early detection, machine learning

Procedia PDF Downloads 292