Search results for: erosion rate prediction
9182 Predictive Factors of Nasal Continuous Positive Airway Pressure (NCPAP) Therapy Success in Preterm Neonates with Hyaline Membrane Disease (HMD)
Authors: Novutry Siregar, Afdal, Emilzon Taslim
Abstract:
Hyaline Membrane Disease (HMD) is the main cause of respiratory failure in preterm neonates caused by surfactant deficiency. Nasal Continuous Positive Airway Pressure (NCPAP) is the therapy for HMD. The success of therapy is determined by gestational age, birth weight, HMD grade, time of NCAP administration, and time of breathing frequency recovery. The aim of this research is to identify the predictive factor of NCPAP therapy success in preterm neonates with HMD. This study used a cross-sectional design by using medical records of patients who were treated in the Perinatology of the Pediatric Department of Dr. M. Djamil Padang Central Hospital from January 2015 to December 2017. The samples were eighty-two neonates that were selected by using the total sampling technique. Data analysis was done by using the Chi-Square Test and the Multiple Logistic Regression Prediction Model. The results showed the success rate of NCPAP therapy reached 53.7%. Birth weight (p = 0.048, OR = 3.34 95% CI 1.01-11.07), HMD grade I (p = 0.018, OR = 4.95 CI 95% 1.31-18.68), HMD grade II (p = 0.044, OR = 5.52 95% CI 1.04-29.15), and time of breathing frequency recovery (p = 0,000, OR = 13.50 95% CI 3.58-50, 83) are the predictive factors of NCPAP therapy success in preterm neonates with HMD. The most significant predictive factor is the time of breathing frequency recovery.Keywords: predictive factors, the success of therapy, NCPAP, preterm neonates, HMD
Procedia PDF Downloads 639181 Multi-Objective Optimization of Electric Discharge Machining for Inconel 718
Authors: Pushpendra S. Bharti, S. Maheshwari
Abstract:
Electric discharge machining (EDM) is one of the most widely used non-conventional manufacturing process to shape difficult-to-cut materials. The process yield, in terms of material removal rate, surface roughness and tool wear rate, of EDM may considerably be improved by selecting the optimal combination(s) of process parameters. This paper employs Multi-response signal-to-noise (MRSN) ratio technique to find the optimal combination(s) of the process parameters during EDM of Inconel 718. Three cases v.i.z. high cutting efficiency, high surface finish, and normal machining have been taken and the optimal combinations of input parameters have been obtained for each case. Analysis of variance (ANOVA) has been employed to find the dominant parameter(s) in all three cases. The experimental verification of the obtained results has also been made. MRSN ratio technique found to be a simple and effective multi-objective optimization technique.Keywords: electric discharge machining, material removal rate, surface roughness, too wear rate, multi-response signal-to-noise ratio, multi response signal-to-noise ratio, optimization
Procedia PDF Downloads 3559180 Analytical Solutions for Tunnel Collapse Mechanisms in Circular Cross-Section Tunnels under Seepage and Seismic Forces
Authors: Zhenyu Yang, Qiunan Chen, Xiaocheng Huang
Abstract:
Reliable prediction of tunnel collapse remains a prominent challenge in the field of civil engineering. In this study, leveraging the nonlinear Hoek-Brown failure criterion and the upper-bound theorem, an analytical solution for the collapse surface of shallowly buried circular tunnels was derived, taking into account the coupled effects of surface loads and pore water pressures. Initially, surface loads and pore water pressures were introduced as external force factors, equating the energy dissipation rate to the external force, yielding our objective function. Subsequently, the variational method was employed for optimization, and the outcomes were juxtaposed with previous research findings. Furthermore, we utilized the deduced equation set to systematically analyze the influence of various rock mass parameters on collapse shape and extent. To validate our analytical solutions, a comparison with prior studies was executed. The corroboration underscored the efficacy of our proposed methodology, offering invaluable insights for collapse risk assessment in practical engineering applications.Keywords: tunnel roof stability, analytical solution, hoek–brown failure criterion, limit analysis
Procedia PDF Downloads 849179 Effect of Tool Size and Cavity Depth on Response Characteristics during Electric Discharge Machining on Superalloy Metal - An Experimental Investigation
Authors: Sudhanshu Kumar
Abstract:
Electrical discharge machining, also known as EDM, process is one of the most applicable machining process for removal of material in hard to machine materials like superalloy metals. EDM process utilizes electrical energy into sparks to erode the metals in presence of dielectric medium. In the present investigation, superalloy, Inconel 718 has been selected as workpiece and electrolytic copper as tool electrode. Attempt has been made to understand the effect of size of tool with varying cavity depth during drilling of hole through EDM process. In order to systematic investigate, tool size in terms of tool diameter and cavity depth along with other important electrical parameters namely, peak current, pulse-on time and servo voltage have been varied at three different values and the experiments has been designed using fractional factorial (Taguchi) method. Each experiment has been repeated twice under the same condition in order to understand the variability within the experiments. The effect of variations in parameters has been evaluated in terms of material removal rate, tool wear rate and surface roughness. Results revel that change in tool diameter during machining affects the response characteristics significantly. Larger tool diameter yielded 13% more material removal rate than smaller tool diameter. Analysis of the effect of variation in cavity depth is notable. There is no significant effect of cavity depth on material removal rate, tool wear rate and surface quality. This indicates that number of experiments can be performed to analyze other parameters effect even at smaller depth of cavity which can reduce the cost and time of experiments. Further, statistical analysis has been carried out to identify the interaction effect between parameters.Keywords: EDM, Inconel 718, material removal rate, roughness, tool wear, tool size
Procedia PDF Downloads 2179178 Verification of Simulated Accumulated Precipitation
Authors: Nato Kutaladze, George Mikuchadze, Giorgi Sokhadze
Abstract:
Precipitation forecasts are one of the most demanding applications in numerical weather prediction (NWP). Georgia, as the whole Caucasian region, is characterized by very complex topography. The country territory is prone to flash floods and mudflows, quantitative precipitation estimation (QPE) and quantitative precipitation forecast (QPF) at any leading time are very important for Georgia. In this study, advanced research weather forecasting model’s skill in QPF is investigated over Georgia’s territory. We have analyzed several convection parameterization and microphysical scheme combinations for different rainy episodes and heavy rainy phenomena. We estimate errors and biases in accumulated 6 h precipitation using different spatial resolution during model performance verification for 12-hour and 24-hour lead time against corresponding rain gouge observations and satellite data. Various statistical parameters have been calculated for the 8-month comparison period, and some skills of model simulation have been evaluated. Our focus is on the formation and organization of convective precipitation systems in a low-mountain region. Several problems in connection with QPF have been identified for mountain regions, which include the overestimation and underestimation of precipitation on the windward and lee side of the mountains, respectively, and a phase error in the diurnal cycle of precipitation leading to the onset of convective precipitation in model forecasts several hours too early.Keywords: extremal dependence index, false alarm, numerical weather prediction, quantitative precipitation forecasting
Procedia PDF Downloads 1519177 Combustion Characteristic of Propane/Acetylene Fuel Blends Pool Fire
Authors: Yubo Bi, Xiao Chen, Shouxiang Lu
Abstract:
A kind of gas-fueled burner, named Burning Rate Emulator, was proposed for the purpose of the emulation of condensed fuel recently. The gaseous fuel can be pure combustible fuel gas or blends of gaseous fuel or inert gas. However, this concept was recently proposed without detailed study on the combustion characteristic of fuel blends. In this study, two kinds of common gaseous fuels were selected, propane and acetylene, to provide the combustion heat as well as a large amount of smoke, which widely exists in liquid and solid fuel burning process. A set of experiments were carried out using a gas-fueled burner with a diameter of 8 cm. The total volume flow rate of propane and acetylene was kept at 3 liters per minute. The volume fraction of propane varied from 0% to 100% at interval of 10%. It is found that the flame height increases with propane volume fraction, which may be caused by the increase of heat release rate, as the energy density of propane is larger than that of acetylene. The dimensionless flame height is correlated against dimensionless heat release rate, which shows a power function relationship. The radiation fraction of the flame does not show a monotonic relationship with propane volume fraction. With the increase of propane volume fraction from 0% to 100%, the value of radiation fraction increases first and reach a maximum value around 0.46 at a propane volume fraction of 10%, and then decreases continuously to a value of 0.25 at the propane volume fraction of 100%. The flame radiation is related to the soot in the flame. The trend of the radiation fraction reflects that there may be a synergistic effect of soot formation between propane and acetylene which can be guessed from the significantly high radiation fraction at a propane volume fraction of 10%. This work provides data for combustion of gaseous fuel blends pool fire and also give reference on the design of Burning Rate Emulator.Keywords: Burning Rate Emulator, fuel blends pool fire, flame height, radiation fraction
Procedia PDF Downloads 2289176 Community Participation of the Villagers: Corporate Social Responsibility Programme in Pantai Harapan Jaya Village, Bekasi Regency, West Java
Authors: Auliya Adzillatin Uzhma, Ismu Rini Dwi Ari, I. Nyoman Suluh Wijaya
Abstract:
Corporate Social Responsibility (CSR) programme in Pantai Harapan Jaya village is cultivation of mangrove and fishery capital distribution, to achieve the goal the CSR programme needed participation from the society in it. Moeliono in Fahrudin (2011) mentioned that participation from society is based by intrinsic reason from inside people it self and extrinsic reason from the other who related to him or from connection with other people. The fundamental connection who caused more boundaries from action which the organization can do called the social structure. The purpose of this research is to know the form of public participation and the density of the villager and people who is participated in CSR programme. This research use Social Network Analysis method by knew the Rate of Participation and Density. The result of the research is people who is involved in the programme is lived in Dusun Pondok Dua and they work in fisheries field. Rate of Participation is 11,61 and that means people involved in 11 or 12 activites of CSR Programme. The rate of participation of CSR Programme is categorized as high rate participation. The density value from the participant is 0.516 it’s mean that 51.6% of the people that participated is involved in the same step of CSR programme.Keywords: community participation, social network analysis, corporate social responsibility, urban and regional studies
Procedia PDF Downloads 5199175 Radial Distribution Network Reliability Improvement by Using Imperialist Competitive Algorithm
Authors: Azim Khodadadi, Sahar Sadaat Vakili, Ebrahim Babaei
Abstract:
This study presents a numerical method to optimize the failure rate and repair time of a typical radial distribution system. Failure rate and repair time are effective parameters in customer and energy based indices of reliability. Decrease of these parameters improves reliability indices. Thus, system stability will be boost. The penalty functions indirectly reflect the cost of investment which spent to improve these indices. Constraints on customer and energy based indices, i.e. SAIFI, SAIDI, CAIDI and AENS have been considered by using a new method which reduces optimization algorithm controlling parameters. Imperialist Competitive Algorithm (ICA) used as main optimization technique and particle swarm optimization (PSO), simulated annealing (SA) and differential evolution (DE) has been applied for further investigation. These algorithms have been implemented on a test system by MATLAB. Obtained results have been compared with each other. The optimized values of repair time and failure rate are much lower than current values which this achievement reduced investment cost and also ICA gives better answer than the other used algorithms.Keywords: imperialist competitive algorithm, failure rate, repair time, radial distribution network
Procedia PDF Downloads 6699174 The Kafrah Dam (The Oldest Dam in History)
Authors: Mohamed Bekhit Gad Khalil
Abstract:
This dam is the oldest dam in history. It was built by the ancient Egyptian around (2650 B.C) control flooding. It is believed to have been built between the third and fourth dynasties .It contains the oldest dam in history. Many studies have been conducted for the dam. This report was prepared under my supervision and in cooperation with the Ministry of Tourism and Antiquities. The dam was re-documented and photographed again. The dam on the northern side Consists of irregularly shaped stones of varying sizes used randomly. Sand and soil fill the gaps between the stones. creating layers to form the body of the dam. The eastern. side of the dam Consists of a series of regular shaped stones that have been cut and constructed into a stepped pyramid-like structure with width of (15,7) meters and height of (10) meters. The surface has significant erosion and wear on the stones due to weather Conditions. which has resulted in deep cavities in most of the stone blocks forming the surface.Keywords: ministry of tourism and antiquities, excavations, registration, documentation
Procedia PDF Downloads 359173 Integrating Artificial Neural Network and Taguchi Method on Constructing the Real Estate Appraisal Model
Authors: Mu-Yen Chen, Min-Hsuan Fan, Chia-Chen Chen, Siang-Yu Jhong
Abstract:
In recent years, real estate prediction or valuation has been a topic of discussion in many developed countries. Improper hype created by investors leads to fluctuating prices of real estate, affecting many consumers to purchase their own homes. Therefore, scholars from various countries have conducted research in real estate valuation and prediction. With the back-propagation neural network that has been popular in recent years and the orthogonal array in the Taguchi method, this study aimed to find the optimal parameter combination at different levels of orthogonal array after the system presented different parameter combinations, so that the artificial neural network obtained the most accurate results. The experimental results also demonstrated that the method presented in the study had a better result than traditional machine learning. Finally, it also showed that the model proposed in this study had the optimal predictive effect, and could significantly reduce the cost of time in simulation operation. The best predictive results could be found with a fewer number of experiments more efficiently. Thus users could predict a real estate transaction price that is not far from the current actual prices.Keywords: artificial neural network, Taguchi method, real estate valuation model, investors
Procedia PDF Downloads 4909172 Scoring System for the Prognosis of Sepsis Patients in Intensive Care Units
Authors: Javier E. García-Gallo, Nelson J. Fonseca-Ruiz, John F. Duitama-Munoz
Abstract:
Sepsis is a syndrome that occurs with physiological and biochemical abnormalities induced by severe infection and carries a high mortality and morbidity, therefore the severity of its condition must be interpreted quickly. After patient admission in an intensive care unit (ICU), it is necessary to synthesize the large volume of information that is collected from patients in a value that represents the severity of their condition. Traditional severity of illness scores seeks to be applicable to all patient populations, and usually assess in-hospital mortality. However, the use of machine learning techniques and the data of a population that shares a common characteristic could lead to the development of customized mortality prediction scores with better performance. This study presents the development of a score for the one-year mortality prediction of the patients that are admitted to an ICU with a sepsis diagnosis. 5650 ICU admissions extracted from the MIMICIII database were evaluated, divided into two groups: 70% to develop the score and 30% to validate it. Comorbidities, demographics and clinical information of the first 24 hours after the ICU admission were used to develop a mortality prediction score. LASSO (least absolute shrinkage and selection operator) and SGB (Stochastic Gradient Boosting) variable importance methodologies were used to select the set of variables that make up the developed score; each of this variables was dichotomized and a cut-off point that divides the population into two groups with different mean mortalities was found; if the patient is in the group that presents a higher mortality a one is assigned to the particular variable, otherwise a zero is assigned. These binary variables are used in a logistic regression (LR) model, and its coefficients were rounded to the nearest integer. The resulting integers are the point values that make up the score when multiplied with each binary variables and summed. The one-year mortality probability was estimated using the score as the only variable in a LR model. Predictive power of the score, was evaluated using the 1695 admissions of the validation subset obtaining an area under the receiver operating characteristic curve of 0.7528, which outperforms the results obtained with Sequential Organ Failure Assessment (SOFA), Oxford Acute Severity of Illness Score (OASIS) and Simplified Acute Physiology Score II (SAPSII) scores on the same validation subset. Observed and predicted mortality rates within estimated probabilities deciles were compared graphically and found to be similar, indicating that the risk estimate obtained with the score is close to the observed mortality, it is also observed that the number of events (deaths) is indeed increasing as the outcome go from the decile with the lowest probabilities to the decile with the highest probabilities. Sepsis is a syndrome that carries a high mortality, 43.3% for the patients included in this study; therefore, tools that help clinicians to quickly and accurately predict a worse prognosis are needed. This work demonstrates the importance of customization of mortality prediction scores since the developed score provides better performance than traditional scoring systems.Keywords: intensive care, logistic regression model, mortality prediction, sepsis, severity of illness, stochastic gradient boosting
Procedia PDF Downloads 2249171 State Estimation of a Biotechnological Process Using Extended Kalman Filter and Particle Filter
Authors: R. Simutis, V. Galvanauskas, D. Levisauskas, J. Repsyte, V. Grincas
Abstract:
This paper deals with advanced state estimation algorithms for estimation of biomass concentration and specific growth rate in a typical fed-batch biotechnological process. This biotechnological process was represented by a nonlinear mass-balance based process model. Extended Kalman Filter (EKF) and Particle Filter (PF) was used to estimate the unmeasured state variables from oxygen uptake rate (OUR) and base consumption (BC) measurements. To obtain more general results, a simplified process model was involved in EKF and PF estimation algorithms. This model doesn’t require any special growth kinetic equations and could be applied for state estimation in various bioprocesses. The focus of this investigation was concentrated on the comparison of the estimation quality of the EKF and PF estimators by applying different measurement noises. The simulation results show that Particle Filter algorithm requires significantly more computation time for state estimation but gives lower estimation errors both for biomass concentration and specific growth rate. Also the tuning procedure for Particle Filter is simpler than for EKF. Consequently, Particle Filter should be preferred in real applications, especially for monitoring of industrial bioprocesses where the simplified implementation procedures are always desirable.Keywords: biomass concentration, extended Kalman filter, particle filter, state estimation, specific growth rate
Procedia PDF Downloads 4309170 Design, Modeling, Fabrication, and Testing of a Scaled down Hybrid Rocket Engine
Authors: Pawthawala Nancy Manish, Syed Alay Hashim
Abstract:
A hybrid rocket is a rocket engine which uses propellants in two different states of matter- one is in solid and the other either gas or liquid. A hybrid rocket exhibit advantages over both liquid rockets and solid rockets especially in terms of simplicity, stop-start-restart capabilities, safety and cost. This paper deals the design and development of a hybrid rocket having paraffin wax as solid fuel and liquid oxygen as oxidizer. Due to variation of pressure in combustion chamber there is significantly change in mass flow rate, burning rate and uneven regression along the length of the grain. This project describes the working model of a hybrid propellant rocket motor. We have designed a hybrid rocket thrust chamber based on the predetermined combustion chamber pressure and the properties of hybrid propellant. This project is all ready in working condition with normal oxygen injector. Now we have planned to modify the injector design to improve the combustion property. We will use spray type injector for injecting the oxidizer. This idea will increase the performance followed by the regression rate of the solid fuel. By employing mass conservation law, oxygen mass flux, oxidizer/fuel ratio and regression rate the thrust coefficient can be obtained for our current design. CATIA V5 R20 is our design software for the complete setup. This project is fully based on experimental evaluation and the collection of combustion and flow parameters. The thrust chamber is made of stainless steel and the duration of test is around 15-20 seconds (Maximum). These experiments indicates that paraffin based fuel provides the opportunity to satisfy a broad range of mission requirements for the next generation of the hybrid rocket system.Keywords: burning rate, liquid oxygen, mass flow rate, paraffin wax and sugar
Procedia PDF Downloads 3369169 Evaluation of Critical Rate in Mature Oil Field with Dynamic Oil Rim Fluid Contacts in the Niger Delta
Authors: Stanley Ibuchukwu Onwukwe
Abstract:
Most reservoir in mature oil fields are vulnerable to challenges of water and/or gas coning as the size of their oil column reduces due to long period of oil production. These often result to low oil production and excessive water and/or gas production. Since over 50 years of oil production in the Niger delta, it is apparent that most of the oil fields in the region have reached their mature stages, thereby susceptible to coning tendencies. As a result of these, a good number of wells have been shut-in and abandoned, with significant amount of oil left unproduced. Analysis of the movement of fluid contacts in the reservoir is a significant aspect of reservoir studies and can assist in the management of coning tendencies and production performance of reservoirs in a mature field. This study, therefore, seeks to evaluate the occurrence of coning through the movement of fluid contacts (GOC and OWC) and determine the critical rate for controlling coning tendencies in mature oil field. This study applies the principle of Nodal analysis to calibrate the thin oil column of a reservoir of a mature field, and was graphically evaluated using the Joshi’s equation of critical rate for gas-oil system and oil-water system respectively. A representative Proxy equation was developed and sensitivity analysis carried out to determine the trend of critical rate as the oil column is been depleted. The result shows the trend in the movement of the GOC and OWC, and the critical rate, beyond which will result in excessive water and gas production, resulting to decreasing oil production from the reservoir. This result of this study can be used as a first pass assessment in the development of mature oil field reservoirs anticipated to experience water and/or gas coning during production.Keywords: coning, fluid contact movement, mature oil field, oil production
Procedia PDF Downloads 2459168 The Direct Deconvolution Model for the Large Eddy Simulation of Turbulence
Authors: Ning Chang, Zelong Yuan, Yunpeng Wang, Jianchun Wang
Abstract:
Large eddy simulation (LES) has been extensively used in the investigation of turbulence. LES calculates the grid-resolved large-scale motions and leaves small scales modeled by sublfilterscale (SFS) models. Among the existing SFS models, the deconvolution model has been used successfully in the LES of the engineering flows and geophysical flows. Despite the wide application of deconvolution models, the effects of subfilter scale dynamics and filter anisotropy on the accuracy of SFS modeling have not been investigated in depth. The results of LES are highly sensitive to the selection of filters and the anisotropy of the grid, which has been overlooked in previous research. In the current study, two critical aspects of LES are investigated. Firstly, we analyze the influence of sub-filter scale (SFS) dynamics on the accuracy of direct deconvolution models (DDM) at varying filter-to-grid ratios (FGR) in isotropic turbulence. An array of invertible filters are employed, encompassing Gaussian, Helmholtz I and II, Butterworth, Chebyshev I and II, Cauchy, Pao, and rapidly decaying filters. The significance of FGR becomes evident, as it acts as a pivotal factor in error control for precise SFS stress prediction. When FGR is set to 1, the DDM models cannot accurately reconstruct the SFS stress due to the insufficient resolution of SFS dynamics. Notably, prediction capabilities are enhanced at an FGR of 2, resulting in accurate SFS stress reconstruction, except for cases involving Helmholtz I and II filters. A remarkable precision close to 100% is achieved at an FGR of 4 for all DDM models. Additionally, the further exploration extends to the filter anisotropy to address its impact on the SFS dynamics and LES accuracy. By employing dynamic Smagorinsky model (DSM), dynamic mixed model (DMM), and direct deconvolution model (DDM) with the anisotropic filter, aspect ratios (AR) ranging from 1 to 16 in LES filters are evaluated. The findings highlight the DDM's proficiency in accurately predicting SFS stresses under highly anisotropic filtering conditions. High correlation coefficients exceeding 90% are observed in the a priori study for the DDM's reconstructed SFS stresses, surpassing those of the DSM and DMM models. However, these correlations tend to decrease as lter anisotropy increases. In the a posteriori studies, the DDM model consistently outperforms the DSM and DMM models across various turbulence statistics, encompassing velocity spectra, probability density functions related to vorticity, SFS energy flux, velocity increments, strain-rate tensors, and SFS stress. It is observed that as filter anisotropy intensify, the results of DSM and DMM become worse, while the DDM continues to deliver satisfactory results across all filter-anisotropy scenarios. The findings emphasize the DDM framework's potential as a valuable tool for advancing the development of sophisticated SFS models for LES of turbulence.Keywords: deconvolution model, large eddy simulation, subfilter scale modeling, turbulence
Procedia PDF Downloads 769167 Investigation of Corrosion of Steel Buried in Unsaturated Soil in the Presence of Cathodic Protection: The Modified Voltammetry Technique
Authors: Mandlenkosi G. R. Mahlobo, Peter A. Olubambi, Philippe Refait
Abstract:
The aim of this study was to use voltammetry as a method to understand the behaviour of steel in unsaturated soil in the presence of cathodic protection (CP). Three carbon steel coupons were buried in artificial soil wetted at 65-70% of saturation for 37 days. All three coupons were left at open circuit potential (OCP) for the first seven days in the unsaturated soil before CP, which was only applied on two of the three coupons at the protection potential -0.8 V vs Cu/CuSO₄ for the remaining 30 days of the experiment. Voltammetry was performed weekly on the coupon without CP, while electrochemical impedance spectroscopy (EIS) was performed daily to monitor and correct the applied CP potential from the ohmic drop. Voltammetry was finally performed on the last day on the coupons under CP. All the voltammograms were modeled with mathematical equations in order to compute the electrochemical parameters and subsequently deduced the corrosion rate of the steel coupons. For the coupon without CP, the corrosion rate was determined at 300 µm/y. For the coupons under CP, the residual corrosion rate under CP was estimated at 12 µm/y while the corrosion rate of the coupons, after interruption of CP, was estimated at 25 µm/y. This showed that CP was efficient due to two effects: a direct effect from the decreased potential and an induced effect associated with the increased interfacial pH that promoted the formation of a protective layer on the steel surface.Keywords: carbon steel, cathodic protection, voltammetry, unsaturated soil, Raman spectroscopy
Procedia PDF Downloads 639166 Comprehensive Machine Learning-Based Glucose Sensing from Near-Infrared Spectra
Authors: Bitewulign Mekonnen
Abstract:
Context: This scientific paper focuses on the use of near-infrared (NIR) spectroscopy to determine glucose concentration in aqueous solutions accurately and rapidly. The study compares six different machine learning methods for predicting glucose concentration and also explores the development of a deep learning model for classifying NIR spectra. The objective is to optimize the detection model and improve the accuracy of glucose prediction. This research is important because it provides a comprehensive analysis of various machine-learning techniques for estimating aqueous glucose concentrations. Research Aim: The aim of this study is to compare and evaluate different machine-learning methods for predicting glucose concentration from NIR spectra. Additionally, the study aims to develop and assess a deep-learning model for classifying NIR spectra. Methodology: The research methodology involves the use of machine learning and deep learning techniques. Six machine learning regression models, including support vector machine regression, partial least squares regression, extra tree regression, random forest regression, extreme gradient boosting, and principal component analysis-neural network, are employed to predict glucose concentration. The NIR spectra data is randomly divided into train and test sets, and the process is repeated ten times to increase generalization ability. In addition, a convolutional neural network is developed for classifying NIR spectra. Findings: The study reveals that the SVMR, ETR, and PCA-NN models exhibit excellent performance in predicting glucose concentration, with correlation coefficients (R) > 0.99 and determination coefficients (R²)> 0.985. The deep learning model achieves high macro-averaging scores for precision, recall, and F1-measure. These findings demonstrate the effectiveness of machine learning and deep learning methods in optimizing the detection model and improving glucose prediction accuracy. Theoretical Importance: This research contributes to the field by providing a comprehensive analysis of various machine-learning techniques for estimating glucose concentrations from NIR spectra. It also explores the use of deep learning for the classification of indistinguishable NIR spectra. The findings highlight the potential of machine learning and deep learning in enhancing the prediction accuracy of glucose-relevant features. Data Collection and Analysis Procedures: The NIR spectra and corresponding references for glucose concentration are measured in increments of 20 mg/dl. The data is randomly divided into train and test sets, and the models are evaluated using regression analysis and classification metrics. The performance of each model is assessed based on correlation coefficients, determination coefficients, precision, recall, and F1-measure. Question Addressed: The study addresses the question of whether machine learning and deep learning methods can optimize the detection model and improve the accuracy of glucose prediction from NIR spectra. Conclusion: The research demonstrates that machine learning and deep learning methods can effectively predict glucose concentration from NIR spectra. The SVMR, ETR, and PCA-NN models exhibit superior performance, while the deep learning model achieves high classification scores. These findings suggest that machine learning and deep learning techniques can be used to improve the prediction accuracy of glucose-relevant features. Further research is needed to explore their clinical utility in analyzing complex matrices, such as blood glucose levels.Keywords: machine learning, signal processing, near-infrared spectroscopy, support vector machine, neural network
Procedia PDF Downloads 959165 Fire Characteristic of Commercial Retardant Flame Polycarbonate under Different Oxygen Concentration: Ignition Time and Heat Blockage
Authors: Xuelin Zhang, Shouxiang Lu, Changhai Li
Abstract:
The commercial retardant flame polycarbonate samples as the main high speed train interior carriage material with different thicknesses were investigated in Fire Propagation Apparatus with different external heat fluxes under different oxygen concentration from 12% to 40% to study the fire characteristics and quantitatively analyze the ignition time, mass loss rate and heat blockage. The additives of commercial retardant flame polycarbonate were intumescent and maintained a steady height before ignition when heated. The results showed the transformed ignition time (1/t_ig)ⁿ increased linearly with external flux under different oxygen concentration after deducting the heat blockage due to pyrolysis products, the mass loss rate was taken on linearly with external heat fluxes and the slop of the fitting line for mass loss rate and external heat fluxes decreased with the enhanced oxygen concentration and the heat blockage independent on external heat fluxes rose with oxygen concentration increasing. The inquired data as the input of the fire simulation model was the most important to be used to evaluate the fire risk of commercial retardant flame polycarbonate.Keywords: ignition time, mass loss rate, heat blockage, fire characteristic
Procedia PDF Downloads 2839164 A Semantic and Concise Structure to Represent Human Actions
Authors: Tobias Strübing, Fatemeh Ziaeetabar
Abstract:
Humans usually manipulate objects with their hands. To represent these actions in a simple and understandable way, we need to use a semantic framework. For this purpose, the Semantic Event Chain (SEC) method has already been presented which is done by consideration of touching and non-touching relations between manipulated objects in a scene. This method was improved by a computational model, the so-called enriched Semantic Event Chain (eSEC), which incorporates the information of static (e.g. top, bottom) and dynamic spatial relations (e.g. moving apart, getting closer) between objects in an action scene. This leads to a better action prediction as well as the ability to distinguish between more actions. Each eSEC manipulation descriptor is a huge matrix with thirty rows and a massive set of the spatial relations between each pair of manipulated objects. The current eSEC framework has so far only been used in the category of manipulation actions, which eventually involve two hands. Here, we would like to extend this approach to a whole body action descriptor and make a conjoint activity representation structure. For this purpose, we need to do a statistical analysis to modify the current eSEC by summarizing while preserving its features, and introduce a new version called Enhanced eSEC or (e2SEC). This summarization can be done from two points of the view: 1) reducing the number of rows in an eSEC matrix, 2) shrinking the set of possible semantic spatial relations. To achieve these, we computed the importance of each matrix row in an statistical way, to see if it is possible to remove a particular one while all manipulations are still distinguishable from each other. On the other hand, we examined which semantic spatial relations can be merged without compromising the unity of the predefined manipulation actions. Therefore by performing the above analyses, we made the new e2SEC framework which has 20% fewer rows, 16.7% less static spatial and 11.1% less dynamic spatial relations. This simplification, while preserving the salient features of a semantic structure in representing actions, has a tremendous impact on the recognition and prediction of complex actions, as well as the interactions between humans and robots. It also creates a comprehensive platform to integrate with the body limbs descriptors and dramatically increases system performance, especially in complex real time applications such as human-robot interaction prediction.Keywords: enriched semantic event chain, semantic action representation, spatial relations, statistical analysis
Procedia PDF Downloads 1269163 Stress Concentration and Strength Prediction of Carbon/Epoxy Composites
Authors: Emre Ozaslan, Bulent Acar, Mehmet Ali Guler
Abstract:
Unidirectional composites are very popular structural materials used in aerospace, marine, energy and automotive industries thanks to their superior material properties. However, the mechanical behavior of composite materials is more complicated than isotropic materials because of their anisotropic nature. Also, a stress concentration availability on the structure, like a hole, makes the problem further complicated. Therefore, enormous number of tests require to understand the mechanical behavior and strength of composites which contain stress concentration. Accurate finite element analysis and analytical models enable to understand mechanical behavior and predict the strength of composites without enormous number of tests which cost serious time and money. In this study, unidirectional Carbon/Epoxy composite specimens with central circular hole were investigated in terms of stress concentration factor and strength prediction. The composite specimens which had different specimen wide (W) to hole diameter (D) ratio were tested to investigate the effect of hole size on the stress concentration and strength. Also, specimens which had same specimen wide to hole diameter ratio, but varied sizes were tested to investigate the size effect. Finite element analysis was performed to determine stress concentration factor for all specimen configurations. For quasi-isotropic laminate, it was found that the stress concentration factor increased approximately %15 with decreasing of W/D ratio from 6 to 3. Point stress criteria (PSC), inherent flaw method and progressive failure analysis were compared in terms of predicting the strength of specimens. All methods could predict the strength of specimens with maximum %8 error. PSC was better than other methods for high values of W/D ratio, however, inherent flaw method was successful for low values of W/D. Also, it is seen that increasing by 4 times of the W/D ratio rises the failure strength of composite specimen as %62.4. For constant W/D ratio specimens, all the strength prediction methods were more successful for smaller size specimens than larger ones. Increasing the specimen width and hole diameter together by 2 times reduces the specimen failure strength as %13.2.Keywords: failure, strength, stress concentration, unidirectional composites
Procedia PDF Downloads 1569162 Cybernetic Modeling of Growth Dynamics of Debaryomyces nepalensis NCYC 3413 and Xylitol Production in Batch Reactor
Authors: J. Sharon Mano Pappu, Sathyanarayana N. Gummadi
Abstract:
Growth of Debaryomyces nepalensis on mixed substrates in batch culture follows diauxic pattern of completely utilizing glucose during the first exponential growth phase, followed by an intermediate lag phase and a second exponential growth phase consuming xylose. The present study deals with the development of cybernetic mathematical model for prediction of xylitol production and yield. Production of xylitol from xylose in batch fermentation is investigated in the presence of glucose as the co-substrate. Different ratios of glucose and xylose concentrations are assessed to study the impact of multi substrate on production of xylitol in batch reactors. The parameters in the model equations were estimated from experimental observations using integral method. The model equations were solved simultaneously by numerical technique using MATLAB. The developed cybernetic model of xylose fermentation in the presence of a co-substrate can provide answers about how the ratio of glucose to xylose influences the yield and rate of production of xylitol. This model is expected to accurately predict the growth of microorganism on mixed substrate, duration of intermediate lag phase, consumption of substrate, production of xylitol. The model developed based on cybernetic modelling framework can be helpful to simulate the dynamic competition between the metabolic pathways.Keywords: co-substrate, cybernetic model, diauxic growth, xylose, xylitol
Procedia PDF Downloads 3299161 Predicting Stack Overflow Accepted Answers Using Features and Models with Varying Degrees of Complexity
Authors: Osayande Pascal Omondiagbe, Sherlock a Licorish
Abstract:
Stack Overflow is a popular community question and answer portal which is used by practitioners to solve technology-related challenges during software development. Previous studies have shown that this forum is becoming a substitute for official software programming languages documentation. While tools have looked to aid developers by presenting interfaces to explore Stack Overflow, developers often face challenges searching through many possible answers to their questions, and this extends the development time. To this end, researchers have provided ways of predicting acceptable Stack Overflow answers by using various modeling techniques. However, less interest is dedicated to examining the performance and quality of typically used modeling methods, and especially in relation to models’ and features’ complexity. Such insights could be of practical significance to the many practitioners that use Stack Overflow. This study examines the performance and quality of various modeling methods that are used for predicting acceptable answers on Stack Overflow, drawn from 2014, 2015 and 2016. Our findings reveal significant differences in models’ performance and quality given the type of features and complexity of models used. Researchers examining classifiers’ performance and quality and features’ complexity may leverage these findings in selecting suitable techniques when developing prediction models.Keywords: feature selection, modeling and prediction, neural network, random forest, stack overflow
Procedia PDF Downloads 1329160 Artificial Intelligence Models for Detecting Spatiotemporal Crop Water Stress in Automating Irrigation Scheduling: A Review
Authors: Elham Koohi, Silvio Jose Gumiere, Hossein Bonakdari, Saeid Homayouni
Abstract:
Water used in agricultural crops can be managed by irrigation scheduling based on soil moisture levels and plant water stress thresholds. Automated irrigation scheduling limits crop physiological damage and yield reduction. Knowledge of crop water stress monitoring approaches can be effective in optimizing the use of agricultural water. Understanding the physiological mechanisms of crop responding and adapting to water deficit ensures sustainable agricultural management and food supply. This aim could be achieved by analyzing and diagnosing crop characteristics and their interlinkage with the surrounding environment. Assessments of plant functional types (e.g., leaf area and structure, tree height, rate of evapotranspiration, rate of photosynthesis), controlling changes, and irrigated areas mapping. Calculating thresholds of soil water content parameters, crop water use efficiency, and Nitrogen status make irrigation scheduling decisions more accurate by preventing water limitations between irrigations. Combining Remote Sensing (RS), the Internet of Things (IoT), Artificial Intelligence (AI), and Machine Learning Algorithms (MLAs) can improve measurement accuracies and automate irrigation scheduling. This paper is a review structured by surveying about 100 recent research studies to analyze varied approaches in terms of providing high spatial and temporal resolution mapping, sensor-based Variable Rate Application (VRA) mapping, the relation between spectral and thermal reflectance and different features of crop and soil. The other objective is to assess RS indices formed by choosing specific reflectance bands and identifying the correct spectral band to optimize classification techniques and analyze Proximal Optical Sensors (POSs) to control changes. The innovation of this paper can be defined as categorizing evaluation methodologies of precision irrigation (applying the right practice, at the right place, at the right time, with the right quantity) controlled by soil moisture levels and sensitiveness of crops to water stress, into pre-processing, processing (retrieval algorithms), and post-processing parts. Then, the main idea of this research is to analyze the error reasons and/or values in employing different approaches in three proposed parts reported by recent studies. Additionally, as an overview conclusion tried to decompose different approaches to optimizing indices, calibration methods for the sensors, thresholding and prediction models prone to errors, and improvements in classification accuracy for mapping changes.Keywords: agricultural crops, crop water stress detection, irrigation scheduling, precision agriculture, remote sensing
Procedia PDF Downloads 719159 An Criterion to Minimize FE Mesh-Dependency in Concrete Plate Subjected to Impact Loading
Authors: Kwak, Hyo-Gyung, Gang, Han Gul
Abstract:
In the context of an increasing need for reliability and safety in concrete structures under blast and impact loading condition, the behavior of concrete under high strain rate condition has been an important issue. Since concrete subjected to impact loading associated with high strain rate shows quite different material behavior from that in the static state, several material models are proposed and used to describe the high strain rate behavior under blast and impact loading. In the process of modelling, in advance, mesh dependency in the used finite element (FE) is the key problem because simulation results under high strain-rate condition are quite sensitive to applied FE mesh size. It means that the accuracy of simulation results may deeply be dependent on FE mesh size in simulations. This paper introduces an improved criterion which can minimize the mesh-dependency of simulation results on the basis of the fracture energy concept, and HJC (Holmquist Johnson Cook), CSC (Continuous Surface Cap) and K&C (Karagozian & Case) models are examined to trace their relative sensitivity to the used FE mesh size. To coincide with the purpose of the penetration test with a concrete plate under a projectile (bullet), the residual velocities of projectile after penetration are compared. The correlation studies between analytical results and the parametric studies associated with them show that the variation of residual velocity with the used FE mesh size is quite reduced by applying a unique failure strain value determined according to the proposed criterion.Keywords: high strain rate concrete, penetration simulation, failure strain, mesh-dependency, fracture energy
Procedia PDF Downloads 5229158 Uniqueness and Repeatability Analysis for Slim Tube Determined Minimum Miscibility Pressure
Authors: Waqar Ahmad Butt, Gholamreza Vakili Nezhaad, Ali Soud Al Bemani, Yahya Al Wahaibi
Abstract:
Miscible gas injection processes as secondary recovery methods can be applied to a huge number of mature reservoirs to improve the trapped oil displacement. Successful miscible gas injection processes require an accurate estimation of the minimum miscibility pressure (MMP) to make injection process feasible, economical, and effective. There are several methods of MMP determination like slim tube approach, vanishing interfacial tension and rising bubble apparatus but slim tube is the deployed experimental technique in this study. Slim tube method is assumed to be non-standardized for MMP determination with respect to both operating procedure and design. Therefore, 25 slim tube runs were being conducted with three different coil lengths (12, 18 and 24 m) of constant diameter using three different injection rates (0.08, 0.1 and 0.15 cc/min) to evaluate uniqueness and repeatability of determined MMP. A trend of decrease in MMP with increase in coil length was found. No unique trend was found between MMP and injection rate. Lowest MMP and highest recovery were observed with highest coil length and lowest injection rate. It shows that slim tube measured MMP does not depend solely on interacting fluids characteristics but also affected by used coil selection and injection rate choice. Therefore, both slim tube design and procedure need to be standardized. It is recommended to use lowest possible injection rate and estimated coil length depending upon the distance between injections and producing wells for accurate and reliable MMP determination.Keywords: coil length, injection rate, minimum miscibility pressure, multiple contacts miscibility
Procedia PDF Downloads 2549157 Intra-miR-ExploreR, a Novel Bioinformatics Platform for Integrated Discovery of MiRNA:mRNA Gene Regulatory Networks
Authors: Surajit Bhattacharya, Daniel Veltri, Atit A. Patel, Daniel N. Cox
Abstract:
miRNAs have emerged as key post-transcriptional regulators of gene expression, however identification of biologically-relevant target genes for this epigenetic regulatory mechanism remains a significant challenge. To address this knowledge gap, we have developed a novel tool in R, Intra-miR-ExploreR, that facilitates integrated discovery of miRNA targets by incorporating target databases and novel target prediction algorithms, using statistical methods including Pearson and Distance Correlation on microarray data, to arrive at high confidence intragenic miRNA target predictions. We have explored the efficacy of this tool using Drosophila melanogaster as a model organism for bioinformatics analyses and functional validation. A number of putative targets were obtained which were also validated using qRT-PCR analysis. Additional features of the tool include downloadable text files containing GO analysis from DAVID and Pubmed links of literature related to gene sets. Moreover, we are constructing interaction maps of intragenic miRNAs, using both micro array and RNA-seq data, focusing on neural tissues to uncover regulatory codes via which these molecules regulate gene expression to direct cellular development.Keywords: miRNA, miRNA:mRNA target prediction, statistical methods, miRNA:mRNA interaction network
Procedia PDF Downloads 5139156 A Study on Prediction Model for Thermally Grown Oxide Layer in Thermal Barrier Coating
Authors: Yongseok Kim, Jeong-Min Lee, Hyunwoo Song, Junghan Yun, Jungin Byun, Jae-Mean Koo, Chang-Sung Seok
Abstract:
Thermal barrier coating(TBC) is applied for gas turbine components to protect the components from extremely high temperature condition. Since metallic substrate cannot endure such severe condition of gas turbines, delamination of TBC can cause failure of the system. Thus, delamination life of TBC is one of the most important issues for designing the components operating at high temperature condition. Thermal stress caused by thermally grown oxide(TGO) layer is known as one of the major failure mechanisms of TBC. Thermal stress by TGO mainly occurs at the interface between TGO layer and ceramic top coat layer, and it is strongly influenced by the thickness and shape of TGO layer. In this study, Isothermal oxidation is conducted on coin-type TBC specimens prepared by APS(air plasma spray) method. After the isothermal oxidation at various temperature and time condition, the thickness and shape(rumpling shape) of the TGO is investigated, and the test data is processed by numerical analysis. Finally, the test data is arranged into a mathematical prediction model with two variables(temperature and exposure time) which can predict the thickness and rumpling shape of TGO.Keywords: thermal barrier coating, thermally grown oxide, thermal stress, isothermal oxidation, numerical analysis
Procedia PDF Downloads 3429155 Soft Computing Approach for Diagnosis of Lassa Fever
Authors: Roseline Oghogho Osaseri, Osaseri E. I.
Abstract:
Lassa fever is an epidemic hemorrhagic fever caused by the Lassa virus, an extremely virulent arena virus. This highly fatal disorder kills 10% to 50% of its victims, but those who survive its early stages usually recover and acquire immunity to secondary attacks. One of the major challenges in giving proper treatment is lack of fast and accurate diagnosis of the disease due to multiplicity of symptoms associated with the disease which could be similar to other clinical conditions and makes it difficult to diagnose early. This paper proposed an Adaptive Neuro Fuzzy Inference System (ANFIS) for the prediction of Lass Fever. In the design of the diagnostic system, four main attributes were considered as the input parameters and one output parameter for the system. The input parameters are Temperature on admission (TA), White Blood Count (WBC), Proteinuria (P) and Abdominal Pain (AP). Sixty-one percent of the datasets were used in training the system while fifty-nine used in testing. Experimental results from this study gave a reliable and accurate prediction of Lassa fever when compared with clinically confirmed cases. In this study, we have proposed Lassa fever diagnostic system to aid surgeons and medical healthcare practictionals in health care facilities who do not have ready access to Polymerase Chain Reaction (PCR) diagnosis to predict possible Lassa fever infection.Keywords: anfis, lassa fever, medical diagnosis, soft computing
Procedia PDF Downloads 2719154 A Mathematical Model of Pulsatile Blood Flow through a Bifurcated Artery
Authors: D. Srinivasacharya, G. Madhava Rao
Abstract:
In this article, the pulsatile flow of blood flow in bifurcated artery with mild stenosis is investigated. Blood is treated to be a micropolar fluid with constant density. The arteries forming bifurcation are assumed to be symmetric about its axes and straight cylinders of restricted length. As the geometry of the stenosed bifurcated artery is irregular, it is changed to regular geometry utilizing the appropriate transformations. The numerical solutions, using the finite difference method, are computed for the flow rate, the shear stress, and the impedance. The influence of time, coupling number, half of the bifurcated angle and Womersley number on shear stress, flow rate and impedance (resistance to the flow) on both sides of the flow divider is shown graphically. It has been observed that the shear stress and flow rate are increasing with increase in the values of Womersley number and bifurcation angle on both sides of the apex. The shear stress is increasing along the inner wall and decreasing along the outer wall of the daughter artery with an increase in the value of coupling number. Further, it has been noticed that the shear stress, flow rate, and impedance are perturbed largely near to the apex in the parent artery due to the presence of backflow near the apex.Keywords: micropolar fluid, bifurcated artery, stenosis, back flow, secondary flow, pulsatile flow, Womersley number
Procedia PDF Downloads 1939153 Estimation of Pressure Profile and Boundary Layer Characteristics over NACA 4412 Airfoil
Authors: Anwar Ul Haque, Waqar Asrar, Erwin Sulaeman, Jaffar S. M. Ali
Abstract:
Pressure distribution data of the standard airfoils is usually used for the calibration purposes in subsonic wind tunnels. Results of such experiments are quite old and obtained by using the model in the spanwise direction. In this manuscript, pressure distribution over NACA 4412 airfoil model was presented by placing the 3D model in the lateral direction. The model is made of metal with pressure ports distributed longitudinally as well as in the lateral direction. The pressure model was attached to the floor of the tunnel with the help of the base plate to give the specified angle of attack to the model. Before the start of the experiments, the pressure tubes of the respective ports of the 128 ports pressure scanner are checked for leakage, and the losses due to the length of the pipes were also incorporated in the results for the specified pressure range. Growth rate maps of the boundary layer thickness were also plotted. It was found that with the increase in the velocity, the dynamic pressure distribution was also increased for the alpha seep. Plots of pressure distribution so obtained were overlapped with those obtained by using XFLR software, a low fidelity tool. It was found that at moderate and high angles of attack, the distribution of the pressure coefficients obtained from the experiments is high when compared with the XFLR ® results obtained along with the span of the wing. This under-prediction by XFLR ® is more obvious on the windward than on the leeward side.Keywords: subsonic flow, boundary layer, wind tunnel, pressure testing
Procedia PDF Downloads 320