Search results for: quantity estimation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2854

Search results for: quantity estimation

2164 Risk Analysis of Leaks from a Subsea Oil Facility Based on Fuzzy Logic Techniques

Authors: Belén Vinaixa Kinnear, Arturo Hidalgo López, Bernardo Elembo Wilasi, Pablo Fernández Pérez, Cecilia Hernández Fuentealba

Abstract:

The expanded use of risk assessment in legislative and corporate decision-making has increased the role of expert judgement in giving data for security-related decision-making. Expert judgements are required in most steps of risk assessment: danger recognizable proof, hazard estimation, risk evaluation, and examination of choices. This paper presents a fault tree analysis (FTA), which implies a probabilistic failure analysis applied to leakage of oil in a subsea production system. In standard FTA, the failure probabilities of items of a framework are treated as exact values while evaluating the failure probability of the top event. There is continuously insufficiency of data for calculating the failure estimation of components within the drilling industry. Therefore, fuzzy hypothesis can be used as a solution to solve the issue. The aim of this paper is to examine the leaks from the Zafiro West subsea oil facility by using fuzzy fault tree analysis (FFTA). As a result, the research has given theoretical and practical contributions to maritime safety and environmental protection. It has been also an effective strategy used traditionally in identifying hazards in nuclear installations and power industries.

Keywords: expert judgment, probability assessment, fault tree analysis, risk analysis, oil pipelines, subsea production system, drilling, quantitative risk analysis, leakage failure, top event, off-shore industry

Procedia PDF Downloads 190
2163 Fabric Softener Deposition on Cellulose Nanocrystals and Cotton Fibers

Authors: Evdokia K. Oikonomou, Nikolay Christov, Galder Cristobal, Graziana Messina, Giovani Marletta, Laurent Heux, Jean-Francois Berret

Abstract:

Fabric softeners are aqueous formulations that contain ~10 wt. % double tailed cationic surfactants. Here, a formulation in which 50% surfactant was replaced with low quantities of natural guar polymers was developed. Thanks to the reduced surfactant quantity this product has less environmental impact while the guars presence was found to maintain the product’s performance. The objective of this work is to elucidate the effect of the guar polymers on the softener deposition and the adsorption mechanism on the cotton surface. The surfactants in these formulations are assembled into large distributed (0.1 – 1 µm) vesicles that are stable in the presence of guars and upon dilution. The effect of guars on the vesicles adsorption on cotton was first estimated by using cellulose nanocrystals (CNC) as a stand-in for cotton. The dispersion of CNC in water permits to follow the interaction between the vesicles, guars, and CNC in the bulk. It was found that guars enhance the deposition on CNC and that the vesicles are deposited intactly on the fibers driven by electrostatics. The mechanism of the vesicles/guars adsorption on cellulose fibers was identified by quartz crystal microbalance with dissipation monitoring. It was found that the guars increase the surfactant deposited quantity, in agreement with the results in the bulk. Also, the structure of the adsorbed surfactant on the fibers' surfaces (vesicle or bilayer) was influenced by the guars presence. Deposition studies on cotton fabrics were also conducted. Attenuated total reflection and scanning electron microscopy were used to study the effect of the polymers on this deposition. Finally, fluorescent microscopy was used to follow the adsorption of surfactant vesicles, labeled with a fluorescent dye, on cotton fabrics in water. It was found that, in the presence or not of polymers, the surfactant vesicles are adsorbed on fiber maintaining their vesicular structure in water (supported vesicular bilayer structure). The guars influence this process. However, upon drying the vesicles are transformed into bilayers and eventually wrap the fibers (supported lipid bilayer structure). This mechanism is proposed for the adsorption of vesicular conditioner on cotton fiber and can be affected by the presence of polymers.

Keywords: cellulose nanocrystals, cotton fibers, fabric softeners, guar polymers, surfactant vesicles

Procedia PDF Downloads 178
2162 A Combination of Mesenchymal Stem Cells and Low-Intensity Ultrasound for Knee Meniscus Regeneration: A Preliminary Study

Authors: Mohammad Nasb, Muhammad Rehan, Chen Hong

Abstract:

Background Meniscus defects critically alter knee function and lead to degenerative changes. Regenerative medicine applications including stem cell transplantation have showed a promising efficacy in finding alternatives to overcome traditional treatment limitations. However, stem cell therapy remains limited due to the substantially reduced viability and inhibitory microenvironment. Since tissue growth and repair are under the control of biochemical and mechanical signals, several approaches have recently been investigated (e.g., low intensity pulsed ultrasound [LIPUS]) to promote the regeneration process. This study employed LIPUS to improve growth and osteogenic differentiation of mesenchymal stem cells derived from human embryonic stem cells to improve the regeneration of meniscus tissue. Methodology: The Mesenchymal stromal cells (MSCs) were transplanted into the epicenter of the injured meniscus in rabbits, which were randomized into two main groups: a treatment group (n=32 New Zealand rabbits) including 4 subgroups of 8 rabbits in each subgroup (LIPUS treatment, MSC treatment, LIPUS with MSC and control), and a second group (n=9) to track implanted cells and their progeny using green fluorescence protein (GFP). GFP consists of the MSC and LIPUS-MSC combination subgroups. Rabbits were then subjected to histological, immunohistochemistry, and MRI assessment. Results: The quantity of the newly regenerated tissue in the combination treatment group that had Ultrasound irradiation after mesenchymal stem cells were better at all end points. Likewise, Tissue quality scores were also greater in knees treated with both approaches compared with controls and single treatment at all end points, achieving significance at twelve and twenty-four weeks [p < 0.05], and [p = 0.008] at twelve weeks. Differentiation into type-I and II collagen-expressing cells were higher in the combination group at up to twenty-four weeks. Conclusions: the combination of mesenchymal stem cells and LIPUS showed greater adhering to the sites of meniscus injury, differentiate into cells resembling meniscal fibrochondrocytes, and improve both quality and quantity of meniscal regeneration.

Keywords: stem cells, regenerative medicine, osteoarthritis, knee

Procedia PDF Downloads 117
2161 Satellite LiDAR-Based Digital Terrain Model Correction using Gaussian Process Regression

Authors: Keisuke Takahata, Hiroshi Suetsugu

Abstract:

Forest height is an important parameter for forest biomass estimation, and precise elevation data is essential for accurate forest height estimation. There are several globally or nationally available digital elevation models (DEMs) like SRTM and ASTER. However, its accuracy is reported to be low particularly in mountainous areas where there are closed canopy or steep slope. Recently, space-borne LiDAR, such as the Global Ecosystem Dynamics Investigation (GEDI), have started to provide sparse but accurate ground elevation and canopy height estimates. Several studies have reported the high degree of accuracy in their elevation products on their exact footprints, while it is not clear how this sparse information can be used for wider area. In this study, we developed a digital terrain model correction algorithm by spatially interpolating the difference between existing DEMs and GEDI elevation products by using Gaussian Process (GP) regression model. The result shows that our GP-based methodology can reduce the mean bias of the elevation data from 3.7m to 0.3m when we use airborne LiDAR-derived elevation information as ground truth. Our algorithm is also capable of quantifying the elevation data uncertainty, which is critical requirement for biomass inventory. Upcoming satellite-LiDAR missions, like MOLI (Multi-footprint Observation Lidar and Imager), are expected to contribute to the more accurate digital terrain model generation.

Keywords: digital terrain model, satellite LiDAR, gaussian processes, uncertainty quantification

Procedia PDF Downloads 181
2160 Investigation and Estimation of State of Health of Battery Pack in Battery Electric Vehicles-Online Battery Characterization

Authors: Ali Mashayekh, Mahdiye Khorasani, Thomas Weyh

Abstract:

The tendency to use the Battery-Electric vehicle (BEV) for the low and medium driving range or even high driving range has been growing more and more. As a result, higher safety, reliability, and durability of the battery pack as a component of electric vehicles, which has a great share of cost and weight of the final product, are the topics to be considered and investigated. Battery aging can be considered as the predominant factor regarding the reliability and durability of BEV. To better understand the aging process, offline battery characterization has been widely used, which is time-consuming and needs very expensive infrastructures. This paper presents the substitute method for the conventional battery characterization methods, which is based on battery Modular Multilevel Management (BM3). According to this Topology, the battery cells can be drained and charged concerning their capacity, which allows varying battery pack structures. Due to the integration of the power electronics, the output voltage of the battery pack is no longer fixed but can be dynamically adjusted in small steps. In other words, each cell can have three different states, namely series, parallel, and bypass in connection with the neighbor cells. With the help of MATLAB/Simulink and by using the BM3 modules, the battery string model is created. This model allows us to switch two cells with the different SoC as parallel, which results in the internal balancing of the cells. But if the parallel switching lasts just for a couple of ms, we can have a perturbation pulse which can stimulate the cells out of the relaxation phase. With the help of modeling the voltage response pulse of the battery, it would be possible to characterize the cell. The Online EIS method, which is discussed in this paper, can be a robust substitute for the conventional battery characterization methods.

Keywords: battery characterization, SoH estimation, RLS, BEV

Procedia PDF Downloads 147
2159 Price Effect Estimation of Tobacco on Low-wage Male Smokers: A Causal Mediation Analysis

Authors: Kawsar Ahmed, Hong Wang

Abstract:

The study's goal was to estimate the causal mediation impact of tobacco tax before and after price hikes among low-income male smokers, with a particular emphasis on the effect estimating pathways framework for continuous and dichotomous variables. From July to December 2021, a cross-sectional investigation of observational data (n=739) was collected from Bangladeshi low-wage smokers. The Quasi-Bayesian technique, binomial probit model, and sensitivity analysis using a simulation of the computational tools R mediation package had been used to estimate the effect. After a price rise for tobacco products, the average number of cigarettes or bidis sticks taken decreased from 6.7 to 4.56. Tobacco product rising prices have a direct effect on low-income people's decisions to quit or lessen their daily smoking habits of Average Causal Mediation Effect (ACME) [effect=2.31, 95 % confidence interval (C.I.) = (4.71-0.00), p<0.01], Average Direct Effect (ADE) [effect=8.6, 95 percent (C.I.) = (6.8-0.11), p<0.001], and overall significant effects (p<0.001). Tobacco smoking choice is described by the mediated proportion of income effect, which is 26.1% less of following price rise. The curve of ACME and ADE is based on observational figures of the coefficients of determination that asses the model of hypothesis as the substantial consequence after price rises in the sensitivity analysis. To reduce smoking product behaviors, price increases through taxation have a positive causal mediation with income that affects the decision to limit tobacco use and promote low-income men's healthcare policy.

Keywords: causal mediation analysis, directed acyclic graphs, tobacco price policy, sensitivity analysis, pathway estimation

Procedia PDF Downloads 111
2158 A Two-Stage Bayesian Variable Selection Method with the Extension of Lasso for Geo-Referenced Data

Authors: Georgiana Onicescu, Yuqian Shen

Abstract:

Due to the complex nature of geo-referenced data, multicollinearity of the risk factors in public health spatial studies is a commonly encountered issue, which leads to low parameter estimation accuracy because it inflates the variance in the regression analysis. To address this issue, we proposed a two-stage variable selection method by extending the least absolute shrinkage and selection operator (Lasso) to the Bayesian spatial setting, investigating the impact of risk factors to health outcomes. Specifically, in stage I, we performed the variable selection using Bayesian Lasso and several other variable selection approaches. Then, in stage II, we performed the model selection with only the selected variables from stage I and compared again the methods. To evaluate the performance of the two-stage variable selection methods, we conducted a simulation study with different distributions for the risk factors, using geo-referenced count data as the outcome and Michigan as the research region. We considered the cases when all candidate risk factors are independently normally distributed, or follow a multivariate normal distribution with different correlation levels. Two other Bayesian variable selection methods, Binary indicator, and the combination of Binary indicator and Lasso were considered and compared as alternative methods. The simulation results indicated that the proposed two-stage Bayesian Lasso variable selection method has the best performance for both independent and dependent cases considered. When compared with the one-stage approach, and the other two alternative methods, the two-stage Bayesian Lasso approach provides the highest estimation accuracy in all scenarios considered.

Keywords: Lasso, Bayesian analysis, spatial analysis, variable selection

Procedia PDF Downloads 142
2157 Comparing Xbar Charts: Conventional versus Reweighted Robust Estimation Methods for Univariate Data Sets

Authors: Ece Cigdem Mutlu, Burak Alakent

Abstract:

Maintaining the quality of manufactured products at a desired level depends on the stability of process dispersion and location parameters and detection of perturbations in these parameters as promptly as possible. Shewhart control chart is the most widely used technique in statistical process monitoring to monitor the quality of products and control process mean and variability. In the application of Xbar control charts, sample standard deviation and sample mean are known to be the most efficient conventional estimators in determining process dispersion and location parameters, respectively, based on the assumption of independent and normally distributed datasets. On the other hand, there is no guarantee that the real-world data would be normally distributed. In the cases of estimated process parameters from Phase I data clouded with outliers, efficiency of traditional estimators is significantly reduced, and performance of Xbar charts are undesirably low, e.g. occasional outliers in the rational subgroups in Phase I data set may considerably affect the sample mean and standard deviation, resulting a serious delay in detection of inferior products in Phase II. For more efficient application of control charts, it is required to use robust estimators against contaminations, which may exist in Phase I. In the current study, we present a simple approach to construct robust Xbar control charts using average distance to the median, Qn-estimator of scale, M-estimator of scale with logistic psi-function in the estimation of process dispersion parameter, and Harrell-Davis qth quantile estimator, Hodge-Lehmann estimator and M-estimator of location with Huber psi-function and logistic psi-function in the estimation of process location parameter. Phase I efficiency of proposed estimators and Phase II performance of Xbar charts constructed from these estimators are compared with the conventional mean and standard deviation statistics both under normality and against diffuse-localized and symmetric-asymmetric contaminations using 50,000 Monte Carlo simulations on MATLAB. Consequently, it is found that robust estimators yield parameter estimates with higher efficiency against all types of contaminations, and Xbar charts constructed using robust estimators have higher power in detecting disturbances, compared to conventional methods. Additionally, utilizing individuals charts to screen outlier subgroups and employing different combination of dispersion and location estimators on subgroups and individual observations are found to improve the performance of Xbar charts.

Keywords: average run length, M-estimators, quality control, robust estimators

Procedia PDF Downloads 189
2156 Wind Resource Estimation and Economic Analysis for Rakiraki, Fiji

Authors: Kaushal Kishore

Abstract:

Immense amount of imported fuels are used in Fiji for electricity generation, transportation and for carrying out miscellaneous household work. To alleviate its dependency on fossil fuel, paramount importance has been given to instigate the utilization of renewable energy sources for power generation and to reduce the environmental dilapidation. Amongst the many renewable energy sources, wind has been considered as one of the best identified renewable sources that are comprehensively available in Fiji. In this study the wind resource assessment for three locations in Rakiraki, Fiji has been carried out. The wind resource estimation at Rokavukavu, Navolau and at Tuvavatu has been analyzed. The average wind speed at 55 m above ground level (a.g.l) at Rokavukavu, Navolau, and Tuvavatu sites are 5.91 m/s, 8.94 m/s and 8.13 m/s with the turbulence intensity of 14.9%, 17.1%, and 11.7% respectively. The moment fitting method has been used to estimate the Weibull parameter and the power density at each sites. A high resolution wind resource map for the three locations has been developed by using Wind Atlas Analysis and Application Program (WAsP). The results obtained from WAsP exhibited good wind potential at Navolau and Tuvavatu sites. A wind farm has been proposed at Navolau and Tuvavatu site that comprises six Vergnet 275 kW wind turbines at each site. The annual energy production (AEP) for each wind farm is estimated and an economic analysis is performed. The economic analysis for the proposed wind farms at Navolau and Tuvavatu sites showed a payback period of 5 and 6 years respectively.

Keywords: annual energy production, Rakiraki Fiji, turbulence intensity, Weibull parameter, wind speed, Wind Atlas Analysis and Application Program

Procedia PDF Downloads 188
2155 Assessment of the Thermal Performance of a Solar Heating System on an Agricultural Greenhouse Microclimate

Authors: Nora Arbaoui, Rachid Tadili

Abstract:

The substantial increase of areas cultivated under glasshouses compels the use of other natural heating and cooling procedures to make a profit as well as avoid both exorbitant fuel consumption and CO₂ emissions. This experimental study is designed to examine the functioning of a solar heating system that will increase positive consequences in terms of both quantity and quality while successfully enhancing greenhouse microclimate during wintertime. Those configurations have been tested in a miniaturized greenhouse simply after having optimized the operating parameters. These were noteworthy results when compared to an unheated witness greenhouse.

Keywords: solar system, agricultural greenhouse, heating, cooling, storage, drying

Procedia PDF Downloads 24
2154 Tracing Sources of Sediment in an Arid River, Southern Iran

Authors: Hesam Gholami

Abstract:

Elevated suspended sediment loads in riverine systems resulting from accelerated erosion due to human activities are a serious threat to the sustainable management of watersheds and ecosystem services therein worldwide. Therefore, mitigation of deleterious sediment effects as a distributed or non-point pollution source in the catchments requires reliable provenance information. Sediment tracing or sediment fingerprinting, as a combined process consisting of sampling, laboratory measurements, different statistical tests, and the application of mixing or unmixing models, is a useful technique for discriminating the sources of sediments. From 1996 to the present, different aspects of this technique, such as grouping the sources (spatial and individual sources), discriminating the potential sources by different statistical techniques, and modification of mixing and unmixing models, have been introduced and modified by many researchers worldwide, and have been applied to identify the provenance of fine materials in agricultural, rural, mountainous, and coastal catchments, and in large catchments with numerous lakes and reservoirs. In the last two decades, efforts exploring the uncertainties associated with sediment fingerprinting results have attracted increasing attention. The frameworks used to quantify the uncertainty associated with fingerprinting estimates can be divided into three groups comprising Monte Carlo simulation, Bayesian approaches and generalized likelihood uncertainty estimation (GLUE). Given the above background, the primary goal of this study was to apply geochemical fingerprinting within the GLUE framework in the estimation of sub-basin spatial sediment source contributions in the arid Mehran River catchment in southern Iran, which drains into the Persian Gulf. The accuracy of GLUE predictions generated using four different sets of statistical tests for discriminating three sub-basin spatial sources was evaluated using 10 virtual sediments (VS) samples with known source contributions using the root mean square error (RMSE) and mean absolute error (MAE). Based on the results, the contributions modeled by GLUE for the western, central and eastern sub-basins are 1-42% (overall mean 20%), 0.5-30% (overall mean 12%) and 55-84% (overall mean 68%), respectively. According to the mean absolute fit (MAF; ≥ 95% for all target sediment samples) and goodness-of-fit (GOF; ≥ 99% for all samples), our suggested modeling approach is an accurate technique to quantify the source of sediments in the catchments. Overall, the estimated source proportions can help watershed engineers plan the targeting of conservation programs for soil and water resources.

Keywords: sediment source tracing, generalized likelihood uncertainty estimation, virtual sediment mixtures, Iran

Procedia PDF Downloads 73
2153 Achieving Design-Stage Elemental Cost Planning Accuracy: Case Study of New Zealand

Authors: Johnson Adafin, James O. B. Rotimi, Suzanne Wilkinson, Abimbola O. Windapo

Abstract:

An aspect of client expenditure management that requires attention is the level of accuracy achievable in design-stage elemental cost planning. This has been a major concern for construction clients and practitioners in New Zealand (NZ). Pre-tender estimating inaccuracies are significantly influenced by the level of risk information available to estimators. Proper cost planning activities should ensure the production of a project’s likely construction costs (initial and final), and subsequent cost control activities should prevent unpleasant consequences of cost overruns, disputes and project abandonment. If risks were properly identified and priced at the design stage, observed variance between design-stage elemental cost plans (ECPs) and final tender sums (FTS) (initial contract sums) could be reduced. This study investigates the variations between design-stage ECPs and FTS of construction projects, with a view to identifying risk factors that are responsible for the observed variance. Data were sourced through interviews, and risk factors were identified by using thematic analysis. Access was obtained to project files from the records of study participants (consultant quantity surveyors), and document analysis was employed in complementing the responses from the interviews. Study findings revealed the discrepancies between ECPs and FTS in the region of -14% and +16%. It is opined in this study that the identified risk factors were responsible for the variability observed. The values obtained from the analysis would enable greater accuracy in the forecast of FTS by Quantity Surveyors. Further, whilst inherent risks in construction project developments are observed globally, these findings have important ramifications for construction projects by expanding existing knowledge on what is needed for reasonable budgetary performance and successful delivery of construction projects. The findings contribute significantly to the study by providing quantitative confirmation to justify the theoretical conclusions generated in the literature from around the world. This therefore adds to and consolidates existing knowledge.

Keywords: accuracy, design-stage, elemental cost plan, final tender sum

Procedia PDF Downloads 268
2152 South African Breast Cancer Mutation Spectrum: Pitfalls to Copy Number Variation Detection Using Internationally Designed Multiplex Ligation-Dependent Probe Amplification and Next Generation Sequencing Panels

Authors: Jaco Oosthuizen, Nerina C. Van Der Merwe

Abstract:

The National Health Laboratory Services in Bloemfontien has been the diagnostic testing facility for 1830 patients for familial breast cancer since 1997. From the cohort, 540 were comprehensively screened using High-Resolution Melting Analysis or Next Generation Sequencing for the presence of point mutations and/or indels. Approximately 90% of these patients stil remain undiagnosed as they are BRCA1/2 negative. Multiplex ligation-dependent probe amplification was initially added to screen for copy number variation detection, but with the introduction of next generation sequencing in 2017, was substituted and is currently used as a confirmation assay. The aim was to investigate the viability of utilizing internationally designed copy number variation detection assays based on mostly European/Caucasian genomic data for use within a South African context. The multiplex ligation-dependent probe amplification technique is based on the hybridization and subsequent ligation of multiple probes to a targeted exon. The ligated probes are amplified using conventional polymerase chain reaction, followed by fragment analysis by means of capillary electrophoresis. The experimental design of the assay was performed according to the guidelines of MRC-Holland. For BRCA1 (P002-D1) and BRCA2 (P045-B3), both multiplex assays were validated, and results were confirmed using a secondary probe set for each gene. The next generation sequencing technique is based on target amplification via multiplex polymerase chain reaction, where after the amplicons are sequenced parallel on a semiconductor chip. Amplified read counts are visualized as relative copy numbers to determine the median of the absolute values of all pairwise differences. Various experimental parameters such as DNA quality, quantity, and signal intensity or read depth were verified using positive and negative patients previously tested internationally. DNA quality and quantity proved to be the critical factors during the verification of both assays. The quantity influenced the relative copy number frequency directly whereas the quality of the DNA and its salt concentration influenced denaturation consistency in both assays. Multiplex ligation-dependent probe amplification produced false positives due to ligation failure when ligation was inhibited due to a variant present within the ligation site. Next generation sequencing produced false positives due to read dropout when primer sequences did not meet optimal multiplex binding kinetics due to population variants in the primer binding site. The analytical sensitivity and specificity for the South African population have been proven. Verification resulted in repeatable reactions with regards to the detection of relative copy number differences. Both multiplex ligation-dependent probe amplification and next generation sequencing multiplex panels need to be optimized to accommodate South African polymorphisms present within the genetically diverse ethnic groups to reduce the false copy number variation positive rate and increase performance efficiency.

Keywords: familial breast cancer, multiplex ligation-dependent probe amplification, next generation sequencing, South Africa

Procedia PDF Downloads 230
2151 Estimation of Constant Coefficients of Bourgoyne and Young Drilling Rate Model for Drill Bit Wear Prediction

Authors: Ahmed Z. Mazen, Nejat Rahmanian, Iqbal Mujtaba, Ali Hassanpour

Abstract:

In oil and gas well drilling, the drill bit is an important part of the Bottom Hole Assembly (BHA), which is installed and designed to drill and produce a hole by several mechanisms. The efficiency of the bit depends on many drilling parameters such as weight on bit, rotary speed, and mud properties. When the bit is pulled out of the hole, the evaluation of the bit damage must be recorded very carefully to guide engineers in order to select the bits for further planned wells. Having a worn bit for hole drilling may cause severe damage to bit leading to cutter or cone losses in the bottom of hole, where a fishing job will have to take place, and all of these will increase the operating cost. The main factor to reduce the cost of drilling operation is to maximize the rate of penetration by analyzing real-time data to predict the drill bit wear while drilling. There are numerous models in the literature for prediction of the rate of penetration based on drilling parameters, mostly based on empirical approaches. One of the most commonly used approaches is Bourgoyne and Young model, where the rate of penetration can be estimated by the drilling parameters as well as a wear index using an empirical correlation, provided all the constants and coefficients are accurately determined. This paper introduces a new methodology to estimate the eight coefficients for Bourgoyne and Young model using the gPROMS parameters estimation GPE (Version 4.2.0). Real data collected form similar formations (12 ¼’ sections) in two different fields in Libya are used to estimate the coefficients. The estimated coefficients are then used in the equations and applied to nearby wells in the same field to predict the bit wear.

Keywords: Bourgoyne and Young model, bit wear, gPROMS, rate of penetration

Procedia PDF Downloads 152
2150 Estimation and Removal of Chlorophenolic Compounds from Paper Mill Waste Water by Electrochemical Treatment

Authors: R. Sharma, S. Kumar, C. Sharma

Abstract:

A number of toxic chlorophenolic compounds are formed during pulp bleaching. The nature and concentration of these chlorophenolic compounds largely depends upon the amount and nature of bleaching chemicals used. These compounds are highly recalcitrant and difficult to remove but are partially removed by the biochemical treatment processes adopted by the paper industry. Identification and estimation of these chlorophenolic compounds has been carried out in the primary and secondary clarified effluents from the paper mill by GCMS. Twenty-six chorophenolic compounds have been identified and estimated in paper mill waste waters. Electrochemical treatment is an efficient method for oxidation of pollutants and has successfully been used to treat textile and oil waste water. Electrochemical treatment using less expensive anode material, stainless steel electrodes has been tried to study their removal. The electrochemical assembly comprised a DC power supply, a magnetic stirrer and stainless steel (316 L) electrode. The optimization of operating conditions has been carried out and treatment has been performed under optimized treatment conditions. Results indicate that 68.7% and 83.8% of cholorphenolic compounds are removed during 2 h of electrochemical treatment from primary and secondary clarified effluent respectively. Further, there is a reduction of 65.1, 60 and 92.6% of COD, AOX and color, respectively for primary clarified and 83.8%, 75.9% and 96.8% of COD, AOX and color, respectively for secondary clarified effluent. EC treatment has also been found to increase significantly the biodegradability index of wastewater because of conversion of non- biodegradable fraction into biodegradable fraction. Thus, electrochemical treatment is an efficient method for the degradation of cholorophenolic compounds, removal of color, AOX and other recalcitrant organic matter present in paper mill waste water.

Keywords: chlorophenolics, effluent, electrochemical treatment, wastewater

Procedia PDF Downloads 385
2149 Government Size and Economic Growth: Testing the Non-Linear Hypothesis for Nigeria

Authors: R. Santos Alimi

Abstract:

Using time-series techniques, this study empirically tested the validity of existing theory which stipulates there is a nonlinear relationship between government size and economic growth; such that government spending is growth-enhancing at low levels but growth-retarding at high levels, with the optimal size occurring somewhere in between. This study employed three estimation equations. First, for the size of government, two measures are considered as follows: (i) share of total expenditures to gross domestic product, (ii) share of recurrent expenditures to gross domestic product. Second, the study adopted real GDP (without government expenditure component), as a variant measure of economic growth other than the real total GDP, in estimating the optimal level of government expenditure. The study is based on annual Nigeria country-level data for the period 1970 to 2012. Estimation results show that the inverted U-shaped curve exists for the two measures of government size and the estimated optimum shares are 19.81% and 10.98%, respectively. Finally, with the adoption of real GDP (without government expenditure component), the optimum government size was found to be 12.58% of GDP. Our analysis shows that the actual share of government spending on average (2000 - 2012) is about 13.4%.This study adds to the literature confirming that the optimal government size exists not only for developed economies but also for developing economy like Nigeria. Thus, a public intervention threshold level that fosters economic growth is a reality; beyond this point economic growth should be left in the hands of the private sector. This finding has a significant implication for the appraisal of government spending and budgetary policy design.

Keywords: public expenditure, economic growth, optimum level, fully modified OLS

Procedia PDF Downloads 420
2148 Identification of Accumulated Hydrocarbon Based on Heat Propagation Analysis in Order to Develop Mature Field: Case Study in South Sumatra Basin, Indonesia

Authors: Kukuh Suprayogi, Muhamad Natsir, Olif Kurniawan, Hot Parulian, Bayu Fitriana, Fery Mustofa

Abstract:

The new approach by utilizing the heat propagation analysis carried out by studying and evaluating the effect of the presence of hydrocarbons to the flow of heat that goes from the bottom surface to surface. Heat propagation is determined by the thermal conductivity of rocks. The thermal conductivity of rock itself is a quantity that describes the ability of a rock to deliver heat. This quantity depends on the constituent rock lithology, large porosity, and pore fluid filler. The higher the thermal conductivity of a rock, the more easily the flow of heat passing through these rocks. With the same sense, the heat flow will more easily pass through the rock when the rock is filled with water than hydrocarbons, given the nature of the hydrocarbons having more insulator against heat. The main objective of this research is to try to make the model the heat propagation calculations in degrees Celsius from the subsurface to the surface which is then compared with the surface temperature is measured directly at the point of location. In calculating the propagation of heat, we need to first determine the thermal conductivity of rocks, where the rocks at the point calculation are not composed of homogeneous but consist of strata. Therefore, we need to determine the mineral constituent and porosity values of each stratum. As for the parameters of pore fluid filler, we assume that all the pores filled with water. Once we get a thermal conductivity value of each unit of the rock, then we begin to model the propagation of heat profile from the bottom to the surface. The initial value of the temperature that we use comes from the data bottom hole temperature (BHT) is obtained from drilling results. Results of calculations per depths the temperature is displayed in plotting temperature versus depth profiles that describe the propagation of heat from the bottom of the well to the surface, note that pore fluid is water. In the technical implementation, we can identify the magnitude of the effect of hydrocarbons in reducing the amount of heat that crept to the surface based on the calculation of propagation of heat at a certain point and compared with measurements of surface temperature at that point, assuming that the surface temperature measured is the temperature that comes from the asthenosphere. This publication proves that the accumulation of hydrocarbon can be identified by analysis of heat propagation profile which could be a method for identifying the presence of hydrocarbons.

Keywords: thermal conductivity, rock, pore fluid, heat propagation

Procedia PDF Downloads 107
2147 The Application of Computer and Technology in Language Teaching and Learning

Authors: Pouya Vakili

Abstract:

Since computers were first introduced into educational facilities, foreign language educators have been faced with the problem of integrating high-tech multimedia techniques into a traditional text-based curriculum. As studies of language teaching have pointed out, ‘Language teaching tends in practice to be eclectic…. There are not only exceptionally many paths and educational means for arriving at a given educational goal, but there are also many types of educational materials which can be used to achieve that goal’. For language educators who are trying to incorporate technology into their curricula, the choices seem endless. Yet the quantity, as well as the limitations, of available computer programs does not guarantee that these programs can be successfully integrated into a curriculum.

Keywords: curriculum, language teaching, learning, multimedia, technology

Procedia PDF Downloads 572
2146 Estimation of Endogenous Brain Noise from Brain Response to Flickering Visual Stimulation Magnetoencephalography Visual Perception Speed

Authors: Alexander N. Pisarchik, Parth Chholak

Abstract:

Intrinsic brain noise was estimated via magneto-encephalograms (MEG) recorded during perception of flickering visual stimuli with frequencies of 6.67 and 8.57 Hz. First, we measured the mean phase difference between the flicker signal and steady-state event-related field (SSERF) in the occipital area where the brain response at the flicker frequencies and their harmonics appeared in the power spectrum. Then, we calculated the probability distribution of the phase fluctuations in the regions of frequency locking and computed its kurtosis. Since kurtosis is a measure of the distribution’s sharpness, we suppose that inverse kurtosis is related to intrinsic brain noise. In our experiments, the kurtosis value varied among subjects from K = 3 to K = 5 for 6.67 Hz and from 2.6 to 4 for 8.57 Hz. The majority of subjects demonstrated leptokurtic kurtosis (K < 3), i.e., the distribution tails approached zero more slowly than Gaussian. In addition, we found a strong correlation between kurtosis and brain complexity measured as the correlation dimension, so that the MEGs of subjects with higher kurtosis exhibited lower complexity. The obtained results are discussed in the framework of nonlinear dynamics and complex network theories. Specifically, in a network of coupled oscillators, phase synchronization is mainly determined by two antagonistic factors, noise, and the coupling strength. While noise worsens phase synchronization, the coupling improves it. If we assume that each neuron and each synapse contribute to brain noise, the larger neuronal network should have stronger noise, and therefore phase synchronization should be worse, that results in smaller kurtosis. The described method for brain noise estimation can be useful for diagnostics of some brain pathologies associated with abnormal brain noise.

Keywords: brain, flickering, magnetoencephalography, MEG, visual perception, perception time

Procedia PDF Downloads 147
2145 Statistical Assessment of Models for Determination of Soil–Water Characteristic Curves of Sand Soils

Authors: S. J. Matlan, M. Mukhlisin, M. R. Taha

Abstract:

Characterization of the engineering behavior of unsaturated soil is dependent on the soil-water characteristic curve (SWCC), a graphical representation of the relationship between water content or degree of saturation and soil suction. A reasonable description of the SWCC is thus important for the accurate prediction of unsaturated soil parameters. The measurement procedures for determining the SWCC, however, are difficult, expensive, and time-consuming. During the past few decades, researchers have laid a major focus on developing empirical equations for predicting the SWCC, with a large number of empirical models suggested. One of the most crucial questions is how precisely existing equations can represent the SWCC. As different models have different ranges of capability, it is essential to evaluate the precision of the SWCC models used for each particular soil type for better SWCC estimation. It is expected that better estimation of SWCC would be achieved via a thorough statistical analysis of its distribution within a particular soil class. With this in view, a statistical analysis was conducted in order to evaluate the reliability of the SWCC prediction models against laboratory measurement. Optimization techniques were used to obtain the best-fit of the model parameters in four forms of SWCC equation, using laboratory data for relatively coarse-textured (i.e., sandy) soil. The four most prominent SWCCs were evaluated and computed for each sample. The result shows that the Brooks and Corey model is the most consistent in describing the SWCC for sand soil type. The Brooks and Corey model prediction also exhibit compatibility with samples ranging from low to high soil water content in which subjected to the samples that evaluated in this study.

Keywords: soil-water characteristic curve (SWCC), statistical analysis, unsaturated soil, geotechnical engineering

Procedia PDF Downloads 336
2144 Impact of Instrument Transformer Secondary Connections on Performance of Protection System: Experiences from Indian POWERGRID

Authors: Pankaj Kumar Jha, Mahendra Singh Hada, Brijendra Singh, Sandeep Yadav

Abstract:

Protective relays are commonly connected to the secondary windings of instrument transformers, i.e., current transformers (CTs) and/or capacitive voltage transformers (CVTs). The purpose of CT and CVT is to provide galvanic isolation from high voltages and reduce primary currents and voltages to a nominal quantity recognized by the protective relays. Selecting the correct instrument transformers for an application is imperative: failing to do so may compromise the relay’s performance, as the output of the instrument transformer may no longer be an accurately scaled representation of the primary quantity. Having an accurately rated instrument transformer is of no use if these devices are not properly connected. The performance of the protective relay is reliant on its programmed settings and on the current and voltage inputs from the instrument transformers secondary. This paper will help in understanding the fundamental concepts of the connections of Instrument Transformers to the protection relays and the effect of incorrect connection on the performance of protective relays. Multiple case studies of protection system mal-operations due to incorrect connections of instrument transformers will be discussed in detail in this paper. Apart from the connection issue of instrument transformers to protective relays, this paper will also discuss the effect of multiple earthing of CTs and CVTs secondary on the performance of the protection system. Case studies presented in this paper will help the readers to analyse the problem through real-world challenges in complex power system networks. This paper will also help the protection engineer in better analysis of disturbance records. CT and CVT connection errors can lead to undesired operations of protection systems. However, many of these operations can be avoided by adhering to industry standards and implementing tried-and-true field testing and commissioning practices. Understanding the effect of missing neutral of CVT, multiple earthing of CVT secondary, and multiple grounding of CT star points on the performance of the protection system through real-world case studies will help the protection engineer in better commissioning the protection system and maintenance of the protection system.

Keywords: bus reactor, current transformer, capacitive voltage transformer, distance protection, differential protection, directional earth fault, disturbance report, instrument transformer, ICT, REF protection, shunt reactor, voltage selection relay, VT fuse failure

Procedia PDF Downloads 81
2143 Don't Just Guess and Slip: Estimating Bayesian Knowledge Tracing Parameters When Observations Are Scant

Authors: Michael Smalenberger

Abstract:

Intelligent tutoring systems (ITS) are computer-based platforms which can incorporate artificial intelligence to provide step-by-step guidance as students practice problem-solving skills. ITS can replicate and even exceed some benefits of one-on-one tutoring, foster transactivity in collaborative environments, and lead to substantial learning gains when used to supplement the instruction of a teacher or when used as the sole method of instruction. A common facet of many ITS is their use of Bayesian Knowledge Tracing (BKT) to estimate parameters necessary for the implementation of the artificial intelligence component, and for the probability of mastery of a knowledge component relevant to the ITS. While various techniques exist to estimate these parameters and probability of mastery, none directly and reliably ask the user to self-assess these. In this study, 111 undergraduate students used an ITS in a college-level introductory statistics course for which detailed transaction-level observations were recorded, and users were also routinely asked direct questions that would lead to such a self-assessment. Comparisons were made between these self-assessed values and those obtained using commonly used estimation techniques. Our findings show that such self-assessments are particularly relevant at the early stages of ITS usage while transaction level data are scant. Once a user’s transaction level data become available after sufficient ITS usage, these can replace the self-assessments in order to eliminate the identifiability problem in BKT. We discuss how these findings are relevant to the number of exercises necessary to lead to mastery of a knowledge component, the associated implications on learning curves, and its relevance to instruction time.

Keywords: Bayesian Knowledge Tracing, Intelligent Tutoring System, in vivo study, parameter estimation

Procedia PDF Downloads 171
2142 Enhancing the Quality of Silage Bales Produced by a Commercial Scale Silage Producer in Northern province, Sri Lanka: A Step Toward Supporting Smallholder Dairy Farmers in the Northern Province Sri Lanka

Authors: Harithas Aruchchunan

Abstract:

Silage production is an essential aspect of dairy farming, used to provide high-quality feed to ruminants. However, dairy farmers in Northern Province Sri Lanka are facing multiple challenges that compromise the quality and quantity of silage produced. To tackle these challenges, promoting silage feeding has become an essential component of sustainable dairy farming practices. In this study, silage bale samples were collected from a newly started silage baling factory in Jaffna, Northern province and their quality was analysed at the Veterinary Research Institute laboratory in Kandy in March 2023. The results show the nutritional composition of three Napier grass cultivars: Super Napier, CO6, and Indian Red Napier (BH18). The main parameters analysed were dry matter, pH, lactic acid, soluble carbohydrate, ammonia nitrogen, ash, crude protein, NDF, and ADF. The results indicate that Super Napier and CO6 have higher crude protein content and lower ADF levels, making them suitable for producing high-quality silage. The pH levels of all three cultivars were safe, and the ammonia nitrogen levels were considered appropriate. However, laboratory results indicate that the quality of silage bales produced can be further enhanced. Dairy farmers should be encouraged to adopt these cultivars to achieve better yields as they are high in protein and are better suited to Northern Province's soil and climate. Therefore, it is vital to educate small-scale fodder producers, who supply the raw material to silage factories, on the best practices of cultivating these new cultivars. To improve silage bale production and quality in Northern Province Sri Lanka, we recommend increasing public awareness about silage feeding, providing education and training to dairy farmers and small-scale fodder producers on modern silage production techniques and improving the availability of raw materials for silage production. Additionally, Napier grass cultivars need to be promoted among dairy farmers for better production and quality of silage bales. Failing to improve the quality and quantity of silage bale production could not only lead to the decline of dairy farming in Northern Province Sri Lanka but also the negative impact on the economy

Keywords: silage bales, dairy farming, economic crisis, Sri Lanka

Procedia PDF Downloads 90
2141 Contribution to the Study of the Fungal Flora Seed-Borne in Cereal: Wheat and Barley

Authors: M’lik Randa, Lakhdari Wassima, Dahliz Abderrahmène, Soud Adila, Hammi Hamida

Abstract:

In cereal culture, as in the most the vegetal productions the seeds play an important role in the development of the future plant. The healthy seeds are very important for the quality and quantity production. This study on a media (P.D.A) shows that an important mycoflora exists in the crops. Among the identified fungical, we notice the presence of Helminthosporium sp, Alternaria sp, Botrytis and Macrosporium. The use of the illness causing facies, especially for Helminthosporium, Alternaria and Botrytis emphasizes the relation between the seminicole inoculums and the appearance of symptoms on young plants noted by authors.

Keywords: seeds, barley, wheat, fungical flora

Procedia PDF Downloads 414
2140 DFT Theoretical Investigation for Evaluating Global Scalar Properties and Validating with Quantum Chemical Based COSMO-RS Theory for Dissolution of Bituminous and Anthracite Coal in Ionic Liquid

Authors: Debanjan Dey, Tamal Banerjee, Kaustubha Mohanty

Abstract:

Global scalar properties are calculated based on higher occupied molecular orbital (HOMO) and lower unoccupied molecular orbital (LUMO) energy to study the interaction between ionic liquids with Bituminous and Anthracite coal using density function theory (DFT) method. B3LYP/6-31G* calculation predicts HOMO-LUMO energy gap, electronegativity, global hardness, global softness, chemical potential and global softness for individual compounds with their clusters. HOMO-LUMO interaction, electron delocalization, electron donating and accepting is the main source of attraction between individual compounds with their complexes. Cation used in this study: 1-butyl-1-methylpyrrolidinium [BMPYR], 1-methyl -3-propylimmidazolium [MPIM], Tributylmethylammonium [TMA] and Tributylmethylphosphonium [MTBP] with the combination of anion: bis(trifluromethylsulfonyl)imide [Tf2N], methyl carbonate [CH3CO3], dicyanamide [N(CN)2] and methylsulfate [MESO4]. Basically three-tier approach comprising HOMO/LUMO energy, Scalar quantity and infinite dilution activity coefficient (IDAC) by sigma profile generation with COSMO-RS (Conductor like screening model for real solvent) model was chosen for simultaneous interaction. [BMPYR]CH3CO3] (1-butyl-1-methylpyrrolidinium methyl carbonate) and [MPIM][CH3CO3] (1-methyl -3-propylimmidazolium methyl carbonate ) are the best effective ILs on the basis of HOMO-LUMO band gap for Anthracite and Bituminous coal respectively and the corresponding band gap is 0.10137 hartree for Anthracite coal and 0.12485 hartree for Bituminous coal. Further ionic liquids are screened quantitatively with all the scalar parameters and got the same result based on CH-π interaction which is found for HOMO-LUMO gap. To check our findings IDAC were predicted using quantum chemical based COSMO-RS methodology which gave the same trend as observed our scalar quantity calculation. Thereafter a qualitative measurement is doing by sigma profile analysis which gives complementary behavior between IL and coal that means highly miscible with each other.

Keywords: coal-ionic liquids cluster, COSMO-RS, DFT method, HOMO-LUMO interaction

Procedia PDF Downloads 304
2139 Release of Legacy Persistent Organic Pollutants and Mitigating Their Effects in Downstream Communities

Authors: Kimberley Rain Miner, Karl Kreutz, Larry LeBlanc

Abstract:

During the period of 1950-1970 persistent organic pollutants such as DDT, dioxin and PCB were released in the atmosphere and distributed through precipitation into glaciers throughout the world. Recent abrupt climate change is increasing the melt rate of these glaciers, introducing the toxins to the watershed. Studies have shown the existence of legacy pollutants in glacial ice, but neither the impact nor quantity of these toxins on downstream populations has been assessed. If these pollutants are released at toxic levels it will be necessary to create a mitigation plan to lower their impact on the affected communities.

Keywords: climate change, adaptation, mitigation, risk management

Procedia PDF Downloads 359
2138 Fatigue Life Prediction under Variable Loading Based a Non-Linear Energy Model

Authors: Aid Abdelkrim

Abstract:

A method of fatigue damage accumulation based upon application of energy parameters of the fatigue process is proposed in the paper. Using this model is simple, it has no parameter to be determined, it requires only the knowledge of the curve W–N (W: strain energy density N: number of cycles at failure) determined from the experimental Wöhler curve. To examine the performance of nonlinear models proposed in the estimation of fatigue damage and fatigue life of components under random loading, a batch of specimens made of 6082 T 6 aluminium alloy has been studied and some of the results are reported in the present paper. The paper describes an algorithm and suggests a fatigue cumulative damage model, especially when random loading is considered. This work contains the results of uni-axial random load fatigue tests with different mean and amplitude values performed on 6082T6 aluminium alloy specimens. The proposed model has been formulated to take into account the damage evolution at different load levels and it allows the effect of the loading sequence to be included by means of a recurrence formula derived for multilevel loading, considering complex load sequences. It is concluded that a ‘damaged stress interaction damage rule’ proposed here allows a better fatigue damage prediction than the widely used Palmgren–Miner rule, and a formula derived in random fatigue could be used to predict the fatigue damage and fatigue lifetime very easily. The results obtained by the model are compared with the experimental results and those calculated by the most fatigue damage model used in fatigue (Miner’s model). The comparison shows that the proposed model, presents a good estimation of the experimental results. Moreover, the error is minimized in comparison to the Miner’s model.

Keywords: damage accumulation, energy model, damage indicator, variable loading, random loading

Procedia PDF Downloads 395
2137 Foreign Direct Investment on Economic Growth by Industries in Central and Eastern European Countries

Authors: Shorena Pharjiani

Abstract:

The Present empirical paper investigates the relationship between FDI and economic growth by 10 selected industries in 10 Central and Eastern European countries from the period 1995 to 2012. Different estimation approaches were used to explore the connection between FDI and economic growth, for example OLS, RE, FE with and without time dummies. Obtained empirical results leads to some main consequences: First, the Central and East European countries (CEEC) attracted foreign direct investment, which raised the productivity of industries they entered in. It should be concluded that the linkage between FDI and output growth by industries is positive and significant enough to suggest that foreign firm’s participation enhanced the productivity of the industries they occupied. There had been an endogeneity problem in the regression and fixed effects estimation approach was used which partially corrected the regression analysis in order to make the results less biased. Second, it should be stressed that the results show that time has an important role in making FDI operational for enhancing output growth by industries via total factor productivity. Third, R&D positively affected economic growth and at the same time, it should take some time for research and development to influence economic growth. Fourth, the general trends masked crucial differences at the country level: over the last 20 years, the analysis of the tables and figures at the country level show that the main recipients of FDI of the 11 Central and Eastern European countries were Hungary, Poland and the Czech Republic. The main reason was that these countries had more open door policies for attracting the FDI. Fifth, according to the graphical analysis, while Hungary had the highest FDI inflow in this region, it was not reflected in the GDP growth as much as in other Central and Eastern European countries.

Keywords: central and East European countries (CEEC), economic growth, FDI, panel data

Procedia PDF Downloads 235
2136 Syngas From Polypropylene Gasification in a Fluidized Bed

Authors: Sergio Rapagnà, Alessandro Antonio Papa, Armando Vitale, Andre Di Carlo

Abstract:

In recent years the world population has enormously increased the use of plastic products for their living needs, in particular for transporting and storing consumer goods such as food and beverage. Plastics are widely used in the automotive industry, in construction of electronic equipment, clothing and home furnishings. Over the last 70 years, the annual production of plastic products has increased from 2 million tons to 460 million tons. About 20% of the last quantity is mismanaged as waste. The consequence of this mismanagement is the release of plastic waste into the terrestrial and marine environments which represents a danger to human health and the ecosystem. Recycling all plastics is difficult because they are often made with mixtures of polymers that are incompatible with each other and contain different additives. The products obtained are always of lower quality and after two/three recycling cycles they must be eliminated either by thermal treatment to produce heat or disposed of in landfill. An alternative to these current solutions is to obtain a mixture of gases rich in H₂, CO and CO₂ suitable for being profitably used for the production of chemicals with consequent savings fossil sources. Obtaining a hydrogen-rich syngas can be achieved by gasification process using the fluidized bed reactor, in presence of steam as the fluidization medium. The fluidized bed reactor allows the gasification process of plastics to be carried out at a constant temperature and allows the use of different plastics with different compositions and different grain sizes. Furthermore, during the gasification process the use of steam increase the gasification of char produced by the first pyrolysis/devolatilization process of the plastic particles. The bed inventory can be made with particles having catalytic properties such as olivine, capable to catalyse the steam reforming reactions of heavy hydrocarbons normally called tars, with a consequent increase in the quantity of gases produced. The plant is composed of a fluidized bed reactor made of AISI 310 steel, having an internal diameter of 0.1 m, containing 3 kg of olivine particles as a bed inventory. The reactor is externally heated by an oven up to 1000 °C. The hot producer gases that exit the reactor, after being cooled, are quantified using a mass flow meter. Gas analyzers are present to measure instantly the volumetric composition of H₂, CO, CO₂, CH₄ and NH₃. At the conference, the results obtained from the continuous gasification of polypropylene (PP) particles in a steam atmosphere at temperatures of 840-860 °C will be presented.

Keywords: gasification, fluidized bed, hydrogen, olivine, polypropyle

Procedia PDF Downloads 26
2135 Nonlinear Estimation Model for Rail Track Deterioration

Authors: M. Karimpour, L. Hitihamillage, N. Elkhoury, S. Moridpour, R. Hesami

Abstract:

Rail transport authorities around the world have been facing a significant challenge when predicting rail infrastructure maintenance work for a long period of time. Generally, maintenance monitoring and prediction is conducted manually. With the restrictions in economy, the rail transport authorities are in pursuit of improved modern methods, which can provide precise prediction of rail maintenance time and location. The expectation from such a method is to develop models to minimize the human error that is strongly related to manual prediction. Such models will help them in understanding how the track degradation occurs overtime under the change in different conditions (e.g. rail load, rail type, rail profile). They need a well-structured technique to identify the precise time that rail tracks fail in order to minimize the maintenance cost/time and secure the vehicles. The rail track characteristics that have been collected over the years will be used in developing rail track degradation prediction models. Since these data have been collected in large volumes and the data collection is done both electronically and manually, it is possible to have some errors. Sometimes these errors make it impossible to use them in prediction model development. This is one of the major drawbacks in rail track degradation prediction. An accurate model can play a key role in the estimation of the long-term behavior of rail tracks. Accurate models increase the track safety and decrease the cost of maintenance in long term. In this research, a short review of rail track degradation prediction models has been discussed before estimating rail track degradation for the curve sections of Melbourne tram track system using Adaptive Network-based Fuzzy Inference System (ANFIS) model.

Keywords: ANFIS, MGT, prediction modeling, rail track degradation

Procedia PDF Downloads 333