Search results for: lumped parameters model
20924 Estimating Evapotranspiration Irrigated Maize in Brazil Using a Hybrid Modelling Approach and Satellite Image Inputs
Authors: Ivo Zution Goncalves, Christopher M. U. Neale, Hiran Medeiros, Everardo Mantovani, Natalia Souza
Abstract:
Multispectral and thermal infrared imagery from satellite sensors coupled with climate and soil datasets were used to estimate evapotranspiration and biomass in center pivots planted to maize in Brazil during the 2016 season. The hybrid remote sensing based model named Spatial EvapoTranspiration Modelling Interface (SETMI) was applied using multispectral and thermal infrared imagery from the Landsat Thematic Mapper instrument. Field data collected by the IRRIGER center pivot management company included daily weather information such as maximum and minimum temperature, precipitation, relative humidity for estimating reference evapotranspiration. In addition, soil water content data were obtained every 0.20 m in the soil profile down to 0.60 m depth throughout the season. Early season soil samples were used to obtain water-holding capacity, wilting point, saturated hydraulic conductivity, initial volumetric soil water content, layer thickness, and saturated volumetric water content. Crop canopy development parameters and irrigation application depths were also inputs of the model. The modeling approach is based on the reflectance-based crop coefficient approach contained within the SETMI hybrid ET model using relationships developed in Nebraska. The model was applied to several fields located in Minas Gerais State in Brazil with approximate latitude: -16.630434 and longitude: -47.192876. The model provides estimates of real crop evapotranspiration (ET), crop irrigation requirements and all soil water balance outputs, including biomass estimation using multi-temporal satellite image inputs. An interpolation scheme based on the growing degree-day concept was used to model the periods between satellite inputs, filling the gaps between image dates and obtaining daily data. Actual and accumulated ET, accumulated cold temperature and water stress and crop water requirements estimated by the model were compared with data measured at the experimental fields. Results indicate that the SETMI modeling approach using data assimilation, showed reliable daily ET and crop water requirements for maize, interpolated between remote sensing observations, confirming the applicability of the SETMI model using new relationships developed in Nebraska for estimating mainly ET and water requirements in Brazil under tropical conditions.Keywords: basal crop coefficient, irrigation, remote sensing, SETMI
Procedia PDF Downloads 14020923 Series Network-Structured Inverse Models of Data Envelopment Analysis: Pitfalls and Solutions
Authors: Zohreh Moghaddas, Morteza Yazdani, Farhad Hosseinzadeh
Abstract:
Nowadays, data envelopment analysis (DEA) models featuring network structures have gained widespread usage for evaluating the performance of production systems and activities (Decision-Making Units (DMUs)) across diverse fields. By examining the relationships between the internal stages of the network, these models offer valuable insights to managers and decision-makers regarding the performance of each stage and its impact on the overall network. To further empower system decision-makers, the inverse data envelopment analysis (IDEA) model has been introduced. This model allows the estimation of crucial information for estimating parameters while keeping the efficiency score unchanged or improved, enabling analysis of the sensitivity of system inputs or outputs according to managers' preferences. This empowers managers to apply their preferences and policies on resources, such as inputs and outputs, and analyze various aspects like production, resource allocation processes, and resource efficiency enhancement within the system. The results obtained can be instrumental in making informed decisions in the future. The top result of this study is an analysis of infeasibility and incorrect estimation that may arise in the theory and application of the inverse model of data envelopment analysis with network structures. By addressing these pitfalls, novel protocols are proposed to circumvent these shortcomings effectively. Subsequently, several theoretical and applied problems are examined and resolved through insightful case studies.Keywords: inverse models of data envelopment analysis, series network, estimation of inputs and outputs, efficiency, resource allocation, sensitivity analysis, infeasibility
Procedia PDF Downloads 5220922 Assessment of Groundwater Quality around a Cement Factory in Ewekoro, Ogun State, Southwest Nigeria
Authors: A. O. David, A. A. Akaho, M. A. Abah, J. O. Ogunjimi
Abstract:
This study focuses on the growing concerns about the quality of groundwater found around cement factories, which have caused several health issues for residents located within two (2) kilometer radius. The qualities of groundwater were determined by an investigative study that involved the determination of some heavy metals and physicochemical properties in drinking water samples. Eight (8) samples of groundwater were collected from the eight sampling sites. The samples were analysed for the following parameters; iron, copper, manganese, zinc, lead, color, dissolved solids, electrical conductivity, pH, dissolved oxygen (DO), biological oxygen demand (BOD), chemical oxygen demand (COD), temperature, turbidity and total hardness using standard methods. The test results showed the variation of the investigated parameters in the samples as follows: temperature 26-31oC, pH 5.9-7.2, electrical conductivity (EC) 0.37 – 0.78 µS/cm, total hardness 181.8 – 333.0 mg/l, turbidity 0.00-0.05 FTU, colour 5-10 TCU, dissolved oxygen 4.31-5.01 mg/l, BOD 0.2-1.0 mg/l, COD 2.0 -4.0 mg/l, Cu 0.04 – 0.09 mg/l, Fe 0.006-0.122 mg/l, Zn 0.016-0.306 mg/l, Mn 0.01-0.05 mg/l and Pb < 0.001 mg/l. The World Health Organization's standard for drinking water quality guidelines was exceeded in several of the analyzed parameters' amounts in the drinking water samples from the study area. The dissolved oxygen was found to exceed 5.0 mg/l, which is the WHO permissible limit; also, Limestone was found to exceed the WHO maximum limit of 170 mg/l. All the above results confirmed the high pollution of the groundwater sources, and hence, they are not suitable for consumption without any prior treatment.Keywords: groundwater, quality, heavy metals, parameters
Procedia PDF Downloads 6520921 Application Difference between Cox and Logistic Regression Models
Authors: Idrissa Kayijuka
Abstract:
The logistic regression and Cox regression models (proportional hazard model) at present are being employed in the analysis of prospective epidemiologic research looking into risk factors in their application on chronic diseases. However, a theoretical relationship between the two models has been studied. By definition, Cox regression model also called Cox proportional hazard model is a procedure that is used in modeling data regarding time leading up to an event where censored cases exist. Whereas the Logistic regression model is mostly applicable in cases where the independent variables consist of numerical as well as nominal values while the resultant variable is binary (dichotomous). Arguments and findings of many researchers focused on the overview of Cox and Logistic regression models and their different applications in different areas. In this work, the analysis is done on secondary data whose source is SPSS exercise data on BREAST CANCER with a sample size of 1121 women where the main objective is to show the application difference between Cox regression model and logistic regression model based on factors that cause women to die due to breast cancer. Thus we did some analysis manually i.e. on lymph nodes status, and SPSS software helped to analyze the mentioned data. This study found out that there is an application difference between Cox and Logistic regression models which is Cox regression model is used if one wishes to analyze data which also include the follow-up time whereas Logistic regression model analyzes data without follow-up-time. Also, they have measurements of association which is different: hazard ratio and odds ratio for Cox and logistic regression models respectively. A similarity between the two models is that they are both applicable in the prediction of the upshot of a categorical variable i.e. a variable that can accommodate only a restricted number of categories. In conclusion, Cox regression model differs from logistic regression by assessing a rate instead of proportion. The two models can be applied in many other researches since they are suitable methods for analyzing data but the more recommended is the Cox, regression model.Keywords: logistic regression model, Cox regression model, survival analysis, hazard ratio
Procedia PDF Downloads 45520920 Numerical Analysis of the Coanda Effect on the Classical Interior Ejectors
Authors: Alexandru Dumitrache, Florin Frunzulica, Octavian Preotu
Abstract:
The flow mitigation detachment problem near solid surfaces, resulting in improved globally aerodynamic performance by exploiting the Coanda effect on surfaces, has been addressed extensively in the literature, since 1940. The research is carried on and further developed, using modern means of calculation and new experimental methods. In this paper, it is shown interest in the detailed behavior of a classical interior ejector assisted by the Coanda effect, used in propulsion systems. For numerical investigations, an implicit formulation of RANS equations for axisymmetric flow with a shear stress transport k- ω (SST model) turbulence model is used. The obtained numerical results emphasize the efficiency of the ejector, depending on the physical parameters of the flow and the geometric configuration. Furthermore, numerical investigations are carried out regarding the evolution of the Reynolds number when the jet is attached to the wall, considering three geometric configurations: sudden expansion, open cavity and sudden expansion with divergent at the inlet. Therefore, further insight into complexities involving issues such as the variety of flow structure and the related bifurcation and flow instabilities are provided. Thus, the conditions and the limits within which one can benefit from the advantages of Coanda-type flows are determined.Keywords: Coanda effect, Coanda ejector, CFD, stationary bifurcation, sudden expansion
Procedia PDF Downloads 21420919 An Algorithm to Find Fractional Edge Domination Number and Upper Fractional Edge Domination Number of an Intuitionistic Fuzzy Graph
Authors: Karunambigai Mevani Govindasamy, Sathishkumar Ayyappan
Abstract:
In this paper, we formulate the algorithm to find out the dominating function parameters of Intuitionistic Fuzzy Graphs(IFG). The methodology we adopted here is converting any physical problem into an IFG, and that has been transformed into Intuitionistic Fuzzy Matrix. Using Linear Program Solver software (LiPS), we found the defined parameters for the given IFG. We obtained these parameters for a path and cycle IFG. This study can be extended to other varieties of IFG. In particular, we obtain the definition of edge dominating function, minimal edge dominating function, fractional edge domination number (γ_if^') and upper fractional edge domination number (Γ_if^') of an intuitionistic fuzzy graph. Also, we formulated an algorithm which is appropriate to work on LiPS to find fractional edge domination number and upper fractional edge domination number of an IFG.Keywords: fractional edge domination number, intuitionistic fuzzy cycle, intuitionistic fuzzy graph, intuitionistic fuzzy path
Procedia PDF Downloads 17420918 Optimized Microwave Pretreatment of Rice Straw for Conversion into Lignin Free and High Crystalline Cellulose
Authors: Mohd Ishfaq Bhat, Navin Chandra Shahi, Umesh Chandra Lohani
Abstract:
The present study aimed to evaluate the effect of microwave application in synergy with the conventional sodium chlorite delignification of rice straw biomass. For the study, Box-Behnken experimental design involving four independent parameters, each with three levels viz. microwave power (480-800 W), irradiation time (4-12 min), bleaching solution concentration (0.4-3.0%), and bleaching time (1-5h) was used. The response was taken in the form of delignification percentage. The optimization of process parameters was done through response surface methodology. The respective optimum parameters of microwave power, irradiation time, bleaching solution concentration, and bleaching time were obtained as 671 W, 8.66 min, 2.67%, and 1h. The delignification percentage achieved at optimum conditions was 93.51%. The spectral, morphological, and x-ray diffraction characteristics of the rice straw powder after delignification showed a complete absence of lignin peaks, deconstruction of lignocellulose complex, and an increase of crystallinity (from 39.8 to 61.6 %).Keywords: lignocellulosic biomass, delignification, microwaves, rice straw, characterization
Procedia PDF Downloads 14720917 Study of Individual Parameters on the Enzymatic Glycosidation of Betulinic Acid by Novozyme-435
Authors: A. U. Adamu, Hamisu Abdu, A. A. Saidu
Abstract:
The enzymatic synthesis of 3-O-β-D-glucopyranoside-betulinic acid using Novozyme-435 as a catalyst was studied. The effect of various parameters such as substrate molar ratio, reaction temperature, reaction time, re-used enzymes and amount of enzymes were investigated. The optimum rection conditions for the enzymatic glycosidation of betulinic acid in an organic solvent using Novozym-435 was found to be at 1:1.2 substrate molar ratio, 55oC, 24 h and 180 mg of enzymes with percentage conversion of 88.69 %.Keywords: betulinic acid, glycosidation, novozyme-435, optimization
Procedia PDF Downloads 42620916 Modeling of the Biodegradation Performance of a Membrane Bioreactor to Enhance Water Reuse in Agri-food Industry - Poultry Slaughterhouse as an Example
Authors: masmoudi Jabri Khaoula, Zitouni Hana, Bousselmi Latifa, Akrout Hanen
Abstract:
Mathematical modeling has become an essential tool for sustainable wastewater management, particularly for the simulation and the optimization of complex processes involved in activated sludge systems. In this context, the activated sludge model (ASM3h) was used for the simulation of a Biological Membrane Reactor (MBR) as it includes the integration of biological wastewater treatment and physical separation by membrane filtration. In this study, the MBR with a useful volume of 12.5 L was fed continuously with poultry slaughterhouse wastewater (PSWW) for 50 days at a feed rate of 2 L/h and for a hydraulic retention time (HRT) of 6.25h. Throughout its operation, High removal efficiency was observed for the removal of organic pollutants in terms of COD with 84% of efficiency. Moreover, the MBR has generated a treated effluent which fits with the limits of discharge into the public sewer according to the Tunisian standards which were set in March 2018. In fact, for the nitrogenous compounds, average concentrations of nitrate and nitrite in the permeat reached 0.26±0.3 mg. L-1 and 2.2±2.53 mg. L-1, respectively. The simulation of the MBR process was performed using SIMBA software v 5.0. The state variables employed in the steady state calibration of the ASM3h were determined using physical and respirometric methods. The model calibration was performed using experimental data obtained during the first 20 days of the MBR operation. Afterwards, kinetic parameters of the model were adjusted and the simulated values of COD, N-NH4+and N- NOx were compared with those reported from the experiment. A good prediction was observed for the COD, N-NH4+and N- NOx concentrations with 467 g COD/m³, 110.2 g N/m³, 3.2 g N/m³ compared to the experimental data which were 436.4 g COD/m³, 114.7 g N/m³ and 3 g N/m³, respectively. For the validation of the model under dynamic simulation, the results of the experiments obtained during the second treatment phase of 30 days were used. It was demonstrated that the model simulated the conditions accurately by yielding a similar pattern on the variation of the COD concentration. On the other hand, an underestimation of the N-NH4+ concentration was observed during the simulation compared to the experimental results and the measured N-NO3 concentrations were lower than the predicted ones, this difference could be explained by the fact that the ASM models were mainly designed for the simulation of biological processes in the activated sludge systems. In addition, more treatment time could be required by the autotrophic bacteria to achieve a complete and stable nitrification. Overall, this study demonstrated the effectiveness of mathematical modeling in the prediction of the performance of the MBR systems with respect to organic pollution, the model can be further improved for the simulation of nutrients removal for a longer treatment period.Keywords: activated sludge model (ASM3h), membrane bioreactor (MBR), poultry slaughter wastewater (PSWW), reuse
Procedia PDF Downloads 5820915 A Numerical Investigation of Segmental Lining Joints Interactions in Tunnels
Authors: M. H. Ahmadi, A. Mortazavi, H. Zarei
Abstract:
Several authors have described the main mechanism of formation of cracks in the segment lining during the construction of tunnels with tunnel boring machines. A comprehensive analysis of segmental lining joints may help to guarantee a safe construction during Tunneling and serviceable stages. The most frequent types of segment damage are caused by a condition of uneven segment matching due to contact deficiencies. This paper investigated the interaction mechanism of precast concrete lining joints in tunnels. The Discrete Element Method (DEM) was used to analyze a typical segmental lining model consisting of six segment rings. In the analyses, typical segmental lining design parameters of the Ghomrood water conveyance tunnel, Iran were employed in the study. In the conducted analysis, the worst-case scenario of loading faced during the boring of Ghomrood tunnel was considered. This was associated with the existence of a crushed zone dipping at 75 degree at the location of the key segment. In the analysis, moreover, the effect of changes in horizontal stress ratio on the loads on the segment was assessed. The boundary condition associated with K (ratio of the horizontal to the vertical stress) values of 0.5, 1, 1.5 and 2 were applied to the model and separate analysis was conducted for each case. Important parameters such as stress, moments, and displacements were measured at joint locations and the surrounding rock. Accordingly, the segment joint interactions were assessed and analyzed. Moreover, rock mass properties of the Ghomrood in Ghom were adopted. In this study, the load acting on segments joints are included a crushed zone stratum force that intersect tunnel with 75 slopes in the location of the key segment, gravity force of segments and earth pressures. A numerical investigation was used for different coefficients of stress concentration of 0.5, 1, 1.5, 2 and different geological conditions of saturated crushed zone under the critical scenario. The numerical results also demonstrate that maximum bending moments in longitudinal joints occurred for crushed zone with the weaken strengths (Sandstone). Besides that, increasing the load in segment-stratum interfaces affected radial stress in longitudinal joints and finally the opening of joints occurred.Keywords: joint, interface, segment, contact
Procedia PDF Downloads 25820914 On Differential Growth Equation to Stochastic Growth Model Using Hyperbolic Sine Function in Height/Diameter Modeling of Pines
Authors: S. O. Oyamakin, A. U. Chukwu
Abstract:
Richard's growth equation being a generalized logistic growth equation was improved upon by introducing an allometric parameter using the hyperbolic sine function. The integral solution to this was called hyperbolic Richard's growth model having transformed the solution from deterministic to a stochastic growth model. Its ability in model prediction was compared with the classical Richard's growth model an approach which mimicked the natural variability of heights/diameter increment with respect to age and therefore provides a more realistic height/diameter predictions using the coefficient of determination (R2), Mean Absolute Error (MAE) and Mean Square Error (MSE) results. The Kolmogorov-Smirnov test and Shapiro-Wilk test was also used to test the behavior of the error term for possible violations. The mean function of top height/Dbh over age using the two models under study predicted closely the observed values of top height/Dbh in the hyperbolic Richard's nonlinear growth models better than the classical Richard's growth model.Keywords: height, Dbh, forest, Pinus caribaea, hyperbolic, Richard's, stochastic
Procedia PDF Downloads 48020913 Design Optimisation of a Novel Cross Vane Expander-Compressor Unit for Refrigeration System
Authors: Y. D. Lim, K. S. Yap, K. T. Ooi
Abstract:
In recent years, environmental issue has been a hot topic in the world, especially the global warming effect caused by conventional non-environmentally friendly refrigerants has increased. Several studies of a more energy-efficient and environmentally friendly refrigeration system have been conducted in order to tackle the issue. In search of a better refrigeration system, CO2 refrigeration system has been proposed as a better option. However, the high throttling loss involved during the expansion process of the refrigeration cycle leads to a relatively low efficiency and thus the system is impractical. In order to improve the efficiency of the refrigeration system, it is suggested by replacing the conventional expansion valve in the refrigeration system with an expander. Based on this issue, a new type of expander-compressor combined unit, named Cross Vane Expander-Compressor (CVEC) was introduced to replace the compressor and the expansion valve of a conventional refrigeration system. A mathematical model was developed to calculate the performance of CVEC, and it was found that the machine is capable of saving the energy consumption of a refrigeration system by as much as 18%. Apart from energy saving, CVEC is also geometrically simpler and more compact. To further improve its efficiency, optimization study of the device is carried out. In this report, several design parameters of CVEC were chosen to be the variables of optimization study. This optimization study was done in a simulation program by using complex optimization method, which is a direct search, multi-variables and constrained optimization method. It was found that the main design parameters, which was shaft radius was reduced around 8% while the inner cylinder radius was remained unchanged at its lower limit after optimization. Furthermore, the port sizes were increased to their upper limit after optimization. The changes of these design parameters have resulted in reduction of around 12% in the total frictional loss and reduction of 4% in power consumption. Eventually, the optimization study has resulted in an improvement in the mechanical efficiency CVEC by 4% and improvement in COP by 6%.Keywords: complex optimization method, COP, cross vane expander-compressor, CVEC, design optimization, direct search, energy saving, improvement, mechanical efficiency, multi variables
Procedia PDF Downloads 37320912 Impact of Climate Variability on Household's Crop Income in Central Highlands and Arssi Grain Plough Areas of Ethiopia
Authors: Arega Shumetie Ademe, Belay Kassa, Degye Goshu, Majaliwa Mwanjalolo
Abstract:
Currently the world economy is suffering from one critical problem, climate change. Some studies done before identified that impact of the problem is region specific means in some part of the world (temperate zone) there is improvement in agricultural performance but in some others like in the tropics there is drastic reduction in crop production and crop income. Climate variability is becoming dominant cause of short-term fluctuation in rain-fed agricultural production and income of developing countries. The purely rain-fed Ethiopian agriculture is the most vulnerable sector to the risks and impacts of climate variability. Thus, this study tried to identify impact of climate variability on crop income of smallholders in Ethiopia. The research used eight rounded unbalanced panel data from 1994- 2014 collected from six villages in the study area. After having all diagnostic tests the research used fixed effect method of regression. Based on the regression result rainfall and temperature deviation from their respective long term averages have negative and significant effect on crop income. Other extreme devastating shocks like flood, storm and frost, which are sourced from climate variability, have significant and negative effect on crop income of households’. Parameters that notify rainfall inconsistency like late start, variation in availability at growing season, and early cessation are critical problems for crop income of smallholder households as to the model result. Given this, impact of climate variability is not consistent in different agro-ecologies of the country. Rainfall variability has similar impact on crop income in different agro-ecology, but variation in temperature affects cold agro-ecology villages negatively and significantly, while it has positive effect in warm villages. Parameters that represent rainfall inconsistency have similar impact in both agro-ecologies and the aggregate model regression. This implies climate variability sourced from rainfall inconsistency is the main problem of Ethiopian agriculture especially the crop production sub-sector of smallholder households.Keywords: climate variability, crop income, household, rainfall, temperature
Procedia PDF Downloads 37620911 Development of a Predictive Model to Prevent Financial Crisis
Authors: Tengqin Han
Abstract:
Delinquency has been a crucial factor in economics throughout the years. Commonly seen in credit card and mortgage, it played one of the crucial roles in causing the most recent financial crisis in 2008. In each case, a delinquency is a sign of the loaner being unable to pay off the debt, and thus may cause a lost of property in the end. Individually, one case of delinquency seems unimportant compared to the entire credit system. China, as an emerging economic entity, the national strength and economic strength has grown rapidly, and the gross domestic product (GDP) growth rate has remained as high as 8% in the past decades. However, potential risks exist behind the appearance of prosperity. Among the risks, the credit system is the most significant one. Due to long term and a large amount of balance of the mortgage, it is critical to monitor the risk during the performance period. In this project, about 300,000 mortgage account data are analyzed in order to develop a predictive model to predict the probability of delinquency. Through univariate analysis, the data is cleaned up, and through bivariate analysis, the variables with strong predictive power are detected. The project is divided into two parts. In the first part, the analysis data of 2005 are split into 2 parts, 60% for model development, and 40% for in-time model validation. The KS of model development is 31, and the KS for in-time validation is 31, indicating the model is stable. In addition, the model is further validation by out-of-time validation, which uses 40% of 2006 data, and KS is 33. This indicates the model is still stable and robust. In the second part, the model is improved by the addition of macroeconomic economic indexes, including GDP, consumer price index, unemployment rate, inflation rate, etc. The data of 2005 to 2010 is used for model development and validation. Compared with the base model (without microeconomic variables), KS is increased from 41 to 44, indicating that the macroeconomic variables can be used to improve the separation power of the model, and make the prediction more accurate.Keywords: delinquency, mortgage, model development, model validation
Procedia PDF Downloads 22820910 Full Analytical Procedure to Derive P-I Diagram of a Steel Beam under Blast Loading
Authors: L. Hamra, J. F. Demonceau, V. Denoël
Abstract:
The aim of this paper is to study a beam extracted from a frame and subjected to blast loading. The demand of ductility depends on six dimensionless parameters: two related to the blast loading, two referring to the bending behavior of the beam and two corresponding to the dynamic behavior of the rest of the structure. We develop a full analytical procedure that provides the ductility demand as a function of these six dimensionless parameters.Keywords: analytical procedure, blast loading, membrane force, P-I diagram
Procedia PDF Downloads 42720909 Simulation-Based Parametric Study for the Hybrid Superplastic Forming of AZ31
Authors: Fatima Ghassan Al-Abtah, Naser Al-Huniti, Elsadig Mahdi
Abstract:
As the lightest constructional metal on earth, magnesium alloys offer excellent potential for weight reduction in the transportation industry, and it was observed that some magnesium alloys exhibit superior ductility and superplastic behavior at high temperatures. The main limitation of the superplastic forming (SPF) includes the low production rate since it needs a long forming time for each part. Through this study, an SPF process that starts with a mechanical pre-forming stage is developed to promote formability and reduce forming time. A two-dimensional finite element model is used to simulate the process. The forming process consists of two steps. At the pre-forming step (deep drawing), the sheet is drawn into the die to a preselected level, using a mechanical punch, and at the second step (SPF) a pressurized gas is applied at a controlled rate. It is shown that a significant reduction in forming time and improved final thickness uniformity can be achieved when the hybrid forming technique is used, where the process achieved a fully formed part at 400°C. Investigation for the impact of different forming process parameters achieved by comparing forming time and the distribution of final thickness that were obtained from the simulation analysis. Maximum thinning decreased from over 67% to less than 55% and forming time significantly decreased by more than 6 minutes, and the required gas pressure profile was predicted for optimum forming process parameters based on the 0.001/sec target constant strain rate within the sheet.Keywords: magnesium, plasticity, superplastic forming, finite element analysis
Procedia PDF Downloads 15620908 Characterization of an Extrapolation Chamber for Dosimetry of Low Energy X-Ray Beams
Authors: Fernanda M. Bastos, Teógenes A. da Silva
Abstract:
Extrapolation chambers were designed to be used as primary standard dosimeter for measuring absorbed dose in a medium in beta radiation and low energy x-rays. The International Organization for Standardization established series of reference x-radiation for calibrating and determining the energy dependence of dosimeters that are to be reproduced in metrology laboratories. Standardization of the low energy x-ray beams with tube potential lower than 30 kV may be affected by the instrument used for dosimetry. In this work, parameters of a 23392 model PTW extrapolation chamber were determined aiming its use in low energy x-ray beams as a reference instrument.Keywords: extrapolation chamber, low energy x-rays, x-ray dosimetry, X-ray metrology
Procedia PDF Downloads 39520907 Powder Flow with Normalized Powder Particles Size Distribution and Temperature Analyses in Laser Melting Deposition: Analytical Modelling and Experimental Validation
Authors: Muhammad Arif Mahmood, Andrei C. Popescu, Mihai Oane, Diana Chioibascu, Carmen Ristoscu, Ion N. Mihailescu
Abstract:
Powder flow and temperature distributions are recognized as influencing factors during laser melting deposition (LMD) process, that not only affect the consolidation rate but also characteristics of the deposited layers. Herewith, two simplified analytical models will be presented to simulate the powder flow with the inclusion of powder particles size distribution in Gaussian form, under three powder jet nozzles, and temperature analyses during LMD process. The output of the 1st model will serve as the input in the 2nd model. The models will be validated with experimental data, i.e., weight measurement method for powder particles distribution and infrared imaging for temperature analyses. This study will increase the cost-efficiency of the LMD process by adjustment of the operating parameters for reaching optimal powder debit and energy. This research has received funds under the Marie Sklodowska-Curie grant agreement No. 764935, from the European Union’s Horizon 2020 research and innovation program.Keywords: laser additive manufacturing, powder particles size distribution in Gaussian form, powder stream distribution, temperature analyses
Procedia PDF Downloads 13520906 Identifying Knowledge Gaps in Incorporating Toxicity of Particulate Matter Constituents for Developing Regulatory Limits on Particulate Matter
Authors: Ananya Das, Arun Kumar, Gazala Habib, Vivekanandan Perumal
Abstract:
Regulatory bodies has proposed limits on Particulate Matter (PM) concentration in air; however, it does not explicitly indicate the incorporation of effects of toxicities of constituents of PM in developing regulatory limits. This study aimed to provide a structured approach to incorporate toxic effects of components in developing regulatory limits on PM. A four-step human health risk assessment framework consists of - (1) hazard identification (parameters: PM and its constituents and their associated toxic effects on health), (2) exposure assessment (parameters: concentrations of PM and constituents, information on size and shape of PM; fate and transport of PM and constituents in respiratory system), (3) dose-response assessment (parameters: reference dose or target toxicity dose of PM and its constituents), and (4) risk estimation (metric: hazard quotient and/or lifetime incremental risk of cancer as applicable). Then parameters required at every step were obtained from literature. Using this information, an attempt has been made to determine limits on PM using component-specific information. An example calculation was conducted for exposures of PM2.5 and its metal constituents from Indian ambient environment to determine limit on PM values. Identified data gaps were: (1) concentrations of PM and its constituents and their relationship with sampling regions, (2) relationship of toxicity of PM with its components.Keywords: air, component-specific toxicity, human health risks, particulate matter
Procedia PDF Downloads 31120905 Proactive WPA/WPA2 Security Using DD-WRT Firmware
Authors: Mustafa Kamoona, Mohamed El-Sharkawy
Abstract:
Although the latest Wireless Local Area Network technology Wi-Fi 802.11i standard addresses many of the security weaknesses of the antecedent Wired Equivalent Privacy (WEP) protocol, there are still scenarios where the network security are still vulnerable. The first security model that 802.11i offers is the Personal model which is very cheap and simple to install and maintain, yet it uses a Pre Shared Key (PSK) and thus has a low to medium security level. The second model that 802.11i provide is the Enterprise model which is highly secured but much more expensive and difficult to install/maintain and requires the installation and maintenance of an authentication server that will handle the authentication and key management for the wireless network. A central issue with the personal model is that the PSK needs to be shared with all the devices that are connected to the specific Wi-Fi network. This pre-shared key, unless changed regularly, can be cracked using offline dictionary attacks within a matter of hours. The key is burdensome to change in all the connected devices manually unless there is some kind of algorithm that coordinate this PSK update. The key idea of this paper is to propose a new algorithm that proactively and effectively coordinates the pre-shared key generation, management, and distribution in the cheap WPA/WPA2 personal security model using only a DD-WRT router.Keywords: Wi-Fi, WPS, TLS, DD-WRT
Procedia PDF Downloads 23320904 The Generalized Pareto Distribution as a Model for Sequential Order Statistics
Authors: Mahdy Esmailian, Mahdi Doostparast, Ahmad Parsian
Abstract:
In this article, sequential order statistics (SOS) censoring type II samples coming from the generalized Pareto distribution are considered. Maximum likelihood (ML) estimators of the unknown parameters are derived on the basis of the available multiple SOS data. Necessary conditions for existence and uniqueness of the derived ML estimates are given. Due to complexity in the proposed likelihood function, a useful re-parametrization is suggested. For illustrative purposes, a Monte Carlo simulation study is conducted and an illustrative example is analysed.Keywords: bayesian estimation, generalized pareto distribution, maximum likelihood estimation, sequential order statistics
Procedia PDF Downloads 50920903 Forecasting Age-Specific Mortality Rates and Life Expectancy at Births for Malaysian Sub-Populations
Authors: Syazreen N. Shair, Saiful A. Ishak, Aida Y. Yusof, Azizah Murad
Abstract:
In this paper, we forecast age-specific Malaysian mortality rates and life expectancy at births by gender and ethnic groups including Malay, Chinese and Indian. Two mortality forecasting models are adopted the original Lee-Carter model and its recent modified version, the product ratio coherent model. While the first forecasts the mortality rates for each subpopulation independently, the latter accounts for the relationship between sub-populations. The evaluation of both models is performed using the out-of-sample forecast errors which are mean absolute percentage errors (MAPE) for mortality rates and mean forecast errors (MFE) for life expectancy at births. The best model is then used to perform the long-term forecasts up to the year 2030, the year when Malaysia is expected to become an aged nation. Results suggest that in terms of overall accuracy, the product ratio model performs better than the original Lee-Carter model. The association of lower mortality group (Chinese) in the subpopulation model can improve the forecasts of high mortality groups (Malay and Indian).Keywords: coherent forecasts, life expectancy at births, Lee-Carter model, product-ratio model, mortality rates
Procedia PDF Downloads 21920902 Efficient Sampling of Probabilistic Program for Biological Systems
Authors: Keerthi S. Shetty, Annappa Basava
Abstract:
In recent years, modelling of biological systems represented by biochemical reactions has become increasingly important in Systems Biology. Biological systems represented by biochemical reactions are highly stochastic in nature. Probabilistic model is often used to describe such systems. One of the main challenges in Systems biology is to combine absolute experimental data into probabilistic model. This challenge arises because (1) some molecules may be present in relatively small quantities, (2) there is a switching between individual elements present in the system, and (3) the process is inherently stochastic on the level at which observations are made. In this paper, we describe a novel idea of combining absolute experimental data into probabilistic model using tool R2. Through a case study of the Transcription Process in Prokaryotes we explain how biological systems can be written as probabilistic program to combine experimental data into the model. The model developed is then analysed in terms of intrinsic noise and exact sampling of switching times between individual elements in the system. We have mainly concentrated on inferring number of genes in ON and OFF states from experimental data.Keywords: systems biology, probabilistic model, inference, biology, model
Procedia PDF Downloads 34920901 Creative Mathematics – Action Research of a Professional Development Program in an Icelandic Compulsory School
Authors: Osk Dagsdottir
Abstract:
Background—Gait classifying allows clinicians to differentiate gait patterns into clinically important categories that help in clinical decision making. Reliable comparison of gait data between normal and patients requires knowledge of the gait parameters of normal children's specific age group. However, there is still a lack of the gait database for normal children of different ages. Objectives—This study aims to investigate the kinematics of the lower limb joints during gait for normal children in different age groups. Methods—Fifty-three normal children (34 boys, 19 girls) were recruited in this study. All the children were aged between 5 to 16 years old. Age groups were defined as three types: young child aged (5-7), child (8-11), and adolescent (12-16). When a participant agreed to take part in the project, their parents signed a consent form. Vicon® motion capture system was used to collect gait data. Participants were asked to walk at their comfortable speed along a 10-meter walkway. Each participant walked up to 20 trials. Three good trials were analyzed using the Vicon Plug-in-Gait model to obtain parameters of the gait, e.g., walking speed, cadence, stride length, and joint parameters, e.g., joint angle, force, moments, etc. Moreover, each gait cycle was divided into 8 phases. The range of motion (ROM) angle of pelvis, hip, knee, and ankle joints in three planes of both limbs were calculated using an in-house program. Results—The temporal-spatial variables of three age groups of normal children were compared between each other; it was found that there was a significant difference (p < 0.05) between the groups. The step length and walking speed were gradually increasing from young child to adolescent, while cadence was gradually decreasing from young child to adolescent group. The mean and standard deviation (SD) of the step length of young child, child and adolescent groups were 0.502 ± 0.067 m, 0.566 ± 0.061 m and 0.672 ± 0.053 m, respectively. The mean and SD of the cadence of the young child, child and adolescent groups were 140.11±15.79 step/min, 129±11.84 step/min, and a 115.96±6.47 step/min, respectively. Moreover, it was observed that there were significant differences in kinematic parameters, either whole gait cycle or each phase. For example, RoM of knee angle in the sagittal plane in the whole cycle of young child group is (65.03±0.52 deg) larger than child group (63.47±0.47 deg). Conclusion—Our result showed that there are significant differences between each age group in the gait phases and thus children walking performance changes with ages. Therefore, it is important for the clinician to consider the age group when analyzing the patients with lower limb disorders before any clinical treatment.Keywords: action research, creative learning, mathematics education, professional development
Procedia PDF Downloads 10820900 A Data-Driven Compartmental Model for Dengue Forecasting and Covariate Inference
Authors: Yichao Liu, Peter Fransson, Julian Heidecke, Jonas Wallin, Joacim Rockloev
Abstract:
Dengue, a mosquito-borne viral disease, poses a significant public health challenge in endemic tropical or subtropical countries, including Sri Lanka. To reveal insights into the complexity of the dynamics of this disease and study the drivers, a comprehensive model capable of both robust forecasting and insightful inference of drivers while capturing the co-circulating of several virus strains is essential. However, existing studies mostly focus on only one aspect at a time and do not integrate and carry insights across the siloed approach. While mechanistic models are developed to capture immunity dynamics, they are often oversimplified and lack integration of all the diverse drivers of disease transmission. On the other hand, purely data-driven methods lack constraints imposed by immuno-epidemiological processes, making them prone to overfitting and inference bias. This research presents a hybrid model that combines machine learning techniques with mechanistic modelling to overcome the limitations of existing approaches. Leveraging eight years of newly reported dengue case data, along with socioeconomic factors, such as human mobility, weekly climate data from 2011 to 2018, genetic data detecting the introduction and presence of new strains, and estimates of seropositivity for different districts in Sri Lanka, we derive a data-driven vector (SEI) to human (SEIR) model across 16 regions in Sri Lanka at the weekly time scale. By conducting ablation studies, the lag effects allowing delays up to 12 weeks of time-varying climate factors were determined. The model demonstrates superior predictive performance over a pure machine learning approach when considering lead times of 5 and 10 weeks on data withheld from model fitting. It further reveals several interesting interpretable findings of drivers while adjusting for the dynamics and influences of immunity and introduction of a new strain. The study uncovers strong influences of socioeconomic variables: population density, mobility, household income and rural vs. urban population. The study reveals substantial sensitivity to the diurnal temperature range and precipitation, while mean temperature and humidity appear less important in the study location. Additionally, the model indicated sensitivity to vegetation index, both max and average. Predictions on testing data reveal high model accuracy. Overall, this study advances the knowledge of dengue transmission in Sri Lanka and demonstrates the importance of incorporating hybrid modelling techniques to use biologically informed model structures with flexible data-driven estimates of model parameters. The findings show the potential to both inference of drivers in situations of complex disease dynamics and robust forecasting models.Keywords: compartmental model, climate, dengue, machine learning, social-economic
Procedia PDF Downloads 8420899 Machine Learning Model Applied for SCM Processes to Efficiently Determine Its Impacts on the Environment
Authors: Elena Puica
Abstract:
This paper aims to investigate the impact of Supply Chain Management (SCM) on the environment by applying a Machine Learning model while pointing out the efficiency of the technology used. The Machine Learning model was used to derive the efficiency and optimization of technology used in SCM and the environmental impact of SCM processes. The model applied is a predictive classification model and was trained firstly to determine which stage of the SCM has more outputs and secondly to demonstrate the efficiency of using advanced technology in SCM instead of recuring to traditional SCM. The outputs are the emissions generated in the environment, the consumption from different steps in the life cycle, the resulting pollutants/wastes emitted, and all the releases to air, land, and water. This manuscript presents an innovative approach to applying advanced technology in SCM and simultaneously studies the efficiency of technology and the SCM's impact on the environment. Identifying the conceptual relationships between SCM practices and their impact on the environment is a new contribution to the research. The authors can take a forward step in developing recent studies in SCM and its effects on the environment by applying technology.Keywords: machine-learning model in SCM, SCM processes, SCM and the environmental impact, technology in SCM
Procedia PDF Downloads 11620898 Design of Black-Seed Pulp biomass-Derived New Bio-Sorbent by Combining Methods of Mineral Acids and High-Temperature for Arsenic Removal
Authors: Mozhgan Mohammadi, Arezoo Ghadi
Abstract:
Arsenic is known as a potential threat to the environment. Therefore, the aim of this research is to assess the arsenic removal efficiency from an aqueous solution, with a new biosorbent composed of a black seed pulp (BSP). To treat BSP, the combination of two methods (i.e. treating with mineral acids and use at high temperature) was used and designed bio-sorbent called BSP-activated/carbonized. The BSP-activated and BSP-carbonized were also prepared using HCL and 400°C temperature, respectively, to compare the results of each three methods. Followed by, adsorption parameters such as pH, initial ion concentration, biosorbent dosage, contact time, and temperature were assessed. It was found that the combination method has provided higher adsorption capacity so that up to ~99% arsenic removal was observed with BSP-activated/carbonized at pH of 7.0 and 40°C. The adsorption capacity for BSP-carbonized and BSP-activated were 87.92% (pH: 7, 60°C) and 78.50% (pH: 6, 90°C), respectively. Moreover, adsorption kinetics data indicated the best fit with the pseudo-second-order model. The maximum biosorption capacity, by the Langmuir isotherm model, was also recorded for BSP-activated/carbonized (53.47 mg/g). It is notable that arsenic adsorption on studied bio sorbents takes place as spontaneous and through chemisorption along with the endothermic nature of the biosorption process and reduction of random collision in the solid-liquid phase.Keywords: black seed pulp, bio-sorbents, treatment of sorbents, adsorption isotherms
Procedia PDF Downloads 9520897 Application of Computational Flow Dynamics (CFD) Analysis for Surge Inception and Propagation for Low Head Hydropower Projects
Authors: M. Mohsin Munir, Taimoor Ahmad, Javed Munir, Usman Rashid
Abstract:
Determination of maximum elevation of a flowing fluid due to sudden rejection of load in a hydropower facility is of great interest to hydraulic engineers to ensure safety of the hydraulic structures. Several mathematical models exist that employ one-dimensional modeling for the determination of surge but none of these perfectly simulate real-time circumstances. The paper envisages investigation of surge inception and propagation for a Low Head Hydropower project using Computational Fluid Dynamics (CFD) analysis on FLOW-3D software package. The fluid dynamic model utilizes its analysis for surge by employing Reynolds’ Averaged Navier-Stokes Equations (RANSE). The CFD model is designed for a case study at Taunsa hydropower Project in Pakistan. Various scenarios have run through the model keeping in view upstream boundary conditions. The prototype results were then compared with the results of physical model testing for the same scenarios. The results of the numerical model proved quite accurate coherence with the physical model testing and offers insight into phenomenon which are not apparent in physical model and shall be adopted in future for the similar low head projects limiting delays and cost incurred in the physical model testing.Keywords: surge, FLOW-3D, numerical model, Taunsa, RANSE
Procedia PDF Downloads 36120896 Optimization of Submerged Arc Welding Parameters for Joining SS304 and MS1018
Authors: Jasvinder Singh, Manjinder Singh
Abstract:
Welding of dissimilar materials is a complicated process due to the difference in melting point of two materials. Thermal conductivity and coefficient of thermal expansion of dissimilar materials also different; therefore, residual stresses produced in the weldment and base metal are the most critical problem associated with the joining of dissimilar materials. Tensile strength and impact toughness also reduced due to the residual stresses. In the present research work, an attempt has been made to weld SS304 and MS1018 dissimilar materials by submerged arc welding (SAW). By conducting trail, runs most effective parameters welding current, Arc voltage, welding speed and nozzle to plate distance were selected to weld these materials. The fractional factorial technique was used to optimize the welding parameters. Effect on tensile strength (TS), fracture toughness (FT) and microhardness of weldment were studied. It was concluded that by optimizing welding current, voltage and welding speed the properties of weldment can be enhanced.Keywords: SAW, Tensile Strength (TS), fracture toughness, micro hardness
Procedia PDF Downloads 53820895 Interaction between Breathiness and Nasality: An Acoustic Analysis
Authors: Pamir Gogoi, Ratree Wayland
Abstract:
This study investigates the acoustic measures of breathiness when coarticulated with nasality. The acoustic correlates of breathiness and nasality that has already been well established after years of empirical research. Some of these acoustic parameters - like low frequency peaks and wider bandwidths- are common for both nasal and breathy voice. Therefore, it is likely that these parameters interact when a sound is coarticulated with breathiness and nasality. This leads to the hypothesis that the acoustic parameters, which usually act as robust cues in differentiating between breathy and modal voice, might not be reliable cues for differentiating between breathy and modal voice when breathiness is coarticulated with nasality. The effect of nasality on the perception of breathiness has been explored in earlier studies using synthesized speech. The results showed that perceptually, nasality and breathiness do interact. The current study investigates if a similar pattern is observed in natural speech. The study is conducted on Marathi, an Indo-Aryan language which has a three-way contrast between nasality and breathiness. That is, there is a phonemic distinction between nasals, breathy voice and breathy-nasals. Voice quality parameters like – H1-H2 (Difference between the amplitude of first and second harmonic), H1-A3 (Difference between the amplitude of first harmonic and third formant, CPP (Cepstral Peak Prominence), HNR (Harmonics to Noise ratio) and B1 (Bandwidth of first formant) were extracted. Statistical models like linear mixed effects regression and Random Forest classifiers show that measures that capture the noise component in the signal- like CPP and HNR- can classify breathy voice from modal voice better than spectral measures when breathy voice is coarticulated with nasality.Keywords: breathiness, marathi, nasality, voice quality
Procedia PDF Downloads 96