Search results for: quantile function model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19841

Search results for: quantile function model

19781 Modelling for Temperature Non-Isothermal Continuous Stirred Tank Reactor Using Fuzzy Logic

Authors: Nasser Mohamed Ramli, Mohamad Syafiq Mohamad

Abstract:

Many types of controllers were applied on the continuous stirred tank reactor (CSTR) unit to control the temperature. In this research paper, Proportional-Integral-Derivative (PID) controller are compared with Fuzzy Logic controller for temperature control of CSTR. The control system for temperature non-isothermal of a CSTR will produce a stable response curve to its set point temperature. A mathematical model of a CSTR using the most general operating condition was developed through a set of differential equations into S-function using MATLAB. The reactor model and S-function are developed using m.file. After developing the S-function of CSTR model, User-Defined functions are used to link to SIMULINK file. Results that are obtained from simulation and temperature control were better when using Fuzzy logic control compared to PID control.

Keywords: CSTR, temperature, PID, fuzzy logic

Procedia PDF Downloads 427
19780 The Effectiveness of Energy-related Tax in Curbing Transport-related Carbon Emissions: The Role of Green Finance and Technology in OECD Economies

Authors: Hassan Taimoor, Piotr Krajewski, Piotr Gabrielzcak

Abstract:

Being responsible for the largest source of energy-related emissions, the transportation sector is driven by more than half of global oil demand and total energy consumption, making it a crucial factor in tackling climate change and environmental degradation. The present study empirically tests the effectives of the energy-related tax (TXEN) in curbing transport-related carbon emissions (CO2TRANSP) in Organization for Economic Cooperation and Development (OECD) economies over the period of 1990-2020. Moreover, Green Finance (GF), Technology (TECH), and Gross domestic product (GDP) have also been added as explanatory factors which might affect CO2TRANSP emissions. The study employs the Method of Moment Quantile Regression (MMQR), an advance econometric technique to observe the variations along each quantile. Based on the results of the preliminary test, we confirm the presence of cross-sectional dependence and slope heterogeneity. Whereas the result of the panel unit root test report mixed order of variables’ integration. The findings reveal that rise in income level activates CO2TRANSP, confirming the first stage of Environmental Kuznet Hypothesis. Surprisingly, the present TXEN policies of OECD member states are not mature enough to tackle the CO2TRANSP emissions. However, the findings confirm that GF and TECH are solely responsible for the reduction in the CO2TRANSP. The outcomes of Bootstrap Quantile Regression (BSQR) further validate and support the earlier findings of MMQR. Based on the findings of this study, it is revealed that the current TXEN policies are too moderate, and an incremental and progressive rise in TXEN may help in a transition toward a cleaner and sustainable transportation sector in the study region.

Keywords: transport-related CO2 emissions, energy-related tax, green finance, technological development, oecd member states

Procedia PDF Downloads 58
19779 Application of Generalized Autoregressive Score Model to Stock Returns

Authors: Katleho Daniel Makatjane, Diteboho Lawrence Xaba, Ntebogang Dinah Moroke

Abstract:

The current study investigates the behaviour of time-varying parameters that are based on the score function of the predictive model density at time t. The mechanism to update the parameters over time is the scaled score of the likelihood function. The results revealed that there is high persistence of time-varying, as the location parameter is higher and the skewness parameter implied the departure of scale parameter from the normality with the unconditional parameter as 1.5. The results also revealed that there is a perseverance of the leptokurtic behaviour in stock returns which implies the returns are heavily tailed. Prior to model estimation, the White Neural Network test exposed that the stock price can be modelled by a GAS model. Finally, we proposed further researches specifically to model the existence of time-varying parameters with a more detailed model that encounters the heavy tail distribution of the series and computes the risk measure associated with the returns.

Keywords: generalized autoregressive score model, South Africa, stock returns, time-varying

Procedia PDF Downloads 481
19778 Confidence Envelopes for Parametric Model Selection Inference and Post-Model Selection Inference

Authors: I. M. L. Nadeesha Jayaweera, Adao Alex Trindade

Abstract:

In choosing a candidate model in likelihood-based modeling via an information criterion, the practitioner is often faced with the difficult task of deciding just how far up the ranked list to look. Motivated by this pragmatic necessity, we construct an uncertainty band for a generalized (model selection) information criterion (GIC), defined as a criterion for which the limit in probability is identical to that of the normalized log-likelihood. This includes common special cases such as AIC & BIC. The method starts from the asymptotic normality of the GIC for the joint distribution of the candidate models in an independent and identically distributed (IID) data framework and proceeds by deriving the (asymptotically) exact distribution of the minimum. The calculation of an upper quantile for its distribution then involves the computation of multivariate Gaussian integrals, which is amenable to efficient implementation via the R package "mvtnorm". The performance of the methodology is tested on simulated data by checking the coverage probability of nominal upper quantiles and compared to the bootstrap. Both methods give coverages close to nominal for large samples, but the bootstrap is two orders of magnitude slower. The methodology is subsequently extended to two other commonly used model structures: regression and time series. In the regression case, we derive the corresponding asymptotically exact distribution of the minimum GIC invoking Lindeberg-Feller type conditions for triangular arrays and are thus able to similarly calculate upper quantiles for its distribution via multivariate Gaussian integration. The bootstrap once again provides a default competing procedure, and we find that similar comparison performance metrics hold as for the IID case. The time series case is complicated by far more intricate asymptotic regime for the joint distribution of the model GIC statistics. Under a Gaussian likelihood, the default in most packages, one needs to derive the limiting distribution of a normalized quadratic form for a realization from a stationary series. Under conditions on the process satisfied by ARMA models, a multivariate normal limit is once again achieved. The bootstrap can, however, be employed for its computation, whence we are once again in the multivariate Gaussian integration paradigm for upper quantile evaluation. Comparisons of this bootstrap-aided semi-exact method with the full-blown bootstrap once again reveal a similar performance but faster computation speeds. One of the most difficult problems in contemporary statistical methodological research is to be able to account for the extra variability introduced by model selection uncertainty, the so-called post-model selection inference (PMSI). We explore ways in which the GIC uncertainty band can be inverted to make inferences on the parameters. This is being attempted in the IID case by pivoting the CDF of the asymptotically exact distribution of the minimum GIC. For inference one parameter at a time and a small number of candidate models, this works well, whence the attained PMSI confidence intervals are wider than the MLE-based Wald, as expected.

Keywords: model selection inference, generalized information criteria, post model selection, Asymptotic Theory

Procedia PDF Downloads 66
19777 Mixed Integer Programing for Multi-Tier Rebate with Discontinuous Cost Function

Authors: Y. Long, L. Liu, K. V. Branin

Abstract:

One challenge faced by procurement decision-maker during the acquisition process is how to compare similar products from different suppliers and allocate orders among different products or services. This work focuses on allocating orders among multiple suppliers considering rebate. The objective function is to minimize the total acquisition cost including purchasing cost and rebate benefit. Rebate benefit is complex and difficult to estimate at the ordering step. Rebate rules vary for different suppliers and usually change over time. In this work, we developed a system to collect the rebate policies, standardized the rebate policies and developed two-stage optimization models for ordering allocation. Rebate policy with multi-tiers is considered in modeling. The discontinuous cost function of rebate benefit is formulated for different scenarios. A piecewise linear function is used to approximate the discontinuous cost function of rebate benefit. And a Mixed Integer Programing (MIP) model is built for order allocation problem with multi-tier rebate. A case study is presented and it shows that our optimization model can reduce the total acquisition cost by considering rebate rules.

Keywords: discontinuous cost function, mixed integer programming, optimization, procurement, rebate

Procedia PDF Downloads 238
19776 Multinomial Dirichlet Gaussian Process Model for Classification of Multidimensional Data

Authors: Wanhyun Cho, Soonja Kang, Sanggoon Kim, Soonyoung Park

Abstract:

We present probabilistic multinomial Dirichlet classification model for multidimensional data and Gaussian process priors. Here, we have considered an efficient computational method that can be used to obtain the approximate posteriors for latent variables and parameters needed to define the multiclass Gaussian process classification model. We first investigated the process of inducing a posterior distribution for various parameters and latent function by using the variational Bayesian approximations and important sampling method, and next we derived a predictive distribution of latent function needed to classify new samples. The proposed model is applied to classify the synthetic multivariate dataset in order to verify the performance of our model. Experiment result shows that our model is more accurate than the other approximation methods.

Keywords: multinomial dirichlet classification model, Gaussian process priors, variational Bayesian approximation, importance sampling, approximate posterior distribution, marginal likelihood evidence

Procedia PDF Downloads 417
19775 Stability Analysis of a Human-Mosquito Model of Malaria with Infective Immigrants

Authors: Nisha Budhwar, Sunita Daniel

Abstract:

In this paper, we analyse the stability of the SEIR model of malaria with infective immigrants which was recently formulated by the authors. The model consists of an SEIR model for the human population and SI Model for the mosquitoes. Susceptible humans become infected after they are bitten by infectious mosquitoes and move on to the Exposed, Infected and Recovered classes respectively. The susceptible mosquito becomes infected after biting an infected person and remains infected till death. We calculate the reproduction number R0 using the next generation method and then discuss about the stability of the equilibrium points. We use the Lyapunov function to show the global stability of the equilibrium points.

Keywords: equilibrium points, exposed, global stability, infective immigrants, Lyapunov function, recovered, reproduction number, susceptible

Procedia PDF Downloads 338
19774 Best Resource Recommendation for a Stochastic Process

Authors: Likewin Thomas, M. V. Manoj Kumar, B. Annappa

Abstract:

The aim of this study was to develop an Artificial Neural Network0 s recommendation model for an online process using the complexity of load, performance, and average servicing time of the resources. Here, the proposed model investigates the resource performance using stochastic gradient decent method for learning ranking function. A probabilistic cost function is implemented to identify the optimal θ values (load) on each resource. Based on this result the recommendation of resource suitable for performing the currently executing task is made. The test result of CoSeLoG project is presented with an accuracy of 72.856%.

Keywords: ADALINE, neural network, gradient decent, process mining, resource behaviour, polynomial regression model

Procedia PDF Downloads 366
19773 Study of Inhibition of the End Effect Based on AR Model Predict of Combined Data Extension and Window Function

Authors: Pan Hongxia, Wang Zhenhua

Abstract:

In this paper, the EMD decomposition in the process of endpoint effect adopted data based on AR model to predict the continuation and window function method of combining the two effective inhibition. Proven by simulation of the simulation signal obtained the ideal effect, then, apply this method to the gearbox test data is also achieved good effect in the process, for the analysis of the subsequent data processing to improve the calculation accuracy. In the end, under various working conditions for the gearbox fault diagnosis laid a good foundation.

Keywords: gearbox, fault diagnosis, ar model, end effect

Procedia PDF Downloads 346
19772 Comparison of Volume of Fluid Model: Experimental and Empirical Results for Flows over Stacked Drop Manholes

Authors: Ramin Mansouri

Abstract:

The manhole is one of the types of structures that are installed at the site of change direction or change in the pipe diameter or sewage pipes as well as in step slope areas to reduce the flow velocity. In this study, the flow characteristics of hydraulic structures in a manhole structure have been investigated with a numerical model. In this research, the types of computational grid coarse, medium, and fines have been used for simulation. In order to simulate flow, k-ε model (standard, RNG, Realizable) and k-w model (standard SST) are used. Also, in order to find the best wall conditions, two types of standard and non-equilibrium wall functions were investigated. The turbulent model k-ε has the highest correlation with experimental results or all models. In terms of boundary conditions, constant speed is set for the flow input boundary, the output pressure is set in the boundaries which are in contact with the air, and the standard wall function is used for the effect of the wall function. In the numerical model, the depth at the output of the second manhole is estimated to be less than that of the laboratory and the output jet from the span. In the second regime, the jet flow collides with the manhole wall and divides into two parts, so hydraulic characteristics are the same as large vertical shaft hydraulic characteristics. In this situation, the turbulence is in a high range since it can be seen more energy loss in it. According to the results, energy loss in numerical is estimated at 9.359%, which is more than experimental data.

Keywords: manhole, energy, depreciation, turbulence model, wall function, flow

Procedia PDF Downloads 53
19771 Fecundity and Egg Laying in Helicoverpa armigera (Hübner) (Lepidoptera: Noctuidae): Model Development and Field Validation

Authors: Muhammad Noor Ul Ane, Dong-Soon Kim, Myron P. Zalucki

Abstract:

Models can be useful to help understand population dynamics of insects under diverse environmental conditions and in developing strategies to manage pest species better. Adult longevity and fecundity of Helicoverpa armigera (Hübner) were evaluated against a wide range of constant temperatures (15, 20, 25, 30, 35 and 37.5ᵒC). The modified Sharpe and DeMichele model described adult aging rate and was used to estimate adult physiological age. Maximum fecundity of H. armigera was 973 egg/female at 25ᵒC decreasing to 72 eggs/female at 37.5ᵒC. The relationship between adult fecundity and temperature was well described by an extreme value function. Age-specific cumulative oviposition rate and age-specific survival rate were well described by a two-parameter Weibull function and sigmoid function, respectively. An oviposition model was developed using three temperature-dependent components: total fecundity, age-specific oviposition rate, and age-specific survival rate. The oviposition model was validated against independent field data and described the field occurrence pattern of egg population of H. armigera very well. Our model should be a useful component for population modeling of H. armigera and can be independently used for the timing of sprays in management programs of this key pest species.

Keywords: cotton bollworm, life table, temperature-dependent adult development, temperature-dependent fecundity

Procedia PDF Downloads 129
19770 Application of ANN and Fuzzy Logic Algorithms for Runoff and Sediment Yield Modelling of Kal River, India

Authors: Mahesh Kothari, K. D. Gharde

Abstract:

The ANN and fuzzy logic (FL) models were developed to predict the runoff and sediment yield for catchment of Kal river, India using 21 years (1991 to 2011) rainfall and other hydrological data (evaporation, temperature and streamflow lag by one and two day) and 7 years data for sediment yield modelling. The ANN model performance improved with increasing the input vectors. The fuzzy logic model was performing with R value more than 0.95 during developmental stage and validation stage. The comparatively FL model found to be performing well to ANN in prediction of runoff and sediment yield for Kal river.

Keywords: transferred function, sigmoid, backpropagation, membership function, defuzzification

Procedia PDF Downloads 543
19769 Modelling Hydrological Time Series Using Wakeby Distribution

Authors: Ilaria Lucrezia Amerise

Abstract:

The statistical modelling of precipitation data for a given portion of territory is fundamental for the monitoring of climatic conditions and for Hydrogeological Management Plans (HMP). This modelling is rendered particularly complex by the changes taking place in the frequency and intensity of precipitation, presumably to be attributed to the global climate change. This paper applies the Wakeby distribution (with 5 parameters) as a theoretical reference model. The number and the quality of the parameters indicate that this distribution may be the appropriate choice for the interpolations of the hydrological variables and, moreover, the Wakeby is particularly suitable for describing phenomena producing heavy tails. The proposed estimation methods for determining the value of the Wakeby parameters are the same as those used for density functions with heavy tails. The commonly used procedure is the classic method of moments weighed with probabilities (probability weighted moments, PWM) although this has often shown difficulty of convergence, or rather, convergence to a configuration of inappropriate parameters. In this paper, we analyze the problem of the likelihood estimation of a random variable expressed through its quantile function. The method of maximum likelihood, in this case, is more demanding than in the situations of more usual estimation. The reasons for this lie, in the sampling and asymptotic properties of the estimators of maximum likelihood which improve the estimates obtained with indications of their variability and, therefore, their accuracy and reliability. These features are highly appreciated in contexts where poor decisions, attributable to an inefficient or incomplete information base, can cause serious damages.

Keywords: generalized extreme values, likelihood estimation, precipitation data, Wakeby distribution

Procedia PDF Downloads 122
19768 SLAMF5 Regulates Myeloid Cells Activation in the Eae Model

Authors: Laura Bellassen, Idit Shachar

Abstract:

Multiple sclerosis (MS) is a chronic neurological disorder characterized by demyelination of the central nervous system (CNS), leading to a wide range of physical and cognitive impairments. Myeloid cells in the CNS, such microglia and border associated macrophage cells, participate in the neuroinflammation in MS. Activation of those cells in MS contributes to the inflammatory response in the CNS and recruitment of immune cells in the this compartment. SLAMF5 is a cell surface receptor that functions as a homophilic adhesion molecule, whose signaling can activate or inhibit leukocyte function. In the current study we followed the expression and function of SLAMF5 in myeloid cells in the CNS and in the periphery in the murine model for MS, the experimental autoimmune encephalomyelitis model (EAE). Our results show that SLAMF5 deficiency or blocking decreases the expression of activation molecules and costimulatory molecules such as MHCII and CD80, resulting in delayed onset and reduced progression of the disease. Moreover, blocking SLAMF5 in peripheral monocytes derived from MS patients and iPSC-derived microglia cells, controls the expression of HLA-DR and CD80. Thus, SLAMF5 is a regulator of myeloid cells function and can serve as a therapeutic target in autoimmune disorders as Multiple Sclerosis.

Keywords: multiple sclerosis, EAE model, myeloid cells, new antibody, neuroimmunology

Procedia PDF Downloads 33
19767 Parametric Modeling for Survival Data with Competing Risks Using the Generalized Gompertz Distribution

Authors: Noora Al-Shanfari, M. Mazharul Islam

Abstract:

The cumulative incidence function (CIF) is a fundamental approach for analyzing survival data in the presence of competing risks, which estimates the marginal probability for each competing event. Parametric modeling of CIF has the advantage of fitting various shapes of CIF and estimates the impact of covariates with maximum efficiency. To calculate the total CIF's covariate influence using a parametric model., it is essential to parametrize the baseline of the CIF. As the CIF is an improper function by nature, it is necessary to utilize an improper distribution when applying parametric models. The Gompertz distribution, which is an improper distribution, is limited in its applicability as it only accounts for monotone hazard shapes. The generalized Gompertz distribution, however, can adapt to a wider range of hazard shapes, including unimodal, bathtub, and monotonic increasing or decreasing hazard shapes. In this paper, the generalized Gompertz distribution is used to parametrize the baseline of the CIF, and the parameters of the proposed model are estimated using the maximum likelihood approach. The proposed model is compared with the existing Gompertz model using the Akaike information criterion. Appropriate statistical test procedures and model-fitting criteria will be used to test the adequacy of the model. Both models are applied to the ‘colon’ dataset, which is available in the “biostat3” package in R.

Keywords: competing risks, cumulative incidence function, improper distribution, parametric modeling, survival analysis

Procedia PDF Downloads 67
19766 Novel Inference Algorithm for Gaussian Process Classification Model with Multiclass and Its Application to Human Action Classification

Authors: Wanhyun Cho, Soonja Kang, Sangkyoon Kim, Soonyoung Park

Abstract:

In this paper, we propose a novel inference algorithm for the multi-class Gaussian process classification model that can be used in the field of human behavior recognition. This algorithm can drive simultaneously both a posterior distribution of a latent function and estimators of hyper-parameters in a Gaussian process classification model with multi-class. Our algorithm is based on the Laplace approximation (LA) technique and variational EM framework. This is performed in two steps: called expectation and maximization steps. First, in the expectation step, using the Bayesian formula and LA technique, we derive approximately the posterior distribution of the latent function indicating the possibility that each observation belongs to a certain class in the Gaussian process classification model. Second, in the maximization step, using a derived posterior distribution of latent function, we compute the maximum likelihood estimator for hyper-parameters of a covariance matrix necessary to define prior distribution for latent function. These two steps iteratively repeat until a convergence condition satisfies. Moreover, we apply the proposed algorithm with human action classification problem using a public database, namely, the KTH human action data set. Experimental results reveal that the proposed algorithm shows good performance on this data set.

Keywords: bayesian rule, gaussian process classification model with multiclass, gaussian process prior, human action classification, laplace approximation, variational EM algorithm

Procedia PDF Downloads 311
19765 Online Robust Model Predictive Control for Linear Fractional Transformation Systems Using Linear Matrix Inequalities

Authors: Peyman Sindareh Esfahani, Jeffery Kurt Pieper

Abstract:

In this paper, the problem of robust model predictive control (MPC) for discrete-time linear systems in linear fractional transformation form with structured uncertainty and norm-bounded disturbance is investigated. The problem of minimization of the cost function for MPC design is converted to minimization of the worst case of the cost function. Then, this problem is reduced to minimization of an upper bound of the cost function subject to a terminal inequality satisfying the l2-norm of the closed loop system. The characteristic of the linear fractional transformation system is taken into account, and by using some mathematical tools, the robust predictive controller design problem is turned into a linear matrix inequality minimization problem. Afterwards, a formulation which includes an integrator to improve the performance of the proposed robust model predictive controller in steady state condition is studied. The validity of the approaches is illustrated through a robust control benchmark problem.

Keywords: linear fractional transformation, linear matrix inequality, robust model predictive control, state feedback control

Procedia PDF Downloads 375
19764 Understanding Consumption Planning Behaviors

Authors: Gaosheng Ju

Abstract:

Our empirical evidence supports a model of consumption planning behaviors with the following two characteristics. First, households formulate a rational consumption target based on their desired target, displaying a diminishing sensitivity to the discrepancy between them. Second, the established target is a reference point for their planned consumption. The diminishing sensitivity leads to opposite reactions in higher and lower quantiles of both consumption targets and consumption growth to changes in economic conditions. This phenomenon accounts for the perplexingly low correlation between consumption and other macroeconomic variables. Furthermore, the opposing movements of consumption targets offer new insights into consumption-based asset pricing.

Keywords: consumption planning, reference point, diminishing sensitivity, quantile regression, asset pricing puzzles

Procedia PDF Downloads 54
19763 Analysis of the Contribution of Drude and Brendel Model Terms to the Dielectric Function

Authors: Christopher Mkirema Maghanga, Maurice Mghendi Mwamburi

Abstract:

Parametric modeling provides a means to deeper understand the properties of materials. Drude, Brendel, Lorentz and OJL incorporated in SCOUT® software are some of the models used to study dielectric films. In our work, we utilized Brendel and Drude models to extract the optical constants from spectroscopic data of fabricated undoped and niobium doped titanium oxide thin films. The individual contributions by the two models were studied to establish how they influence the dielectric function. The effect of dopants on their influences was also analyzed. For the undoped films, results indicate minimal contribution from the Drude term due to the dielectric nature of the films. However as doping levels increase, the rise in the concentration of free electrons favors the use of Drude model. Brendel model was confirmed to work well with dielectric films - the undoped titanium Oxide films in our case.

Keywords: modeling, Brendel model, optical constants, titanium oxide, Drude Model

Procedia PDF Downloads 160
19762 A Stochastic Diffusion Process Based on the Two-Parameters Weibull Density Function

Authors: Meriem Bahij, Ahmed Nafidi, Boujemâa Achchab, Sílvio M. A. Gama, José A. O. Matos

Abstract:

Stochastic modeling concerns the use of probability to model real-world situations in which uncertainty is present. Therefore, the purpose of stochastic modeling is to estimate the probability of outcomes within a forecast, i.e. to be able to predict what conditions or decisions might happen under different situations. In the present study, we present a model of a stochastic diffusion process based on the bi-Weibull distribution function (its trend is proportional to the bi-Weibull probability density function). In general, the Weibull distribution has the ability to assume the characteristics of many different types of distributions. This has made it very popular among engineers and quality practitioners, who have considered it the most commonly used distribution for studying problems such as modeling reliability data, accelerated life testing, and maintainability modeling and analysis. In this work, we start by obtaining the probabilistic characteristics of this model, as the explicit expression of the process, its trends, and its distribution by transforming the diffusion process in a Wiener process as shown in the Ricciaardi theorem. Then, we develop the statistical inference of this model using the maximum likelihood methodology. Finally, we analyse with simulated data the computational problems associated with the parameters, an issue of great importance in its application to real data with the use of the convergence analysis methods. Overall, the use of a stochastic model reflects only a pragmatic decision on the part of the modeler. According to the data that is available and the universe of models known to the modeler, this model represents the best currently available description of the phenomenon under consideration.

Keywords: diffusion process, discrete sampling, likelihood estimation method, simulation, stochastic diffusion process, trends functions, bi-parameters weibull density function

Procedia PDF Downloads 285
19761 A Rapid Prototyping Tool for Suspended Biofilm Growth Media

Authors: Erifyli Tsagkari, Stephanie Connelly, Zhaowei Liu, Andrew McBride, William Sloan

Abstract:

Biofilms play an essential role in treating water in biofiltration systems. The biofilm morphology and function are inextricably linked to the hydrodynamics of flow through a filter, and yet engineers rarely explicitly engineer this interaction. We develop a system that links computer simulation and 3-D printing to optimize and rapidly prototype filter media to optimize biofilm function with the hypothesis that biofilm function is intimately linked to the flow passing through the filter. A computational model that numerically solves the incompressible time-dependent Navier Stokes equations coupled to a model for biofilm growth and function is developed. The model is imbedded in an optimization algorithm that allows the model domain to adapt until criteria on biofilm functioning are met. This is applied to optimize the shape of filter media in a simple flow channel to promote biofilm formation. The computer code links directly to a 3-D printer, and this allows us to prototype the design rapidly. Its validity is tested in flow visualization experiments and by microscopy. As proof of concept, the code was constrained to explore a small range of potential filter media, where the medium acts as an obstacle in the flow that sheds a von Karman vortex street that was found to enhance the deposition of bacteria on surfaces downstream. The flow visualization and microscopy in the 3-D printed realization of the flow channel validated the predictions of the model and hence its potential as a design tool. Overall, it is shown that the combination of our computational model and the 3-D printing can be effectively used as a design tool to prototype filter media to optimize biofilm formation.

Keywords: biofilm, biofilter, computational model, von karman vortices, 3-D printing.

Procedia PDF Downloads 121
19760 Structural Damage Detection Using Modal Data Employing Teaching Learning Based Optimization

Authors: Subhajit Das, Nirjhar Dhang

Abstract:

Structural damage detection is a challenging work in the field of structural health monitoring (SHM). The damage detection methods mainly focused on the determination of the location and severity of the damage. Model updating is a well known method to locate and quantify the damage. In this method, an error function is defined in terms of difference between the signal measured from ‘experiment’ and signal obtained from undamaged finite element model. This error function is minimised with a proper algorithm, and the finite element model is updated accordingly to match the measured response. Thus, the damage location and severity can be identified from the updated model. In this paper, an error function is defined in terms of modal data viz. frequencies and modal assurance criteria (MAC). MAC is derived from Eigen vectors. This error function is minimized by teaching-learning-based optimization (TLBO) algorithm, and the finite element model is updated accordingly to locate and quantify the damage. Damage is introduced in the model by reduction of stiffness of the structural member. The ‘experimental’ data is simulated by the finite element modelling. The error due to experimental measurement is introduced in the synthetic ‘experimental’ data by adding random noise, which follows Gaussian distribution. The efficiency and robustness of this method are explained through three examples e.g., one truss, one beam and one frame problem. The result shows that TLBO algorithm is efficient to detect the damage location as well as the severity of damage using modal data.

Keywords: damage detection, finite element model updating, modal assurance criteria, structural health monitoring, teaching learning based optimization

Procedia PDF Downloads 196
19759 Prediction of Phonon Thermal Conductivity of F.C.C. Al by Molecular Dynamics Simulation

Authors: Leila Momenzadeh, Alexander V. Evteev, Elena V. Levchenko, Tanvir Ahmed, Irina Belova, Graeme Murch

Abstract:

In this work, the phonon thermal conductivity of f.c.c. Al is investigated in detail in the temperature range 100 – 900 K within the framework of equilibrium molecular dynamics simulations making use of the Green-Kubo formalism and one of the most reliable embedded-atom method potentials. It is found that the heat current auto-correlation function of the f.c.c. Al model demonstrates a two-stage temporal decay similar to the previously observed for f.c.c Cu model. After the first stage of decay, the heat current auto-correlation function of the f.c.c. Al model demonstrates a peak in the temperature range 100-800 K. The intensity of the peak decreases as the temperature increases. At 900 K, it transforms to a shoulder. To describe the observed two-stage decay of the heat current auto-correlation function of the f.c.c. Al model, we employ decomposition model recently developed for phonon-mediated thermal transport in a monoatomic lattice. We found that the electronic contribution to the total thermal conductivity of f.c.c. Al dominates over the whole studied temperature range. However, the phonon contribution to the total thermal conductivity of f.c.c. Al increases as temperature decreases. It is about 1.05% at 900 K and about 12.5% at 100 K.

Keywords: aluminum, gGreen-Kubo formalism, molecular dynamics, phonon thermal conductivity

Procedia PDF Downloads 400
19758 Generalized Additive Model Approach for the Chilean Hake Population in a Bio-Economic Context

Authors: Selin Guney, Andres Riquelme

Abstract:

The traditional bio-economic method for fisheries modeling uses some estimate of the growth parameters and the system carrying capacity from a biological model for the population dynamics (usually a logistic population growth model) which is then analyzed as a traditional production function. The stock dynamic is transformed into a revenue function and then compared with the extraction costs to estimate the maximum economic yield. In this paper, the logistic population growth model for the population is combined with a forecast of the abundance and location of the stock by using a generalized additive model approach. The paper focuses on the Chilean hake population. This method allows for the incorporation of climatic variables and the interaction with other marine species, which in turn will increase the reliability of the estimates and generate better extraction paths for different conservation objectives, such as the maximum biological yield or the maximum economic yield.

Keywords: bio-economic, fisheries, GAM, production

Procedia PDF Downloads 230
19757 Estimation of Fuel Cost Function Characteristics Using Cuckoo Search

Authors: M. R. Al-Rashidi, K. M. El-Naggar, M. F. Al-Hajri

Abstract:

The fuel cost function describes the electric power generation-cost relationship in thermal plants, hence, it sheds light on economical aspects of power industry. Different models have been proposed to describe this relationship with the quadratic function model being the most popular one. Parameters of second order fuel cost function are estimated in this paper using cuckoo search algorithm. It is a new population based meta-heuristic optimization technique that has been used in this study primarily as an accurate estimation tool. Its main features are flexibility, simplicity, and effectiveness when compared to other estimation techniques. The parameter estimation problem is formulated as an optimization one with the goal being minimizing the error associated with the estimated parameters. A case study is considered in this paper to illustrate cuckoo search promising potential as a valuable estimation and optimization technique.

Keywords: cuckoo search, parameters estimation, fuel cost function, economic dispatch

Procedia PDF Downloads 554
19756 Steady State Charge Transport in Quantum Dots: Nonequilibrium Green's Function (NEGF) vs. Single Electron Analysis

Authors: Mahesh Koti

Abstract:

In this paper, we present a quantum transport study of a quantum dot in steady state in the presence of static gate potential. We consider a quantum dot coupled to the two metallic leads. The quantum dot under study is modeled through Anderson Impurity Model (AIM) with hopping parameter modulated through voltage drop between leads and the central dot region. Based on the Landauer's formula derived from Nonequilibrium Green's Function and Single Electron Theory, the essential ingredients of transport properties are revealed. We show that the results out of two approaches closely agree with each other. We demonstrate that Landauer current response derived from single electron approach converges with non-zero interaction through gate potential whereas Landauer current response derived from Nonequilibrium Green's Function (NEGF) hits a pole.

Keywords: Anderson impurity model (AIM), nonequilibrium Green's function (NEGF), Landauer's formula, single electron analysis

Procedia PDF Downloads 451
19755 Study on Optimal Control Strategy of PM2.5 in Wuhan, China

Authors: Qiuling Xie, Shanliang Zhu, Zongdi Sun

Abstract:

In this paper, we analyzed the correlation relationship among PM2.5 from other five Air Quality Indices (AQIs) based on the grey relational degree, and built a multivariate nonlinear regression equation model of PM2.5 and the five monitoring indexes. For the optimal control problem of PM2.5, we took the partial large Cauchy distribution of membership equation as satisfaction function. We established a nonlinear programming model with the goal of maximum performance to price ratio. And the optimal control scheme is given.

Keywords: grey relational degree, multiple linear regression, membership function, nonlinear programming

Procedia PDF Downloads 278
19754 On a Univalent Function and the Integral Means of Its Derivative

Authors: Shatha S. Alhily

Abstract:

The purpose of this research paper is to show all the possible values of the pth power of the integrable function which make the integral means of the derivative of univalent function existing and finite.

Keywords: derivative, integral means, self conformal maps, univalent function

Procedia PDF Downloads 611
19753 Structure Function and Violation of Scale Invariance in NCSM: Theory and Numerical Analysis

Authors: M. R. Bekli, N. Mebarki, I. Chadou

Abstract:

In this study, we focus on the structure functions and violation of scale invariance in the context of non-commutative standard model (NCSM). We find that this violation appears in the first order of perturbation theory and a non-commutative version of the DGLAP evolution equation is deduced. Numerical analysis and comparison with experimental data imposes a new bound on the non-commutative parameter.

Keywords: NCSM, structure function, DGLAP equation, standard model

Procedia PDF Downloads 594
19752 Acoustic Room Impulse Response Computation with Image Sources and Frequency Dependent Boundary Reflection Coefficients

Authors: Pratik Gandhi, Kavitha Chandra, Charles Thompson

Abstract:

A computational model of the acoustic room impulse response between transmitters and receivers located in an enclosed cavity under the influence of frequency-dependent reflection coefficients of the walls is presented. The characteristic features of the impulse responses that differentiate these results from frequency-independent reflecting surfaces are discussed. The image-source model is derived from the first principle solution to Green's function of the acoustic wave equation. The post-processing of the computed impulse response with a band-pass filter to better represents the response of a loud-speaker is demonstrated.

Keywords: acoustic room impulse response, frequency dependent reflection coefficients, Green's function, image model

Procedia PDF Downloads 210