Search results for: model order reduction.
8831 Emission Constrained Hydrothermal Scheduling Algorithm
Authors: Sayeed Salam
Abstract:
This paper presents an efficient emission constrained hydrothermal scheduling algorithm that deals with nonlinear functions such as the water discharge characteristics, thermal cost, and transmission loss. It is then incorporated into the hydrothermal coordination program. The program has been tested on a practical utility system having 32 thermal and 12 hydro generating units. Test results show that a slight increase in production cost causes a substantial reduction in emission.Keywords: Emission constraint, Hydrothermal coordination, and Hydrothermal scheduling algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13728830 Implicit Two Step Continuous Hybrid Block Methods with Four Off-Steps Points for Solving Stiff Ordinary Differential Equation
Authors: O. A. Akinfenwa, N.M. Yao, S. N. Jator
Abstract:
In this paper, a self starting two step continuous block hybrid formulae (CBHF) with four Off-step points is developed using collocation and interpolation procedures. The CBHF is then used to produce multiple numerical integrators which are of uniform order and are assembled into a single block matrix equation. These equations are simultaneously applied to provide the approximate solution for the stiff ordinary differential equations. The order of accuracy and stability of the block method is discussed and its accuracy is established numerically.Keywords: Collocation and Interpolation, Continuous HybridBlock Formulae, Off-Step Points, Stability, Stiff ODEs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20988829 Performance, Emission and Combustion Characteristics of a Variable Compression Ratio Diesel Engine Fueled with Karanj Biodiesel and Its Blends
Authors: Ajay V. Kolhe, R. E. Shelke, S. S. Khandare
Abstract:
The use of biodiesel in conventional diesel engines results in substantial reduction of unburned hydrocarbon, carbon monoxide and particulate matters. The performance, emission and combustion characteristics of a single cylinder four stroke variable compression ratio engine when fueled with Karanja (Pongamia) methyl ester and its 10-50 % blends with diesel (on a volume basis) are investigated and compared with standard diesel. The suitability of karanja methyl ester as a biofuel has been established in this study. The useful brake power obtained is similar to diesel fuel for all loads. Experiment has been conducted at a fixed engine speed of 1500 rpm, variable load and at compression ratios of 17.5:1 and 18.5:1. The impact of compression ratio on fuel consumption, combustion pressures and exhaust gas emissions has been investigated and presented. Optimum compression ratio which gives best performance has been identified. The results indicate longer ignition delay, maximum rate of pressure rise, lower heat release rate and higher mass fraction burnt at higher compression ratio for pongamia oil methyl ester when compared to that of diesel. The brake thermal efficiency for pongamia oil methyl ester blends and diesel has been calculated and the blend B20 is found to give maximum thermal efficiency. The blends when used as fuel results in reduction of carbon monoxide, hydrocarbon and increase in nitrogen oxides emissions. PME as an oxygenated fuel generated more complete combustion, which means increased torque and power. This is also supported with higher thermal efficiencies of the PME blends. NOx is slightly increased due to the higher combustion temperature and the presence of fuel oxygen with the blend at full load. PME as a new Biodiesel and its blends can be used in diesel engines without any engine modification.
Keywords: Variable compression ratio CI engine, performance, combustion, emissions, biodiesel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32998828 Adaption Model for Building Agile Pronunciation Dictionaries Using Phonemic Distance Measurements
Authors: Akella Amarendra Babu, Rama Devi Yellasiri, Natukula Sainath
Abstract:
Where human beings can easily learn and adopt pronunciation variations, machines need training before put into use. Also humans keep minimum vocabulary and their pronunciation variations are stored in front-end of their memory for ready reference, while machines keep the entire pronunciation dictionary for ready reference. Supervised methods are used for preparation of pronunciation dictionaries which take large amounts of manual effort, cost, time and are not suitable for real time use. This paper presents an unsupervised adaptation model for building agile and dynamic pronunciation dictionaries online. These methods mimic human approach in learning the new pronunciations in real time. A new algorithm for measuring sound distances called Dynamic Phone Warping is presented and tested. Performance of the system is measured using an adaptation model and the precision metrics is found to be better than 86 percent.Keywords: Pronunciation variations, dynamic programming, machine learning, natural language processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8008827 Coding of DWT Coefficients using Run-length Coding and Huffman Coding for the Purpose of Color Image Compression
Authors: Varun Setia, Vinod Kumar
Abstract:
In present paper we proposed a simple and effective method to compress an image. Here we found success in size reduction of an image without much compromising with it-s quality. Here we used Haar Wavelet Transform to transform our original image and after quantization and thresholding of DWT coefficients Run length coding and Huffman coding schemes have been used to encode the image. DWT is base for quite populate JPEG 2000 technique.
Keywords: Lossy compression, DWT, quantization, Run length coding, Huffman coding, JPEG2000.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29228826 Analysis of Building Response from Vertical Ground Motions
Authors: George C. Yao, Chao-Yu Tu, Wei-Chung Chen, Fung-Wen Kuo, Yu-Shan Chang
Abstract:
Building structures are subjected to both horizontal and vertical ground motions during earthquakes, but only the horizontal ground motion has been extensively studied and considered in design. Most of the prevailing seismic codes assume the vertical component to be 1/2 to 2/3 of the horizontal one. In order to understand the building responses from vertical ground motions, many earthquakes records are studied in this paper. System identification methods (ARX Model) are used to analyze the strong motions and to find out the characteristics of the vertical amplification factors and the natural frequencies of buildings. Analysis results show that the vertical amplification factors for high-rise buildings and low-rise building are 1.78 and 2.52 respectively, and the average vertical amplification factor of all buildings is about 2. The relationship between the vertical natural frequency and building height was regressed to a suggested formula in this study. The result points out an important message; the taller the building is, the greater chance of resonance of vertical vibration on the building will be.
Keywords: Vertical ground motion, vertical amplification factor, natural frequency, component.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10648825 A Study of Neuro-Fuzzy Inference System for Gross Domestic Product Growth Forecasting
Authors: Ε. Giovanis
Abstract:
In this paper we present a Adaptive Neuro-Fuzzy System (ANFIS) with inputs the lagged dependent variable for the prediction of Gross domestic Product growth rate in six countries. We compare the results with those of Autoregressive (AR) model. We conclude that the forecasting performance of neuro-fuzzy-system in the out-of-sample period is much more superior and can be a very useful alternative tool used by the national statistical services and the banking and finance industry.Keywords: Autoregressive model, Forecasting, Gross DomesticProduct, Neuro-Fuzzy
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16038824 Linear Pocket Profile based Threshold Voltage Model for sub-100 nm n-MOSFET
Authors: Muhibul Haque Bhuyan, Quazi Deen Mohd Khosru
Abstract:
This paper presents a threshold voltage model of pocket implanted sub-100 nm n-MOSFETs incorporating the drain and substrate bias effects using two linear pocket profiles. Two linear equations are used to simulate the pocket profiles along the channel at the surface from the source and drain edges towards the center of the n-MOSFET. Then the effective doping concentration is derived and is used in the threshold voltage equation that is obtained by solving the Poisson-s equation in the depletion region at the surface. Simulated threshold voltages for various gate lengths fit well with the experimental data already published in the literature. The simulated result is compared with the two other pocket profiles used to derive the threshold voltage models of n-MOSFETs. The comparison shows that the linear model has a simple compact form that can be utilized to study and characterize the pocket implanted advanced ULSI devices.
Keywords: Linear pocket profile, pocket implantation, nMOSFET, threshold voltage, short channel effect (SCE), reverse short channeleffect (RSCE).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18018823 A Materialized Approach to the Integration of XML Documents: the OSIX System
Authors: H. Ahmad, S. Kermanshahani, A. Simonet, M. Simonet
Abstract:
The data exchanged on the Web are of different nature from those treated by the classical database management systems; these data are called semi-structured data since they do not have a regular and static structure like data found in a relational database; their schema is dynamic and may contain missing data or types. Therefore, the needs for developing further techniques and algorithms to exploit and integrate such data, and extract relevant information for the user have been raised. In this paper we present the system OSIX (Osiris based System for Integration of XML Sources). This system has a Data Warehouse model designed for the integration of semi-structured data and more precisely for the integration of XML documents. The architecture of OSIX relies on the Osiris system, a DL-based model designed for the representation and management of databases and knowledge bases. Osiris is a viewbased data model whose indexing system supports semantic query optimization. We show that the problem of query processing on a XML source is optimized by the indexing approach proposed by Osiris.Keywords: Data integration, semi-structured data, views, XML.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15908822 Using “Eckel” Model to Measure Income Smoothing Practices: The Case of French Companies
Authors: Feddaoui Amina
Abstract:
Income smoothing represents an attempt on the part of the company's management to reduce variations in earnings through the manipulation of the accounting principles. In this study, we aimed to measure income smoothing practices in a sample of 30 French joint stock companies during the period (2007-2009), we used Dummy variables method and “ECKEL” model to measure income smoothing practices and Binomial test accourding to SPSS program, to confirm or refute our hypothesis. This study concluded that there are no significant statistical indicators of income smoothing practices in the sample studied of French companies during the period (2007-2009), so the income series in the same sample studied of is characterized by stability and non-volatility without any intervention of management through accounting manipulation. However, this type of accounting manipulation should be taken into account and efforts should be made by control bodies to apply Eckel model and generalize its use at the global level.
Keywords: Income, smoothing, “Eckel”, French companies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10048821 CLASS, A New Tool for Nuclear Scenarios: Description and First Application
Authors: B. Mouginot, J.B. Clavel, N Thiolliere
Abstract:
The presented work is motivated by a french law regarding nuclear waste management. In order to avoid the limitation coming with the usage of the existing scenario codes, as COSI, VISION or FAMILY, the Core Library for Advance Scenario Simulation (CLASS) is being develop. CLASS is an open source tool, which allows any user to simulate an electronuclear scenario. The main CLASS asset, is the possibility to include any type of reactor, even a complitely new concept, through the generation of its ACSII evolution database. In the present article, the CLASS working basis will be presented as well as a simple exemple in order to show his potentiel. In the considered exemple, the effect of the transmutation will be assessed on Minor Actinide Inventory produced by PWR reactors.
Keywords: Electronuclear scenario, reactor, simulation, nuclear waste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16838820 Green Function and Eshelby Tensor Based on Mindlin’s 2nd Gradient Model: An Explicit Study of Spherical Inclusion Case
Authors: A. Selmi, A. Bisharat
Abstract:
Using Fourier transform and based on the Mindlin's 2nd gradient model that involves two length scale parameters, the Green's function, the Eshelby tensor, and the Eshelby-like tensor for a spherical inclusion are derived. It is proved that the Eshelby tensor consists of two parts; the classical Eshelby tensor and a gradient part including the length scale parameters which enable the interpretation of the size effect. When the strain gradient is not taken into account, the obtained Green's function and Eshelby tensor reduce to its analogue based on the classical elasticity. The Eshelby tensor in and outside the inclusion, the volume average of the gradient part and the Eshelby-like tensor are explicitly obtained. Unlike the classical Eshelby tensor, the results show that the components of the new Eshelby tensor vary with the position and the inclusion dimensions. It is demonstrated that the contribution of the gradient part should not be neglected.
Keywords: Eshelby tensor, Eshelby-like tensor, Green’s function, Mindlin’s 2nd gradient model, Spherical inclusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7248819 A New Quantile Based Fuzzy Time Series Forecasting Model
Authors: Tahseen A. Jilani, Aqil S. Burney, C. Ardil
Abstract:
Time series models have been used to make predictions of academic enrollments, weather, road accident, casualties and stock prices, etc. Based on the concepts of quartile regression models, we have developed a simple time variant quantile based fuzzy time series forecasting method. The proposed method bases the forecast using prediction of future trend of the data. In place of actual quantiles of the data at each point, we have converted the statistical concept into fuzzy concept by using fuzzy quantiles using fuzzy membership function ensemble. We have given a fuzzy metric to use the trend forecast and calculate the future value. The proposed model is applied for TAIFEX forecasting. It is shown that proposed method work best as compared to other models when compared with respect to model complexity and forecasting accuracy.
Keywords: Quantile Regression, Fuzzy time series, fuzzy logicalrelationship groups, heuristic trend prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19978818 Granger Causal Nexus between Financial Development and Energy Consumption: Evidence from Cross Country Panel Data
Authors: Rudra P. Pradhan
Abstract:
This paper examines the Granger causal nexus between financial development and energy consumption in the group of 35 Financial Action Task Force (FATF) Countries over the period 1988-2012. The study uses two financial development indicators such as private sector credit and stock market capitalization and seven energy consumption indicators such as coal, oil, gas, electricity, hydro-electrical, nuclear and biomass. Using panel cointegration tests, the study finds that financial development and energy consumption are cointegrated, indicating the presence of a long-run relationship between the two. Using a panel vector error correction model (VECM), the study detects both bidirectional and unidirectional causality between financial development and energy consumption. The variation of this causality is due to the use of different proxies for both financial development and energy consumption. The policy implication of this study is that economic policies should recognize the differences in the financial development-energy consumption nexus in order to maintain sustainable development in the selected 35 FATF countries.Keywords: Financial development, energy consumption, Panel VECM, FATF countries.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15138817 Comparing the Efficiency of Simpson’s 1/3 and 3/8 Rules for the Numerical Solution of First Order Volterra Integro-Differential Equations
Authors: N. M. Kamoh, D. G. Gyemang, M. C. Soomiyol
Abstract:
This paper compared the efficiency of Simpson’s 1/3 and 3/8 rules for the numerical solution of first order Volterra integro-differential equations. In developing the solution, collocation approximation method was adopted using the shifted Legendre polynomial as basis function. A block method approach is preferred to the predictor corrector method for being self-starting. Experimental results confirmed that the Simpson’s 3/8 rule is more efficient than the Simpson’s 1/3 rule.
Keywords: Collocation shifted Legendre polynomials, Simpson’s rule and Volterra integro-differential equations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9758816 Wicking and Evaporation of Liquids in Knitted Fabrics: Analytic Solution of Capillary Rise Restrained by Gravity and Evaporation
Authors: N. S. Achour, M. Hamdaoui, S. Ben Nasrallah
Abstract:
Wicking and evaporation of water in porous knitted fabrics is investigated by combining experimental and analytical approaches: The standard wicking model from Lucas and Washburn is enhanced to account for evaporation and gravity effects. The goal is to model the effect of gravity and evaporation on wicking using simple analytical expressions and investigate the influence of fabrics geometrical parameters, such as porosity and thickness on evaporation impact on maximum reachable height values. The results show that fabric properties have a significant influence on evaporation effect. In this paper, an experimental study of determining water kinetics from different knitted fabrics were gravimetrically investigated permitting the measure of the mass and the height of liquid rising in fabrics in various atmospheric conditions. From these measurements, characteristic pore parameters (capillary radius and permeability) can be determined.Keywords: Evaporation, experimental study, geometrical parameters, model, porous knitted fabrics, wicking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20798815 Application of Modified Maxwell-Stefan Equation for Separation of Aqueous Phenol by Pervaporation
Authors: Ujjal K Ghosh, Ling Teen
Abstract:
Pervaporation has the potential to be an alternative to the other traditional separation processes such as distillation, adsorption, reverse osmosis and extraction. This study investigates the separation of phenol from water using a polyurethane membrane by pervaporation by applying the modified Maxwell-Stephen model. The modified Maxwell-Stefan model takes into account the non-ideal multi-component solubility effect, nonideal diffusivity of all permeating components, concentration dependent density of the membrane and diffusion coupling to predict various fluxes. Four cases has been developed to investigate the process parameters effects on the flux and weight fraction of phenol in the permeate values namely feed concentration, membrane thickness, operating temperature and operating downstream pressure. The model could describe semi-quantitatively the performance of the pervaporation membrane for the given system as a very good agreement between the observed and theoretical fluxes was observed.
Keywords: Pervaporation, Phenol, Polyurethane, Modified Maxwell-Stefan equation, Solution Diffusion
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27408814 Yield Prediction Using Support Vectors Based Under-Sampling in Semiconductor Process
Authors: Sae-Rom Pak, Seung Hwan Park, Jeong Ho Cho, Daewoong An, Cheong-Sool Park, Jun Seok Kim, Jun-Geol Baek
Abstract:
It is important to predict yield in semiconductor test process in order to increase yield. In this study, yield prediction means finding out defective die, wafer or lot effectively. Semiconductor test process consists of some test steps and each test includes various test items. In other world, test data has a big and complicated characteristic. It also is disproportionably distributed as the number of data belonging to FAIL class is extremely low. For yield prediction, general data mining techniques have a limitation without any data preprocessing due to eigen properties of test data. Therefore, this study proposes an under-sampling method using support vector machine (SVM) to eliminate an imbalanced characteristic. For evaluating a performance, randomly under-sampling method is compared with the proposed method using actual semiconductor test data. As a result, sampling method using SVM is effective in generating robust model for yield prediction.
Keywords: Yield Prediction, Semiconductor Test Process, Support Vector Machine, Under Sampling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23988813 Optimization of Thermal and Discretization Parameters in Laser Welding Simulation Nd:YAG Applied for Shin Plate Transparent Mode Of DP600
Authors: Chansopheak Seang, Afia David Kouadri, Eric Ragneau
Abstract:
Three dimensional analysis of thermal model in laser full penetration welding, Nd:YAG, by transparent mode DP600 alloy steel 1.25mm of thickness and gap of 0.1mm. Three models studied the influence of thermal dependent temperature properties, thermal independent temperature and the effect of peak value of specific heat at phase transformation temperature, AC1, on the transient temperature. Another seven models studied the influence of discretization, meshes on the temperature distribution in weld plate. It is shown that for the effects of thermal properties, the errors less 4% of maximum temperature in FZ and HAZ have identified. The minimum value of discretization are at least one third increment per radius for temporal discretization and the spatial discretization requires two elements per radius and four elements through thickness of the assembled plate, which therefore represent the minimum requirements of modeling for the laser welding in order to get minimum errors less than 5% compared to the fine mesh.Keywords: FEA, welding, discretization, ABAQUS user subroutine DFLUX
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18188812 Development of a Project Selection Method on Information System Using ANP and Fuzzy Logic
Authors: Ingu Kim, Shangmun Shin, Yongsun Choi, Nguyen Manh Thang, Edwin R. Ramos, Won-Joo Hwang
Abstract:
Project selection problems on management information system (MIS) are often considered a multi-criteria decision-making (MCDM) for a solving method. These problems contain two aspects, such as interdependencies among criteria and candidate projects and qualitative and quantitative factors of projects. However, most existing methods reported in literature consider these aspects separately even though these two aspects are simultaneously incorporated. For this reason, we proposed a hybrid method using analytic network process (ANP) and fuzzy logic in order to represent both aspects. We then propose a goal programming model to conduct an optimization for the project selection problems interpreted by a hybrid concept. Finally, a numerical example is conducted as verification purposes.Keywords: Analytic Network Process (ANP), Multi-Criteria Decision-Making (MCDM), Fuzzy Logic, Information System Project Selection, Goal Programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20898811 Tests for Gaussianity of a Stationary Time Series
Authors: Adnan Al-Smadi
Abstract:
One of the primary uses of higher order statistics in signal processing has been for detecting and estimation of non- Gaussian signals in Gaussian noise of unknown covariance. This is motivated by the ability of higher order statistics to suppress additive Gaussian noise. In this paper, several methods to test for non- Gaussianity of a given process are presented. These methods include histogram plot, kurtosis test, and hypothesis testing using cumulants and bispectrum of the available sequence. The hypothesis testing is performed by constructing a statistic to test whether the bispectrum of the given signal is non-zero. A zero bispectrum is not a proof of Gaussianity. Hence, other tests such as the kurtosis test should be employed. Examples are given to demonstrate the performance of the presented methods.Keywords: Non-Gaussian, bispectrum, kurtosis, hypothesistesting, histogram.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19168810 Confidence Intervals for Double Exponential Distribution: A Simulation Approach
Authors: M. Alrasheedi
Abstract:
The double exponential model (DEM), or Laplace distribution, is used in various disciplines. However, there are issues related to the construction of confidence intervals (CI), when using the distribution.In this paper, the properties of DEM are considered with intention of constructing CI based on simulated data. The analysis of pivotal equations for the models here in comparisons with pivotal equations for normal distribution are performed, and the results obtained from simulation data are presented.Keywords: Confidence intervals, double exponential model, pivotal equations, simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35528809 Estimating Shortest Circuit Path Length Complexity
Authors: Azam Beg, P. W. Chandana Prasad, S.M.N.A Senenayake
Abstract:
When binary decision diagrams are formed from uniformly distributed Monte Carlo data for a large number of variables, the complexity of the decision diagrams exhibits a predictable relationship to the number of variables and minterms. In the present work, a neural network model has been used to analyze the pattern of shortest path length for larger number of Monte Carlo data points. The neural model shows a strong descriptive power for the ISCAS benchmark data with an RMS error of 0.102 for the shortest path length complexity. Therefore, the model can be considered as a method of predicting path length complexities; this is expected to lead to minimum time complexity of very large-scale integrated circuitries and related computer-aided design tools that use binary decision diagrams.Keywords: Monte Carlo circuit simulation data, binary decision diagrams, neural network modeling, shortest path length estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13788808 Modeling Influence on Petty Corruption Attitudes
Authors: Nina Bijedic, Drazena Gaspar, Mirsad Hadzikadic
Abstract:
Corruption is an influential and widespread problem. One part of it is so-called petty corruption, related to large-scale bribe giving by ordinary citizens trying to influence the works of public administration or public services. As it is with all means of corruption, petty corruption is related to the level of democracy (or administration efficiency) in a society. The developed model captures some of the factors related to corruptive behavior, as well as people’s attitude towards petty corruption. It has four basic elements: user’s perception of corruption in the society of interest, the influence of social interactions, the influence of penalizing mechanism, and influence of campaigns against petty corruption. The model is agent-based, developed in NetLogo, with a lot of random settings that provide a wider scope of responses. Interactions of different settings for variables of elements provide insight into the influence of each element on attitude towards petty corruption, as well as petty corruptive behavior.
Keywords: Agent based model, attitude, influence, petty corruption, society.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13498807 Implementation of Technology Concept for the Reduction of Cyanobacteria in Laboratory
Authors: D. Šebo, M. Fedorčáková
Abstract:
Following the research in the Department of environmental engineering in Faculty of mechanical engineering on Technical University of Kosice and experiences with electrocoagulation style of disposal waste water, there were designed and partly examining the equipment of two stage revitalization on the standing and little fusible water of tenet electrolysis on the little tarns. With the cooperation with vet experts was that manners prove and it is innocuous for animals, during which time cyanobacteria are totally paralyzed. For the implementation of science and research results have been obtained by means EU funds for structural development.
Keywords: Cyanobacteria, Equipment, Pollution, Stagnant Water, Technology
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13318806 Radial Basis Surrogate Model Integrated to Evolutionary Algorithm for Solving Computation Intensive Black-Box Problems
Authors: Abdulbaset Saad, Adel Younis, Zuomin Dong
Abstract:
For design optimization with high-dimensional expensive problems, an effective and efficient optimization methodology is desired. This work proposes a series of modification to the Differential Evolution (DE) algorithm for solving computation Intensive Black-Box Problems. The proposed methodology is called Radial Basis Meta-Model Algorithm Assisted Differential Evolutionary (RBF-DE), which is a global optimization algorithm based on the meta-modeling techniques. A meta-modeling assisted DE is proposed to solve computationally expensive optimization problems. The Radial Basis Function (RBF) model is used as a surrogate model to approximate the expensive objective function, while DE employs a mechanism to dynamically select the best performing combination of parameters such as differential rate, cross over probability, and population size. The proposed algorithm is tested on benchmark functions and real life practical applications and problems. The test results demonstrate that the proposed algorithm is promising and performs well compared to other optimization algorithms. The proposed algorithm is capable of converging to acceptable and good solutions in terms of accuracy, number of evaluations, and time needed to converge.
Keywords: Differential evolution, engineering design, expensive computations, meta-modeling, radial basis function, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11738805 Numerical Simulation on Deformation Behaviour of Additively Manufactured AlSi10Mg Alloy
Authors: Racholsan Raj Nirmal, B. S. V. Patnaik, R. Jayaganthan
Abstract:
The deformation behaviour of additively manufactured AlSi10Mg alloy under low strains, high strain rates and elevated temperature conditions is essential to analyse and predict its response against dynamic loading such as impact and thermomechanical fatigue. The constitutive relation of Johnson-Cook is used to capture the strain rate sensitivity and thermal softening effect in AlSi10Mg alloy. Johnson-Cook failure model is widely used for exploring damage mechanics and predicting the fracture in many materials. In this present work, Johnson-Cook material and damage model parameters for additively manufactured AlSi10Mg alloy have been determined numerically from four types of uniaxial tensile test. Three different uniaxial tensile tests with dynamic strain rates (0.1, 1, 10, 50, and 100 s-1) and elevated temperature tensile test with three different temperature conditions (450 K, 500 K and 550 K) were performed on 3D printed AlSi10Mg alloy in ABAQUS/Explicit. Hexahedral elements are used to discretize tensile specimens and fracture energy value of 43.6 kN/m was used for damage initiation. Levenberg Marquardt optimization method was used for the evaluation of Johnson-Cook model parameters. It was observed that additively manufactured AlSi10Mg alloy has shown relatively higher strain rate sensitivity and lower thermal stability as compared to the other Al alloys.
Keywords: ABAQUS, additive manufacturing, AlSi10Mg, Johnson-Cook model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11368804 Dynamic Process Monitoring of an Ammonia Synthesis Fixed-Bed Reactor
Authors: Bothinah Altaf, Gary Montague, Elaine B. Martin
Abstract:
This study involves the modeling and monitoring of an ammonia synthesis fixed-bed reactor using partial least squares (PLS) and its variants. The process exhibits complex dynamic behavior due to the presence of heat recycling and feed quench. One limitation of static PLS model in this situation is that it does not take account of the process dynamics and hence dynamic PLS was used. Although it showed, superior performance to static PLS in terms of prediction, the monitoring scheme was inappropriate hence adaptive PLS was considered. A limitation of adaptive PLS is that non-conforming observations also contribute to the model, therefore, a new adaptive approach was developed, robust adaptive dynamic PLS. This approach updates a dynamic PLS model and is robust to non-representative data. The developed methodology showed a clear improvement over existing approaches in terms of the modeling of the reactor and the detection of faults.Keywords: Ammonia synthesis fixed-bed reactor, dynamic partial least squares modeling, recursive partial least squares, robust modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19228803 Modeling Child Development Factors for the Early Introduction of ICTs in Schools
Authors: K. E. Oyetade, S. D. Eyono Obono
Abstract:
One of the fundamental characteristics of Information and Communication Technology (ICT) has been the ever-changing nature of continuous release and models of ICTs with its impact on the academic, social, and psychological benefits of its introduction in schools. However, there seems to be a growing concern about its negative impact on students when introduced early in schools for teaching and learning. This study aims to design a model of child development factors affecting the early introduction of ICTs in schools in an attempt to improve the understanding of child development and introduction of ICTs in schools. The proposed model is based on a sound theoretical framework. It was designed following a literature review of child development theories and child development factors. The child development theoretical framework that fitted to the best of all child development factors was then chosen as the basis for the proposed model. This study hence found that the Jean Piaget cognitive developmental theory is the most adequate theoretical frameworks for modeling child development factors for ICT introduction in schools.Keywords: Child development factors, child development theories, ICTs, theory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20318802 A Model to Study the Effect of Excess Buffers and Na+ Ions on Ca2+ Diffusion in Neuron Cell
Authors: Vikas Tewari, Shivendra Tewari, K. R. Pardasani
Abstract:
Calcium is a vital second messenger used in signal transduction. Calcium controls secretion, cell movement, muscular contraction, cell differentiation, ciliary beating and so on. Two theories have been used to simplify the system of reaction-diffusion equations of calcium into a single equation. One is excess buffer approximation (EBA) which assumes that mobile buffer is present in excess and cannot be saturated. The other is rapid buffer approximation (RBA), which assumes that calcium binding to buffer is rapid compared to calcium diffusion rate. In the present work, attempt has been made to develop a model for calcium diffusion under excess buffer approximation in neuron cells. This model incorporates the effect of [Na+] influx on [Ca2+] diffusion,variable calcium and sodium sources, sodium-calcium exchange protein, Sarcolemmal Calcium ATPase pump, sodium and calcium channels. The proposed mathematical model leads to a system of partial differential equations which have been solved numerically using Forward Time Centered Space (FTCS) approach. The numerical results have been used to study the relationships among different types of parameters such as buffer concentration, association rate, calcium permeability.
Keywords: Excess buffer approximation, Na+ influx, sodium calcium exchange protein, sarcolemmal calcium atpase pump, forward time centred space.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1597