Search results for: distribution system and optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 23233

Search results for: distribution system and optimization

21403 Modelling Operational Risk Using Extreme Value Theory and Skew t-Copulas via Bayesian Inference

Authors: Betty Johanna Garzon Rozo, Jonathan Crook, Fernando Moreira

Abstract:

Operational risk losses are heavy tailed and are likely to be asymmetric and extremely dependent among business lines/event types. We propose a new methodology to assess, in a multivariate way, the asymmetry and extreme dependence between severity distributions, and to calculate the capital for Operational Risk. This methodology simultaneously uses (i) several parametric distributions and an alternative mix distribution (the Lognormal for the body of losses and the Generalized Pareto Distribution for the tail) via extreme value theory using SAS®, (ii) the multivariate skew t-copula applied for the first time for operational losses and (iii) Bayesian theory to estimate new n-dimensional skew t-copula models via Markov chain Monte Carlo (MCMC) simulation. This paper analyses a newly operational loss data set, SAS Global Operational Risk Data [SAS OpRisk], to model operational risk at international financial institutions. All the severity models are constructed in SAS® 9.2. We implement the procedure PROC SEVERITY and PROC NLMIXED. This paper focuses in describing this implementation.

Keywords: operational risk, loss distribution approach, extreme value theory, copulas

Procedia PDF Downloads 593
21402 Pilot Scale Deproteinization Study on Fish Scale Using Response Surface Methodology

Authors: Fatima Bellali, Mariem Kharroubi

Abstract:

Fish scale wastes are one of the main sources of production of value-added products such as collagen. The main aim of this study is to investigate the optimization conditions of the sardine scale deproteinization using response surface methodology (RSM) on a pilot scale. In order to look for the optimal conditions, a Box–Behnken-based design of experiment (DOE) method was carried out. The model predicted values of product coal ash content were in good agreement with the experiment values (R2 = 0.9813). Finally, model-based optimization was carried out to identify the operating parameters (reaction time=4h and the solid-liquid ratio= 1/10) and to obtain the lowest collagen content.

Keywords: pilot scale, Plackett and Burman design, fish waste, deproteinization

Procedia PDF Downloads 154
21401 A Hierarchical Bayesian Calibration of Data-Driven Models for Composite Laminate Consolidation

Authors: Nikolaos Papadimas, Joanna Bennett, Amir Sakhaei, Timothy Dodwell

Abstract:

Composite modeling of consolidation processes is playing an important role in the process and part design by indicating the formation of possible unwanted prior to expensive experimental iterative trial and development programs. Composite materials in their uncured state display complex constitutive behavior, which has received much academic interest, and this with different models proposed. Errors from modeling and statistical which arise from this fitting will propagate through any simulation in which the material model is used. A general hyperelastic polynomial representation was proposed, which can be readily implemented in various nonlinear finite element packages. In our case, FEniCS was chosen. The coefficients are assumed uncertain, and therefore the distribution of parameters learned using Markov Chain Monte Carlo (MCMC) methods. In engineering, the approach often followed is to select a single set of model parameters, which on average, best fits a set of experiments. There are good statistical reasons why this is not a rigorous approach to take. To overcome these challenges, A hierarchical Bayesian framework was proposed in which population distribution of model parameters is inferred from an ensemble of experiments tests. The resulting sampled distribution of hyperparameters is approximated using Maximum Entropy methods so that the distribution of samples can be readily sampled when embedded within a stochastic finite element simulation. The methodology is validated and demonstrated on a set of consolidation experiments of AS4/8852 with various stacking sequences. The resulting distributions are then applied to stochastic finite element simulations of the consolidation of curved parts, leading to a distribution of possible model outputs. With this, the paper, as far as the authors are aware, represents the first stochastic finite element implementation in composite process modelling.

Keywords: data-driven , material consolidation, stochastic finite elements, surrogate models

Procedia PDF Downloads 140
21400 Identification of Outliers in Flood Frequency Analysis: Comparison of Original and Multiple Grubbs-Beck Test

Authors: Ayesha S. Rahman, Khaled Haddad, Ataur Rahman

Abstract:

At-site flood frequency analysis is used to estimate flood quantiles when at-site record length is reasonably long. In Australia, FLIKE software has been introduced for at-site flood frequency analysis. The advantage of FLIKE is that, for a given application, the user can compare a number of most commonly adopted probability distributions and parameter estimation methods relatively quickly using a windows interface. The new version of FLIKE has been incorporated with the multiple Grubbs and Beck test which can identify multiple numbers of potentially influential low flows. This paper presents a case study considering six catchments in eastern Australia which compares two outlier identification tests (original Grubbs and Beck test and multiple Grubbs and Beck test) and two commonly applied probability distributions (Generalized Extreme Value (GEV) and Log Pearson type 3 (LP3)) using FLIKE software. It has been found that the multiple Grubbs and Beck test when used with LP3 distribution provides more accurate flood quantile estimates than when LP3 distribution is used with the original Grubbs and Beck test. Between these two methods, the differences in flood quantile estimates have been found to be up to 61% for the six study catchments. It has also been found that GEV distribution (with L moments) and LP3 distribution with the multiple Grubbs and Beck test provide quite similar results in most of the cases; however, a difference up to 38% has been noted for flood quantiles for annual exceedance probability (AEP) of 1 in 100 for one catchment. These findings need to be confirmed with a greater number of stations across other Australian states.

Keywords: floods, FLIKE, probability distributions, flood frequency, outlier

Procedia PDF Downloads 445
21399 FEM for Stress Reduction by Optimal Auxiliary Holes in a Uniaxially Loaded Plate

Authors: Basavaraj R. Endigeri, Shriharsh Desphande

Abstract:

Optimization and reduction of stress concentration around holes in a uniaxially loaded plate is one of the important design criteria in many of the engineering applications. These stress risers will lead to failure of the component at the region of high stress concentration which has to be avoided by means of providing auxiliary holes on either side of the parent hole. By literature survey it is known that till date, there is no analytical solution documented to reduce the stress concentration by providing auxiliary holes expect for fever geometries. In the present work, plate with a hole subjected to uniaxial load is analyzed with the numerical method to determine the optimum sizes and locations for the auxillary holes for different center hole diameter to plate width ratios. The introduction of auxiliary holes at a optimum location and radii with its effect on stress concentration is also represented graphically. The finite element analysis package ANSYS 8.0 is used to carry out analysis and optimization is performed to determine the location and radii for optimum values of auxiliary holes to reduce stress concentration. All the results for different diameter to plate width ratio are presented graphically. It is found from the work that introduction of auxiliary holes on either side of central circular hole will reduce stress concentration factor by a factor of 19 to 21 percentage.

Keywords: finite element method, optimization, stress concentration factor, auxiliary holes

Procedia PDF Downloads 436
21398 Improved Dynamic Bayesian Networks Applied to Arabic On Line Characters Recognition

Authors: Redouane Tlemsani, Abdelkader Benyettou

Abstract:

Work is in on line Arabic character recognition and the principal motivation is to study the Arab manuscript with on line technology. This system is a Markovian system, which one can see as like a Dynamic Bayesian Network (DBN). One of the major interests of these systems resides in the complete models training (topology and parameters) starting from training data. Our approach is based on the dynamic Bayesian Networks formalism. The DBNs theory is a Bayesians networks generalization to the dynamic processes. Among our objective, amounts finding better parameters, which represent the links (dependences) between dynamic network variables. In applications in pattern recognition, one will carry out the fixing of the structure, which obliges us to admit some strong assumptions (for example independence between some variables). Our application will relate to the Arabic isolated characters on line recognition using our laboratory database: NOUN. A neural tester proposed for DBN external optimization. The DBN scores and DBN mixed are respectively 70.24% and 62.50%, which lets predict their further development; other approaches taking account time were considered and implemented until obtaining a significant recognition rate 94.79%.

Keywords: Arabic on line character recognition, dynamic Bayesian network, pattern recognition, computer vision

Procedia PDF Downloads 424
21397 Smart Irrigation System

Authors: Levent Seyfi, Ertan Akman, Tuğrul C. Topak

Abstract:

In this study, irrigation automation with electronic sensors and its control with smartphones were aimed. In this context, temperature and soil humidity measurements of the area irrigated were obtained by temperature and humidity sensors. A micro controller (Arduino) was utilized for accessing values of these parameters and controlling the proposed irrigation system. The irrigation system could automatically be worked according to obtained measurement values. Besides, a GSM module used together with Arduino provided that the irrigation system was in connection to smartphones. Thus, the irrigation system can be remotely controlled. Not only can we observe whether the irrigation system is working or not via developed special android application but also we can see temperature and humidity measurement values. In addition to this, if desired, the irrigation system can be remotely and manually started or stopped regardless of measured sensor vales thanks to the developed android application. In addition to smartphones, the irrigation system can be alternatively controlled via the designed website (www.sulamadenetim.com).

Keywords: smartphone, Android Operating System, sensors, irrigation System, arduino

Procedia PDF Downloads 609
21396 Optimization of Fenton Process for the Treatment of Young Municipal Leachate

Authors: Bouchra Wassate, Younes Karhat, Khadija El Falaki

Abstract:

Leachate is a source of surface water and groundwater contamination if it has not been pretreated. Indeed, due to its complex structure and its pollution load make its treatment extremely difficult to achieve the standard limits required. The objective of this work is to show the interest of advanced oxidation processes on leachate treatment of urban waste containing high concentrations of organic pollutants. The efficiency of Fenton (Fe2+ +H2O2 + H+) reagent for young leachate recovered from collection trucks household waste in the city of Casablanca, Morocco, was evaluated with the objectives of chemical oxygen demand (COD) and discoloration reductions. The optimization of certain physicochemical parameters (initial pH value, reaction time, and [Fe2+], [H2O2]/ [Fe2+] ratio) has yielded good results in terms of reduction of COD and discoloration of the leachate.

Keywords: COD removal, color removal, Fenton process, oxidation process, leachate

Procedia PDF Downloads 284
21395 Hybrid Approach for the Min-Interference Frequency Assignment

Authors: F. Debbat, F. T. Bendimerad

Abstract:

The efficient frequency assignment for radio communications becomes more and more crucial when developing new information technologies and their applications. It is consists in defining an assignment of frequencies to radio links, to be established between base stations and mobile transmitters. Separation of the frequencies assigned is necessary to avoid interference. However, unnecessary separation causes an excess requirement for spectrum, the cost of which may be very high. This problem is NP-hard problem which cannot be solved by conventional optimization algorithms. It is therefore necessary to use metaheuristic methods to solve it. This paper proposes Hybrid approach based on simulated annealing (SA) and Tabu Search (TS) methods to solve this problem. Computational results, obtained on a number of standard problem instances, testify the effectiveness of the proposed approach.

Keywords: cellular mobile communication, frequency assignment problem, optimization, tabu search, simulated annealing

Procedia PDF Downloads 379
21394 Development of a General Purpose Computer Programme Based on Differential Evolution Algorithm: An Application towards Predicting Elastic Properties of Pavement

Authors: Sai Sankalp Vemavarapu

Abstract:

This paper discusses the application of machine learning in the field of transportation engineering for predicting engineering properties of pavement more accurately and efficiently. Predicting the elastic properties aid us in assessing the current road conditions and taking appropriate measures to avoid any inconvenience to commuters. This improves the longevity and sustainability of the pavement layer while reducing its overall life-cycle cost. As an example, we have implemented differential evolution (DE) in the back-calculation of the elastic modulus of multi-layered pavement. The proposed DE global optimization back-calculation approach is integrated with a forward response model. This approach treats back-calculation as a global optimization problem where the cost function to be minimized is defined as the root mean square error in measured and computed deflections. The optimal solution which is elastic modulus, in this case, is searched for in the solution space by the DE algorithm. The best DE parameter combinations and the most optimum value is predicted so that the results are reproducible whenever the need arises. The algorithm’s performance in varied scenarios was analyzed by changing the input parameters. The prediction was well within the permissible error, establishing the supremacy of DE.

Keywords: cost function, differential evolution, falling weight deflectometer, genetic algorithm, global optimization, metaheuristic algorithm, multilayered pavement, pavement condition assessment, pavement layer moduli back calculation

Procedia PDF Downloads 159
21393 Optimization of Biomass Production and Lipid Formation from Chlorococcum sp. Cultivation on Dairy and Paper-Pulp Wastewater

Authors: Emmanuel C. Ngerem

Abstract:

The ever-increasing depletion of the dominant global form of energy (fossil fuels) calls for the development of sustainable and green alternative energy sources such as bioethanol, biohydrogen, and biodiesel. The production of the major biofuels relies on biomass feedstocks that are mainly derived from edible food crops and some inedible plants. One suitable feedstock with great potential as raw material for biofuel production is microalgal biomass. Despite the tremendous attributes of microalgae as a source of biofuel, their cultivation requires huge volumes of freshwater, thus posing a serious threat to commercial-scale production and utilization of algal biomass. In this study, a multi-media wastewater mixture for microalgae growth was formulated and optimized. Moreover, the obtained microalgae biomass was pre-treated to reduce sugar recovery and was compared with previous studies on microalgae biomass pre-treatment. The formulated and optimized mixed wastewater media for biomass and lipid accumulation was established using the simplex lattice mixture design. Based on the superposition approach of the potential results, numerical optimization was conducted, followed by the analysis of biomass concentration and lipid accumulation. The coefficients of regression (R²) of 0.91 and 0.98 were obtained for biomass concentration and lipid accumulation models, respectively. The developed optimization model predicted optimal biomass concentration and lipid accumulation of 1.17 g/L and 0.39 g/g, respectively. It suggested 64.69% dairy wastewater (DWW) and 35.31% paper and pulp wastewater (PWW) mixture for biomass concentration, 34.21% DWW, and 65.79% PWW for lipid accumulation. Experimental validation generated 0.94 g/L and 0.39 g/g of biomass concentration and lipid accumulation, respectively. The obtained microalgae biomass was pre-treated, enzymatically hydrolysed, and subsequently assessed for reducing sugars. The optimization of microwave pre-treatment of Chlorococcum sp. was achieved using response surface methodology (RSM). Microwave power (100 – 700 W), pre-treatment time (1 – 7 min), and acid-liquid ratio (1 – 5%) were selected as independent variables for RSM optimization. The optimum conditions were achieved at microwave power, pre-treatment time, and acid-liquid ratio of 700 W, 7 min, and 32.33:1, respectively. These conditions provided the highest amount of reducing sugars at 10.73 g/L. Process optimization predicted reducing sugar yields of 11.14 g/L on microwave-assisted pre-treatment of 2.52% HCl for 4.06 min at 700 watts. Experimental validation yielded reducing sugars of 15.67 g/L. These findings demonstrate that dairy wastewater and paper and pulp wastewater that could pose a serious environmental nuisance. They could be blended to form a suitable microalgae growth media, consolidating the potency of microalgae as a viable feedstock for fermentable sugars. Also, the outcome of this study supports the microalgal wastewater biorefinery concept, where wastewater remediation is coupled with bioenergy production.

Keywords: wastewater cultivation, mixture design, lipid, biomass, nutrient removal, microwave, Chlorococcum, raceway pond, fermentable sugar, modelling, optimization

Procedia PDF Downloads 29
21392 Enhanced Magnetic Hyperthermic Efficiency of Ferrite Based Nanoparticles

Authors: J. P. Borah, R. D. Raland

Abstract:

Hyperthermia is one of many techniques used destroys cancerous cell. It uses the physical methods to heat certain organ or tissue delivering an adequate temperature in an appropriate period of time, to the entire tumor volume for achieving optimal therapeutic results. Magnetic Metal ferrites nanoparticles (MFe₂O₄ where M = Mn, Zn, Ni, Co, Mg, etc.) are one of the most potential candidates for hyperthermia due to their tunability, biocompatibility, chemical stability and notable ability to mediate high rate of heat induction. However, to obtain the desirable properties for these applications, it is important to optimize their chemical composition, structure and magnetic properties. These properties are mainly sensitive to cation distribution of tetrahedral and octahedral sites. Among the ferrites, zinc ferrite (ZnFe₂O₄) and Manganese ferrite ((MnFe₂O₄) is one of a strong candidate for hyperthermia application because Mn and zinc have a non-magnetic cation and therefore the magnetic property is determined only by the cation distribution of iron, which provides a better platform to manipulate or tailor the properties. In this talk, influence of doping and surfactant towards cation re-distribution leading to an enhancement of magnetic properties of ferrite nanoparticles will be demonstrated. The efficiency of heat generation in association with the enhanced magnetic property is also well discussed in this talk.

Keywords: magnetic nanoparticle, hyperthermia, x-ray diffraction, TEM study

Procedia PDF Downloads 161
21391 A Remotely Piloted Aerial Application System to Control Rangeland Grasshoppers

Authors: Daniel Martin, Roberto Rodriguez, Derek Woller, Chris Reuter, Lonnie Black, Mohamed Latheef

Abstract:

The grasshoppers comprised of heterogeneous assemblages of Acrididae (Family: Orthoptera) species periodically reach outbreak levels by their gregarious behavior and voracious feeding habits, devouring stems and leaves of food crops and rangeland pasture. Cattle consume about 1.5-2.5% of their body weight in forage per day, so pound for pound, a grasshopper will eat 12-20 times as much plant material as a steer and cause serious economic damage to the cattle industry, especially during a drought when forage is already scarce. Grasshoppers annually consume more than 20% of rangeland forages in the western United States at an estimated loss of $1.25 billion per year in forage. A remotely piloted aerial application system with both a spreader and spray application system was used to apply granular insect bait and a liquid formulation of Carbaryl for control of grasshopper infestations on rangeland in New Mexico, United States. Pattern testing and calibration of both the granular and liquid application systems were conducted to determine proper application rate set up and distribution pattern. From these tests, an effective swath was calculated. Results showed that 14 days after application, granular baits were only effective on those grasshopper species that accepted the baits. The liquid formulation at 16 ounces per acre was highly successful at controlling all grasshopper species. Results of this study indicated that a remotely piloted aerial application system can be used to effectively deliver grasshopper control products in both granular and liquid form. However, the spray application treatment proved to be most effective and efficient for all grasshopper species present.

Keywords: Carbaryl, Grasshopper, Insecticidal Efficacy, Remotely Piloted Aerial Application System

Procedia PDF Downloads 210
21390 Size Distribution Effect of InAs/InP Self–Organized Quantum Dots on Optical Properties

Authors: Abdelkader Nouri, M’hamed Bouslama, Faouzi Saidi, Hassan Maaref, Michel Gendry

Abstract:

Self-organized InAs quantum dots (QDs) have been grown on 3,1% InP (110) lattice mismatched substrate by Solid Source Molecular Beam Epitaxy (SSMBE). Stranski-Krastanov mode growth has been used to create self-assembled 3D islands on InAs wetting layer (WL). The optical quality depending on the temperature and power is evaluated. In addition, Atomic Force Microscopy (AFM) images shows inhomogeneous island dots size distribution due to temperature coalescence. The quantum size effect was clearly observed through the spectra photoluminescence (PL) shape.

Keywords: AFM, InAs QDs, PL, SSMBE

Procedia PDF Downloads 681
21389 Heat Transfer Process Parameter Optimization in SI/Ge Using TAGUCHI Method

Authors: Evln Ranga Charyulu, S. P. Venu Madhavarao, S. Udaya kumar, S. V. S. S. N. V. G. Krishna Murthy

Abstract:

With the advent of new nanometer process technologies, it is possible to integrate billion transistors on a single substrate. When more and more functionality included there is the possibility of multi-million transistors switching simultaneously consuming more power and dissipating more power along with more leakage of current into the substrate of porous silicon or germanium material. These results in substrate heating and thermal noise generation coupled to signals of interest. The heating process is represented by coupled nonlinear partial differential equations in porous silicon and germanium. By identifying heat sources and heat fluxes may results in designing of ultra-low power circuits. The PDEs are solved by finite difference scheme assuming that boundary layer equations in porous silicon and germanium. Local heat fluxes along the vertical isothermal surface immersed in porous SI/Ge are considered. The parameters considered for optimization are thermal diffusivity, thermal expansion coefficient, thermal diffusion ratio, permeability, specific heat at constant temperatures, Rayleigh number, amplitude of wavy surface, mass expansion coefficient. The diffusion of heat was caused by the concentration gradient. Thermal physical properties are homogeneous and isotropic. By using L8, TAGUCHI method the parameters are optimized.

Keywords: heat transfer, pde, taguchi optimization, SI/Ge

Procedia PDF Downloads 332
21388 Microwave Heating and Catalytic Activity of Iron/Carbon Materials for H₂ Production from the Decomposition of Plastic Wastes

Authors: Peng Zhang, Cai Liang

Abstract:

The non-biodegradable plastic wastes have posed severe environmental and ecological contaminations. Numerous technologies, such as pyrolysis, incineration, and landfilling, have already been employed for the treatment of plastic waste. Compared with conventional methods, microwave has displayed unique advantages in the rapid production of hydrogen from plastic wastes. Understanding the interaction between microwave radiation and materials would promote the optimization of several parameters for the microwave reaction system. In this work, various carbon materials have been investigated to reveal microwave heating performance and the ensuing catalytic activity. Results showed that the diversity in the heating characteristic was mainly due to the dielectric properties and the individual microstructures. Furthermore, the gaps and steps among the surface of carbon materials would lead to the distortion of the electromagnetic field, which correspondingly induced plasma discharging. The intensity and location of local plasma were also studied. For high-yield H₂ production, iron nanoparticles were selected as the active sites, and a series of iron/carbon bifunctional catalysts were synthesized. Apart from the high catalytic activity, the iron particles in nano-size close to the microwave skin depth would transfer microwave irradiation to the heat, intensifying the decomposition of plastics. Under microwave radiation, iron is supported on activated carbon material with 10wt.% loading exhibited the best catalytic activity for H₂ production. Specifically, the plastics were rapidly heated up and subsequently converted into H₂ with a hydrogen efficiency of 85%. This work demonstrated a deep understanding of microwave reaction systems and provided the optimization for plastic treatment.

Keywords: plastic waste, recycling, hydrogen, microwave

Procedia PDF Downloads 63
21387 Strain DistributionProfiles of EDD Steel at Elevated Temperatures

Authors: Eshwara Prasad Koorapati, R. Raman Goud, Swadesh Kumar Singh

Abstract:

In the present work forming limit diagrams and strain distribution profile diagrams for extra deep drawing steel at room and elevated temperatures have been determined experimentally by conducting stretch forming experiments by using designed and fabricated warm stretch forming tooling setup. With the help of forming Limit Diagrams (FLDs) and strain distribution profile diagrams the formability of Extra Deep Drawing steel has been analyzed and co-related with mechanical properties like strain hardening coefficient (n) and normal anisotropy (r−).Mechanical properties of EDD steel from room temperature to 4500 C were determined and discussed the impact of temperature on the properties like work hardening exponent (n) anisotropy (r-) and strength coefficient of the material. Also, the fractured surfaces after stretching have undergone the some metallurgical investigations and attempt has been made to co-relate with the formability of EDD steel sheets. They are co-related and good agreement with FLDs at various temperatures.

Keywords: FLD, micro hardness, strain distribution profile, stretch forming

Procedia PDF Downloads 417
21386 Statistical Feature Extraction Method for Wood Species Recognition System

Authors: Mohd Iz'aan Paiz Bin Zamri, Anis Salwa Mohd Khairuddin, Norrima Mokhtar, Rubiyah Yusof

Abstract:

Effective statistical feature extraction and classification are important in image-based automatic inspection and analysis. An automatic wood species recognition system is designed to perform wood inspection at custom checkpoints to avoid mislabeling of timber which will results to loss of income to the timber industry. The system focuses on analyzing the statistical pores properties of the wood images. This paper proposed a fuzzy-based feature extractor which mimics the experts’ knowledge on wood texture to extract the properties of pores distribution from the wood surface texture. The proposed feature extractor consists of two steps namely pores extraction and fuzzy pores management. The total number of statistical features extracted from each wood image is 38 features. Then, a backpropagation neural network is used to classify the wood species based on the statistical features. A comprehensive set of experiments on a database composed of 5200 macroscopic images from 52 tropical wood species was used to evaluate the performance of the proposed feature extractor. The advantage of the proposed feature extraction technique is that it mimics the experts’ interpretation on wood texture which allows human involvement when analyzing the wood texture. Experimental results show the efficiency of the proposed method.

Keywords: classification, feature extraction, fuzzy, inspection system, image analysis, macroscopic images

Procedia PDF Downloads 419
21385 Energy Efficiency Index Applied to Reactive Systems

Authors: P. Góes, J. Manzi

Abstract:

This paper focuses on the development of an energy efficiency index that will be applied to reactive systems, which is based in the First and Second Law of Thermodynamics, by giving particular consideration to the concept of maximum entropy. Among the requirements of such energy efficiency index, the practical feasibility must be essential. To illustrate the performance of the proposed index, such an index was used as decisive factor of evaluation for the optimization process of an industrial reactor. The results allow the conclusion to be drawn that the energy efficiency index applied to the reactive system is consistent because it extracts the information expected of an efficient indicator, and that it is useful as an analytical tool besides being feasible from a practical standpoint. Furthermore, it has proved to be much simpler to use than tools based on traditional methodologies.

Keywords: energy, efficiency, entropy, reactive

Procedia PDF Downloads 406
21384 A Convolution Neural Network Approach to Predict Pes-Planus Using Plantar Pressure Mapping Images

Authors: Adel Khorramrouz, Monireh Ahmadi Bani, Ehsan Norouzi, Morvarid Lalenoor

Abstract:

Background: Plantar pressure distribution measurement has been used for a long time to assess foot disorders. Plantar pressure is an important component affecting the foot and ankle function and Changes in plantar pressure distribution could indicate various foot and ankle disorders. Morphologic and mechanical properties of the foot may be important factors affecting the plantar pressure distribution. Accurate and early measurement may help to reduce the prevalence of pes planus. With recent developments in technology, new techniques such as machine learning have been used to assist clinicians in predicting patients with foot disorders. Significance of the study: This study proposes a neural network learning-based flat foot classification methodology using static foot pressure distribution. Methodologies: Data were collected from 895 patients who were referred to a foot clinic due to foot disorders. Patients with pes planus were labeled by an experienced physician based on clinical examination. Then all subjects (with and without pes planus) were evaluated for static plantar pressures distribution. Patients who were diagnosed with the flat foot in both feet were included in the study. In the next step, the leg length was normalized and the network was trained for plantar pressure mapping images. Findings: From a total of 895 image data, 581 were labeled as pes planus. A computational neural network (CNN) ran to evaluate the performance of the proposed model. The prediction accuracy of the basic CNN-based model was performed and the prediction model was derived through the proposed methodology. In the basic CNN model, the training accuracy was 79.14%, and the test accuracy was 72.09%. Conclusion: This model can be easily and simply used by patients with pes planus and doctors to predict the classification of pes planus and prescreen for possible musculoskeletal disorders related to this condition. However, more models need to be considered and compared for higher accuracy.

Keywords: foot disorder, machine learning, neural network, pes planus

Procedia PDF Downloads 353
21383 Effect of Depth on the Distribution of Zooplankton in Wushishi Lake Minna, Niger State, Nigeria

Authors: Adamu Zubairu Mohammed, Fransis Oforum Arimoro, Salihu Maikudi Ibrahim, Y. I. Auta, T. I. Arowosegbe, Y. Abdullahi

Abstract:

The present study was conducted to evaluate the effect of depth on the distribution of zooplankton and some physicochemical parameters in Tungan Kawo Lake (Wushishi dam). Water and zooplankton samples were collected from the surface, 3.0 meters deep and 6.0 meters deep, for a period of 24 hours for six months. Standard procedures were adopted for the determination of physicochemical parameters. Results have shown significant differences in the pH, DO, BOD Hardness, Na, and Mg. A total of 1764 zooplankton were recorded, comprising 35 species, with cladocera having 18 species (58%), 14 species of copepoda (41%), 3 species of diptera (1.0%). Results show that more of the zooplankton were recorded in the 3.0 meters-deep region compared to the two other depts and a significant difference was observed in the distribution of Ceriodaphnia dubia, Daphnia laevis, and Leptodiaptomus coloradensis. Though the most abundant zooplankton was recorded in the 3.0 meters deep, Leptodiaptomus coloradesnsis, which was observed in the 6.0 meters deep as the most individual observed, this was followed by Daphnia laevis. Canonical correspondence analysis between physicochemical parameters and the zooplankton indicated a good relationship in the Lake. Ceriodaphnia dubia was found to have a good association with oxygen, sodium, and potassium, while Daphnia laevis and Leptodiaptomus coloradensis are in good relationship with magnesium and phosphorus. It was generally observed that this depth does not have much influence on the distribution of zooplankton in Wushishi Lake.

Keywords: zooplankton, standard procedures, canonical correspondence analysis, Wushishi, canonical, physicochemical parameter

Procedia PDF Downloads 84
21382 A New Framework for ECG Signal Modeling and Compression Based on Compressed Sensing Theory

Authors: Siavash Eftekharifar, Tohid Yousefi Rezaii, Mahdi Shamsi

Abstract:

The purpose of this paper is to exploit compressed sensing (CS) method in order to model and compress the electrocardiogram (ECG) signals at a high compression ratio. In order to obtain a sparse representation of the ECG signals, first a suitable basis matrix with Gaussian kernels, which are shown to nicely fit the ECG signals, is constructed. Then the sparse model is extracted by applying some optimization technique. Finally, the CS theory is utilized to obtain a compressed version of the sparse signal. Reconstruction of the ECG signal from the compressed version is also done to prove the reliability of the algorithm. At this stage, a greedy optimization technique is used to reconstruct the ECG signal and the Mean Square Error (MSE) is calculated to evaluate the precision of the proposed compression method.

Keywords: compressed sensing, ECG compression, Gaussian kernel, sparse representation

Procedia PDF Downloads 460
21381 Ficus Carica as Adsorbent for Removal of Phenol from Aqueous Solutions: Modelling and Optimization

Authors: Tizi Hayet, Berrama Tarek, Bounif Nadia

Abstract:

Phenol and its derivatives are organic compounds utilized in the chemical industry. They are introduced into the environment by accidental spills and illegal release of industrial and municipal wastewater. Phenols are organic intermediaries that considered as potential pollutants. Adsorption is one of the purification and separation techniques used in this area. Algeria produces annually 131000 tones of fig; therefore, a large amount of fig leaves is generated, and the conversion of this waste into adsorbent allows the valorization of agricultural residue. The main purpose of this present work is to describe an application of the statistical method for modeling and optimization of the conditions of the phenol (Ph) adsorption from agricultural by-product locally available (fig leaves). The best experimental performance of Ph elimination on the adsorbent was obtained with: Adsorbent concentration (X2) = 0.2 g L-1; Initial concentration (X3) = 150 mg L-1; Speed agitation (X1) = 300 rpm.

Keywords: low-cost adsorbents, fig leaves, full factorial design, phenol, biosorption

Procedia PDF Downloads 86
21380 Control of Oil Content of Fried Zucchini Slices by Partial Predrying and Process Optimization

Authors: E. Karacabey, Ş. G. Özçelik, M. S. Turan, C. Baltacıoğlu, E. Küçüköner

Abstract:

Main concern about deep-fat-fried food materials is their high final oil contents absorbed during frying process and/or after cooling period, since diet including high content of oil is accepted unhealthy by consumers. Different methods have been evaluated to decrease oil content of fried food stuffs. One promising method is partially drying of food material before frying. In the present study it was aimed to control and decrease the final oil content of zucchini slices by means of partial drying and to optimize process conditions. Conventional oven drying was used to decrease moisture content of zucchini slices at a certain extent. Process performance in terms of oil uptake was evaluated by comparing oil content of predried and then fried zucchini slices with those determined for directly fried ones. For predrying and frying processes, oven temperature and weight loss and frying oil temperature and time pairs were controlled variables, respectively. Zucchini slices were also directly fried for sensory evaluations revealing preferred properties of final product in terms of surface color, moisture content, texture and taste. These properties of directly fried zucchini slices taking the highest score at the end of sensory evaluation were determined and used as targets in optimization procedure. Response surface methodology was used for process optimization. The properties, determined after sensory evaluation, were selected as targets; meanwhile oil content was aimed to be minimized. Results indicated that final oil content of zucchini slices could be reduced from 58% to 46% by controlling conditions of predrying and frying processes. As a result, it was suggested that predrying could be one choose to reduce oil content of fried zucchini slices for health diet. This project (113R015) has been supported by TUBITAK.

Keywords: health process, optimization, response surface methodology, oil uptake, conventional oven

Procedia PDF Downloads 365
21379 First Investigation on CZTS Electron affinity and Thickness Optimization using SILVACO-Atlas 2D Simulation

Authors: Zeineb Seboui, Samar Dabbabi

Abstract:

In this paper, we study the performance of Cu₂ZnSnS₄ (CZTS) based solar cell. In our knowledge, it is for the first time that the FTO/ZnO:Co/CZTS structure is simulated using the SILVACO-Atlas 2D simulation. Cu₂ZnSnS₄ (CZTS), ZnO:Co and FTO (SnO₂:F) layers have been deposited on glass substrates by the spray pyrolysis technique. The extracted physical properties, such as thickness and optical parameters of CZTS layer, are considered to create a new input data of CZTS based solar cell. The optimization of CZTS electron affinity and thickness is performed to have the best FTO/ZnO: Co/CZTS efficiency. The use of CZTS absorber layer with 3.99 eV electron affinity and 3.2 µm in thickness leads to the higher efficiency of 16.86 %, which is very important in the development of new technologies and new solar cell devices.

Keywords: CZTS solar cell, characterization, electron affinity, thickness, SILVACO-atlas 2D simulation

Procedia PDF Downloads 72
21378 The Design Optimization for Sound Absorption Material of Multi-Layer Structure

Authors: Un-Hwan Park, Jun-Hyeok Heo, In-Sung Lee, Tae-Hyeon Oh, Dae-Kyu Park

Abstract:

Sound absorbing material is used as automotive interior material. Sound absorption coefficient should be predicted to design it. But it is difficult to predict sound absorbing coefficient because it is comprised of several material layers. So, its targets are achieved through many experimental tunings. It causes a lot of cost and time. In this paper, we propose the process to estimate the sound absorption coefficient with multi-layer structure. In order to estimate the coefficient, physical properties of each material are used. These properties also use predicted values by Foam-X software using the sound absorption coefficient data measured by impedance tube. Since there are many physical properties and the measurement equipment is expensive, the values predicted by software are used. Through the measurement of the sound absorption coefficient of each material, its physical properties are calculated inversely. The properties of each material are used to calculate the sound absorption coefficient of the multi-layer material. Since the absorption coefficient of multi-layer can be calculated, optimization design is possible through simulation. Then, we will compare and analyze the calculated sound absorption coefficient with the data measured by scaled reverberation chamber and impedance tubes for a prototype. If this method is used when developing automotive interior materials with multi-layer structure, the development effort can be reduced because it can be optimized by simulation. So, cost and time can be saved.

Keywords: sound absorption material, sound impedance tube, sound absorption coefficient, optimization design

Procedia PDF Downloads 285
21377 A Stepwise Approach for Piezoresistive Microcantilever Biosensor Optimization

Authors: Amal E. Ahmed, Levent Trabzon

Abstract:

Due to the low concentration of the analytes in biological samples, the use of Biological Microelectromechanical System (Bio-MEMS) biosensors for biomolecules detection results in a minuscule output signal that is not good enough for practical applications. In response to this, a need has arisen for an optimized biosensor capable of giving high output signal in response the detection of few analytes in the sample; the ultimate goal is being able to convert the attachment of a single biomolecule into a measurable quantity. For this purpose, MEMS microcantilevers based biosensors emerged as a promising sensing solution because it is simple, cheap, very sensitive and more importantly does not need analytes optical labeling (Label-free). Among the different microcantilever transducing techniques, piezoresistive based microcantilever biosensors became more prominent because it works well in liquid environments and has an integrated readout system. However, the design of piezoresistive microcantilevers is not a straightforward problem due to coupling between the design parameters, constraints, process conditions, and performance. It was found that the parameters that can be optimized to enhance the sensitivity of Piezoresistive microcantilever-based sensors are: cantilever dimensions, cantilever material, cantilever shape, piezoresistor material, piezoresistor doping level, piezoresistor dimensions, piezoresistor position, Stress Concentration Region's (SCR) shape and position. After a systematic analyzation of the effect of each design and process parameters on the sensitivity, a step-wise optimization approach was developed in which almost all these parameters were variated one at each step while fixing the others to get the maximum possible sensitivity at the end. At each step, the goal was to optimize the parameter in a way that it maximizes and concentrates the stress in the piezoresistor region for the same applied force thus get the higher sensitivity. Using this approach, an optimized sensor that has 73.5x times higher electrical sensitivity (ΔR⁄R) than the starting sensor was obtained. In addition to that, this piezoresistive microcantilever biosensor it is more sensitive than the other similar sensors previously reported in the open literature. The mechanical sensitivity of the final senior is -1.5×10-8 Ω/Ω ⁄pN; which means that for each 1pN (10-10 g) biomolecules attach to this biosensor; the piezoresistor resistivity will decrease by 1.5×10-8 Ω. Throughout this work COMSOL Multiphysics 5.0, a commercial Finite Element Analysis (FEA) tool, has been used to simulate the sensor performance.

Keywords: biosensor, microcantilever, piezoresistive, stress concentration region (SCR)

Procedia PDF Downloads 567
21376 Enhancement of Long Term Peak Demand Forecast in Peninsular Malaysia Using Hourly Load Profile

Authors: Nazaitul Idya Hamzah, Muhammad Syafiq Mazli, Maszatul Akmar Mustafa

Abstract:

The peak demand forecast is crucial to identify the future generation plant up needed in the long-term capacity planning analysis for Peninsular Malaysia as well as for the transmission and distribution network planning activities. Currently, peak demand forecast (in Mega Watt) is derived from the generation forecast by using load factor assumption. However, a forecast using this method has underperformed due to the structural changes in the economy, emerging trends and weather uncertainty. The dynamic changes of these drivers will result in many possible outcomes of peak demand for Peninsular Malaysia. This paper will look into the independent model of peak demand forecasting. The model begins with the selection of driver variables to capture long-term growth. This selection and construction of variables, which include econometric, emerging trend and energy variables, will have an impact on the peak forecast. The actual framework begins with the development of system energy and load shape forecast by using the system’s hourly data. The shape forecast represents the system shape assuming all embedded technology and use patterns to continue in the future. This is necessary to identify the movements in the peak hour or changes in the system load factor. The next step would be developing the peak forecast, which involves an iterative process to explore model structures and variables. The final step is combining the system energy, shape, and peak forecasts into the hourly system forecast then modifying it with the forecast adjustments. Forecast adjustments are among other sales forecasts for electric vehicles, solar and other adjustments. The framework will result in an hourly forecast that captures growth, peak usage and new technologies. The advantage of this approach as compared to the current methodology is that the peaks capture new technology impacts that change the load shape.

Keywords: hourly load profile, load forecasting, long term peak demand forecasting, peak demand

Procedia PDF Downloads 160
21375 A Study on Design for Parallel Test Based on Embedded System

Authors: Zheng Sun, Weiwei Cui, Xiaodong Ma, Hongxin Jin, Dongpao Hong, Jinsong Yang, Jingyi Sun

Abstract:

With the improvement of the performance and complexity of modern equipment, automatic test system (ATS) becomes widely used for condition monitoring and fault diagnosis. However, the conventional ATS mainly works in a serial mode, and lacks the ability of testing several equipments at the same time. That leads to low test efficiency and ATS redundancy. Especially for a large majority of equipment under test, the conventional ATS cannot meet the requirement of efficient testing. To reduce the support resource and increase test efficiency, we propose a method of design for the parallel test based on the embedded system in this paper. Firstly, we put forward the general framework of the parallel test system, and the system contains a central management system (CMS) and several distributed test subsystems (DTS). Then we give a detailed design of the system. For the hardware of the system, we use embedded architecture to design DTS. For the software of the system, we use test program set to improve the test adaption. By deploying the parallel test system, the time to test five devices is now equal to the time to test one device in the past. Compared with the conventional test system, the proposed test system reduces the size and improves testing efficiency. This is of great significance for equipment to be put into operation swiftly. Finally, we take an industrial control system as an example to verify the effectiveness of the proposed method. The result shows that the method is reasonable, and the efficiency is improved up to 500%.

Keywords: parallel test, embedded system, automatic test system, automatic test system (ATS), central management system, central management system (CMS), distributed test subsystems, distributed test subsystems (DTS)

Procedia PDF Downloads 299
21374 A Game-Theory-Based Price-Optimization Algorithm for the Simulation of Markets Using Agent-Based Modelling

Authors: Juan Manuel Sanchez-Cartas, Gonzalo Leon

Abstract:

A price competition algorithm for ABMs based on game theory principles is proposed to deal with the simulation of theoretical market models. The algorithm is applied to the classical Hotelling’s model and to a two-sided market model to show it leads to the optimal behavior predicted by theoretical models. However, when theoretical models fail to predict the equilibrium, the algorithm is capable of reaching a feasible outcome. Results highlight that the algorithm can be implemented in other simulation models to guarantee rational users and endogenous optimal behaviors. Also, it can be applied as a tool of verification given that is theoretically based.

Keywords: agent-based models, algorithmic game theory, multi-sided markets, price optimization

Procedia PDF Downloads 448