Search results for: single error upset
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6257

Search results for: single error upset

5177 On the Question of Ideology: Criticism of the Enlightenment Approach and Theory of Ideology as Objective Force in Gramsci and Althusser

Authors: Edoardo Schinco

Abstract:

Studying the Marxist intellectual tradition, it is possible to verify that there were numerous cases of philosophical regression, in which the important achievements of detailed studies have been replaced by naïve ideas and previous misunderstandings: one of most important example of this tendency is related to the question of ideology. According to a common Enlightenment approach, the ideology is essentially not a reality, i.e., a factor capable of having an effect on the reality itself; in other words, the ideology is a mere error without specific historical meaning, which is only due to ignorance or inability of subjects to understand the truth. From this point of view, the consequent and immediate practice against every form of ideology are the rational dialogue, the reasoning based on common sense, in order to dispel the obscurity of ignorance through the light of pure reason. The limits of this philosophical orientation are however both theoretical and practical: on the one hand, the Enlightenment criticism of ideology is not an historicistic thought, since it cannot grasp the inner connection that ties an historical context and its peculiar ideology together; moreover, on the other hand, when the Enlightenment approach fails to release people from their illusions (e.g., when the ideology persists, despite the explanation of its illusoriness), it usually becomes a racist or elitarian thought. Unlike this first conception of ideology, Gramsci attempts to recover Marx’s original thought and to valorize its dialectical methodology with respect to the reality of ideology. As Marx suggests, the ideology – in negative meaning – is surely an error, a misleading knowledge, which aims to defense the current state of things and to conceal social, political or moral contradictions; but, that is precisely why the ideological error is not casual: every ideology mediately roots in a particular material context, from which it takes its reason being. Gramsci avoids, however, any mechanistic interpretation of Marx and, for this reason; he underlines the dialectic relation that exists between material base and ideological superstructure; in this way, a specific ideology is not only a passive product of base but also an active factor that reacts on the base itself and modifies it. Therefore, there is a considerable revaluation of ideology’s role in maintenance of status quo and the consequent thematization of both ideology as objective force, active in history, and ideology as cultural hegemony of ruling class on subordinate groups. Among the Marxists, the French philosopher Louis Althusser also gives his contribution to this crucial question; as follower of Gramsci’s thought, he develops the idea of ideology as an objective force through the notions of Repressive State Apparatus (RSA) and Ideological State Apparatuses (ISA). In addition to this, his philosophy is characterized by the presence of structuralist elements, which must be studied, since they deeply change the theoretical foundation of his Marxist thought.

Keywords: Althusser, enlightenment, Gramsci, ideology

Procedia PDF Downloads 183
5176 Conceptional Design of a Hyperloop Capsule with Linear Induction Propulsion System

Authors: Ahmed E. Hodaib, Samar F. Abdel Fattah

Abstract:

High-speed transportation is a growing concern. To develop high-speed rails and to increase high-speed efficiencies, the idea of Hyperloop was introduced. The challenge is to overcome the difficulties of managing friction and air-resistance which become substantial when vehicles approach high speeds. In this paper, we are presenting the methodologies of the capsule design which got a design concept innovation award at SpaceX competition in January, 2016. MATLAB scripts are written for the levitation and propulsion calculations and iterations. Computational Fluid Dynamics (CFD) is used to simulate the air flow around the capsule considering the effect of the axial-flow air compressor and the levitation cushion on the air flow. The design procedures of a single-sided linear induction motor are analyzed in detail and its geometric and magnetic parameters are determined. A structural design is introduced and Finite Element Method (FEM) is used to analyze the stresses in different parts. The configuration and the arrangement of the components are illustrated. Moreover, comments on manufacturing are made.

Keywords: high-speed transportation, hyperloop, railways transportation, single-sided linear induction Motor (SLIM)

Procedia PDF Downloads 268
5175 Using Equipment Telemetry Data for Condition-Based maintenance decisions

Authors: John Q. Todd

Abstract:

Given that modern equipment can provide comprehensive health, status, and error condition data via built-in sensors, maintenance organizations have a new and valuable source of insight to take advantage of. This presentation will expose what these data payloads might look like and how they can be filtered, visualized, calculated into metrics, used for machine learning, and generate alerts for further action.

Keywords: condition based maintenance, equipment data, metrics, alerts

Procedia PDF Downloads 169
5174 Influence of Single and Multiple Skin-Core Debonding on Free Vibration Characteristics of Innovative GFRP Sandwich Panels

Authors: Indunil Jayatilake, Warna Karunasena, Weena Lokuge

Abstract:

An Australian manufacturer has fabricated an innovative GFRP sandwich panel made from E-glass fiber skin and a modified phenolic core for structural applications. Debonding, which refers to separation of skin from the core material in composite sandwiches, is one of the most common types of damage in composites. The presence of debonding is of great concern because it not only severely affects the stiffness but also modifies the dynamic behaviour of the structure. Generally, it is seen that the majority of research carried out has been concerned about the delamination of laminated structures whereas skin-core debonding has received relatively minor attention. Furthermore, it is observed that research done on composite slabs having multiple skin-core debonding is very limited. To address this gap, a comprehensive research investigating dynamic behaviour of composite panels with single and multiple debonding is presented. The study uses finite-element modelling and analyses for investigating the influence of debonding on free vibration behaviour of single and multilayer composite sandwich panels. A broad parametric investigation has been carried out by varying debonding locations, debonding sizes and support conditions of the panels in view of both single and multiple debonding. Numerical models were developed with Strand7 finite element package by innovatively selecting the suitable elements to diligently represent their actual behavior. Three-dimensional finite element models were employed to simulate the physically real situation as close as possible, with the use of an experimentally and numerically validated finite element model. Comparative results and conclusions based on the analyses are presented. For similar extents and locations of debonding, the effect of debonding on natural frequencies appears greatly dependent on the end conditions of the panel, giving greater decrease in natural frequency when the panels are more restrained. Some modes are more sensitive to debonding and this sensitivity seems to be related to their vibration mode shapes. The fundamental mode seems generally the least sensitive mode to debonding with respect to the variation in free vibration characteristics. The results indicate the effectiveness of the developed three-dimensional finite element models in assessing debonding damage in composite sandwich panels

Keywords: debonding, free vibration behaviour, GFRP sandwich panels, three dimensional finite element modelling

Procedia PDF Downloads 303
5173 Homeless Population Modeling and Trend Prediction Through Identifying Key Factors and Machine Learning

Authors: Shayla He

Abstract:

Background and Purpose: According to Chamie (2017), it’s estimated that no less than 150 million people, or about 2 percent of the world’s population, are homeless. The homeless population in the United States has grown rapidly in the past four decades. In New York City, the sheltered homeless population has increased from 12,830 in 1983 to 62,679 in 2020. Knowing the trend on the homeless population is crucial at helping the states and the cities make affordable housing plans, and other community service plans ahead of time to better prepare for the situation. This study utilized the data from New York City, examined the key factors associated with the homelessness, and developed systematic modeling to predict homeless populations of the future. Using the best model developed, named HP-RNN, an analysis on the homeless population change during the months of 2020 and 2021, which were impacted by the COVID-19 pandemic, was conducted. Moreover, HP-RNN was tested on the data from Seattle. Methods: The methodology involves four phases in developing robust prediction methods. Phase 1 gathered and analyzed raw data of homeless population and demographic conditions from five urban centers. Phase 2 identified the key factors that contribute to the rate of homelessness. In Phase 3, three models were built using Linear Regression, Random Forest, and Recurrent Neural Network (RNN), respectively, to predict the future trend of society's homeless population. Each model was trained and tuned based on the dataset from New York City for its accuracy measured by Mean Squared Error (MSE). In Phase 4, the final phase, the best model from Phase 3 was evaluated using the data from Seattle that was not part of the model training and tuning process in Phase 3. Results: Compared to the Linear Regression based model used by HUD et al (2019), HP-RNN significantly improved the prediction metrics of Coefficient of Determination (R2) from -11.73 to 0.88 and MSE by 99%. HP-RNN was then validated on the data from Seattle, WA, which showed a peak %error of 14.5% between the actual and the predicted count. Finally, the modeling results were collected to predict the trend during the COVID-19 pandemic. It shows a good correlation between the actual and the predicted homeless population, with the peak %error less than 8.6%. Conclusions and Implications: This work is the first work to apply RNN to model the time series of the homeless related data. The Model shows a close correlation between the actual and the predicted homeless population. There are two major implications of this result. First, the model can be used to predict the homeless population for the next several years, and the prediction can help the states and the cities plan ahead on affordable housing allocation and other community service to better prepare for the future. Moreover, this prediction can serve as a reference to policy makers and legislators as they seek to make changes that may impact the factors closely associated with the future homeless population trend.

Keywords: homeless, prediction, model, RNN

Procedia PDF Downloads 110
5172 Jean-Francois Lyotrard's Concept of Different and the Conceptual Problems of Beauty in Philosophy of Contemporary Art

Authors: Sunandapriya Bhikkhu, Shimo Sraman

Abstract:

The main objective of this research is to analytically study the concept of Lyotard’s different that rejects the monopoly criteria and single rule with the incommensurable, which can explain about conceptual problems of beauty in the philosophy of contemporary art. In Lyotard’s idea that basic value judgment of human should be a value like a phrase that is a small unit and an individual such as the aesthetic value that to explain the art world. From the concept of the anti-war artist that rejects the concept of the traditional aesthetic which cannot be able to explain the changing in contemporary society but emphasizes the meaning of individual beauty that is at the beginning of contemporary art today. In the analysis of the problem, the researcher supports the concept of Lyotard’s different that emphasizes the artistic expression which opens the space of perception and beyond the limitations of language process. Art is like phrase or small units that can convey a sense of humanity through the aesthetic value of the individual, not social criteria or universal. The concept of Lyotard’s different awakens and challenge us to the rejection of the single rule that is not open the social space to minorities by not accepting the monopoly criteria.

Keywords: difference, Jean-Francois Lyotard, postmodern, beauty, contemporary art

Procedia PDF Downloads 292
5171 Temperature Dependence of Relative Permittivity: A Measurement Technique Using Split Ring Resonators

Authors: Sreedevi P. Chakyar, Jolly Andrews, V. P. Joseph

Abstract:

A compact method for measuring the relative permittivity of a dielectric material at different temperatures using a single circular Split Ring Resonator (SRR) metamaterial unit working as a test probe is presented in this paper. The dielectric constant of a material is dependent upon its temperature and the LC resonance of the SRR depends on its dielectric environment. Hence, the temperature of the dielectric material in contact with the resonator influences its resonant frequency. A single SRR placed between transmitting and receiving probes connected to a Vector Network Analyser (VNA) is used as a test probe. The dependence of temperature between 30 oC and 60 oC on resonant frequency of SRR is analysed. Relative permittivities ‘ε’ of test samples for different temperatures are extracted from a calibration graph drawn between the relative permittivity of samples of known dielectric constant and their corresponding resonant frequencies. This method is found to be an easy and efficient technique for analysing the temperature dependent permittivity of different materials.

Keywords: metamaterials, negative permeability, permittivity measurement techniques, split ring resonators, temperature dependent dielectric constant

Procedia PDF Downloads 396
5170 Correction Factors for Soil-Structure Interaction Predicted by Simplified Models: Axisymmetric 3D Model versus Fully 3D Model

Authors: Fu Jia

Abstract:

The effects of soil-structure interaction (SSI) are often studied using axial-symmetric three-dimensional (3D) models to avoid the high computational cost of the more realistic, fully 3D models, which require 2-3 orders of magnitude more computer time and storage. This paper analyzes the error and presents correction factors for system frequency, system damping, and peak amplitude of structural response computed by axisymmetric models, embedded in uniform or layered half-space. The results are compared with those for fully 3D rectangular foundations of different aspect ratios. Correction factors are presented for a range of the model parameters, such as fixed-base frequency, structure mass, height and length-to-width ratio, foundation embedment, soil-layer stiffness and thickness. It is shown that the errors are larger for stiffer, taller and heavier structures, deeper foundations and deeper soil layer. For example, for a stiff structure like Millikan Library (NS response; length-to-width ratio 1), the error is 6.5% in system frequency, 49% in system damping and 180% in peak amplitude. Analysis of a case study shows that the NEHRP-2015 provisions for reduction of base shear force due to SSI effects may be unsafe for some structures and need revision. The presented correction factor diagrams can be used in practical design and other applications.

Keywords: 3D soil-structure interaction, correction factors for axisymmetric models, length-to-width ratio, NEHRP-2015 provisions for reduction of base shear force, rectangular embedded foundations, SSI system frequency, SSI system damping

Procedia PDF Downloads 248
5169 Experimental Investigation on Effect of the Zirconium + Magnesium Coating of the Piston and Valve of the Single-Cylinder Diesel Engine to the Engine Performance and Emission

Authors: Erdinç Vural, Bülent Özdalyan, Serkan Özel

Abstract:

The four-stroke single cylinder diesel engine has been used in this study, the pistons and valves of the engine have been stabilized, the aluminum oxide (Al2O3) in different ratios has been added in the power of zirconium (ZrO2) magnesium oxide (MgO), and has been coated with the plasma spray method. The pistons and valves of the combustion chamber of the engine are coated with 5 different (ZrO2 + MgO), (ZrO2 + MgO + 25% Al2O3), (ZrO2 + MgO + 50% Al2O3), (ZrO2 + MgO + 75% Al2O3), (Al2O3) sample. The material tests have been made for each of the coated engine parts with the scanning electron microscopy (SEM), energy dispersive X-ray spectroscopy (EDX) and X-ray diffraction (XRD) using Cu Kα radiation surface analysis methods. The engine tests have been repeated for each sample in any electric dynamometer in full power 1600 rpm, 2000 rpm, 2400 rpm and 2800 rpm engine speeds. The material analysis and engine tests have shown that the best performance has been performed with (ZrO2 + MgO + 50% Al2O3). Thus, there is no significant change in HC and Smoke emissions, but NOx emission is increased, as the engine improves power, torque, specific fuel consumption and CO emissions in the tests made with sample A3.

Keywords: ceramic coating, material characterization, engine performance, exhaust emissions

Procedia PDF Downloads 361
5168 Reaction Rate Behavior of a Methane-Air Mixture over a Platinum Catalyst in a Single Channel Catalytic Reactor

Authors: Doo Ki Lee, Kumaresh Selvakumar, Man Young Kim

Abstract:

Catalytic combustion is an environmentally friendly technique to combust fuels in gas turbines. In this paper, the behavior of surface reaction rate on catalytic combustion is studied with respect to the heterogeneous oxidation of methane-air mixture in a catalytic reactor. Plug flow reactor (PFR), the simplified single catalytic channel assists in investigating the catalytic combustion phenomenon over the Pt catalyst by promoting the desired chemical reactions. The numerical simulation with multi-step elementary surface reactions is governed by the availability of free surface sites onto the catalytic surface and thereby, the catalytic combustion characteristics are demonstrated by examining the rate of the reaction for lean fuel mixture. Further, two different surface reaction mechanisms are adopted and compared for surface reaction rates to indicate the controlling heterogeneous reaction for better fuel conversion. The performance of platinum catalyst under heterogeneous reaction is analyzed under the same temperature condition, where the catalyst with the higher kinetic rate of reaction would have a maximum catalytic activity for enhanced methane catalytic combustion.

Keywords: catalytic combustion, heterogeneous reaction, plug flow reactor, surface reaction rate

Procedia PDF Downloads 262
5167 The Transport of Radical Species to Single and Double Strand Breaks in the Liver’s DNA Molecule by a Hybrid Method of Type Monte Carlo - Diffusion Equation

Authors: H. Oudira, A. Saifi

Abstract:

The therapeutic utility of certain Auger emitters such as iodine-125 depends on their position within the cell nucleus . Or diagnostically, and to maintain as low as possible cell damage, it is preferable to have radionuclide localized outside the cell or at least the core. One solution to this problem is to consider markers capable of conveying anticancer drugs to the tumor site regardless of their location within the human body. The objective of this study is to simulate the impact of a complex such as bleomycin on single and double strand breaks in the DNA molecule. Indeed, this simulation consists of the following transactions: - Construction of BLM -Fe- DNA complex. - Simulation of the electron’s transport from the metastable state excitation of Fe 57 by the Monte Carlo method. - Treatment of chemical reactions in the considered environment by the diffusion equation. For physical, physico-chemical and finally chemical steps, the geometry of the complex is considered as a sphere of 50 nm centered on the binding site , and the mathematical method used is called step by step based on Monte Carlo codes.

Keywords: concentration, yield, radical species, bleomycin, excitation, DNA

Procedia PDF Downloads 447
5166 Application of Artificial Neural Network for Single Horizontal Bare Tube and Bare Tube Bundles (Staggered) of Large Particles: Heat Transfer Prediction

Authors: G. Ravindranath, S. Savitha

Abstract:

This paper presents heat transfer analysis of single horizontal bare tube and heat transfer analysis of staggered arrangement of bare tube bundles bare tube bundles in gas-solid (air-solid) fluidized bed and predictions are done by using Artificial Neural Network (ANN) based on experimental data. Fluidized bed provide nearly isothermal environment with high heat transfer rate to submerged objects i.e. due to through mixing and large contact area between the gas and the particle, a fully fluidized bed has little temperature variation and gas leaves at a temperature which is close to that of the bed. Measurement of average heat transfer coefficient was made by local thermal simulation technique in a cold bubbling air-fluidized bed of size 0.305 m. x 0.305 m. Studies were conducted for single horizontal Bare Tube of length 305mm and 28.6mm outer diameter and for bare tube bundles of staggered arrangement using beds of large (average particle diameter greater than 1 mm) particle (raagi and mustard). Within the range of experimental conditions influence of bed particle diameter ( Dp), Fluidizing Velocity (U) were studied, which are significant parameters affecting heat transfer. Artificial Neural Networks (ANNs) have been receiving an increasing attention for simulating engineering systems due to some interesting characteristics such as learning capability, fault tolerance, and non-linearity. Here, feed-forward architecture and trained by back-propagation technique is adopted to predict heat transfer analysis found from experimental results. The ANN is designed to suit the present system which has 3 inputs and 2 out puts. The network predictions are found to be in very good agreement with the experimental observed values of bare heat transfer coefficient (hb) and nusselt number of bare tube (Nub).

Keywords: fluidized bed, large particles, particle diameter, ANN

Procedia PDF Downloads 354
5165 One vs. Rest and Error Correcting Output Codes Principled Rebalancing Schemes for Solving Imbalanced Multiclass Problems

Authors: Alvaro Callejas-Ramos, Lorena Alvarez-Perez, Alexander Benitez-Buenache, Anibal R. Figueiras-Vidal

Abstract:

This contribution presents a promising formulation which allows to extend the principled binary rebalancing procedures, also known as neutral re-balancing mechanisms in the sense that they do not alter the likelihood ratio

Keywords: Bregman divergences, imbalanced multiclass classifi-cation, informed re-balancing, invariant likelihood ratio

Procedia PDF Downloads 198
5164 A One-Dimensional Modeling Analysis of the Influence of Swirl and Tumble Coefficient in a Single-Cylinder Research Engine

Authors: Mateus Silva Mendonça, Wender Pereira de Oliveira, Gabriel Heleno de Paula Araújo, Hiago Tenório Teixeira Santana Rocha, Augusto César Teixeira Malaquias, José Guilherme Coelho Baeta

Abstract:

The stricter legislation and the greater demand of the population regard to gas emissions and their effects on the environment as well as on human health make the automotive industry reinforce research focused on reducing levels of contamination. This reduction can be achieved through the implementation of improvements in internal combustion engines in such a way that they promote the reduction of both specific fuel consumption and air pollutant emissions. These improvements can be obtained through numerical simulation, which is a technique that works together with experimental tests. The aim of this paper is to build, with support of the GT-Suite software, a one-dimensional model of a single-cylinder research engine to analyze the impact of the variation of swirl and tumble coefficients on the performance and on the air pollutant emissions of an engine. Initially, the discharge coefficient is calculated through the software Converge CFD 3D, given that it is an input parameter in GT-Power. Mesh sensitivity tests are made in 3D geometry built for this purpose, using the mass flow rate in the valve as a reference. In the one-dimensional simulation is adopted the non-predictive combustion model called Three Pressure Analysis (TPA) is, and then data such as mass trapped in cylinder, heat release rate, and accumulated released energy are calculated, aiming that the validation can be performed by comparing these data with those obtained experimentally. Finally, the swirl and tumble coefficients are introduced in their corresponding objects so that their influences can be observed when compared to the results obtained previously.

Keywords: 1D simulation, single-cylinder research engine, swirl coefficient, three pressure analysis, tumble coefficient

Procedia PDF Downloads 90
5163 Complex Cooling Approach in Microchannel Heat Exchangers Using Solid and Hollow Fins

Authors: Nahum Yustus Godi

Abstract:

A three-dimensional numerical optimisation of combined microchannels with constructal solid, half hollow, and hollow circular fins is documented in this paper. The technique seeks to minimize peak temperature in the entire volume of the microchannel heat sink. The volume and axial length were all fixed, while the width of the microchannel could morph. High-density heat flux was applied at the bottom wall of the microchannel. The coolant employed to remove the heat deposited at the bottom surface of the microchannel was a single-phase fluid (water) in a forced convection laminar condition, and heat transfer was a conjugate problem. The unit cell symmetrical computation domain was discretised, and governing equations were solved using computational fluid dynamic (CFD) code. The results reveal that the combined microchannel with hollow circular fins and solid fins performed better at different Reynolds numbers. The numerical study was validated for the single microchannel without fins and found to be in good agreement with previous studies.

Keywords: constructal fins, complex heat exchangers, cooling technique, numerical optimisation

Procedia PDF Downloads 211
5162 New Concept for Real Time Selective Harmonics Elimination Based on Lagrange Interpolation Polynomials

Authors: B. Makhlouf, O. Bouchhida, M. Nibouche, K. Laidi

Abstract:

A variety of methods for selective harmonics elimination pulse width modulation have been developed, the most frequently used for real-time implementation based on look-up tables method. To address real-time requirements based in modified carrier signal is proposed in the presented work, with a general formulation to real-time harmonics control/elimination in switched inverters. Firstly, the proposed method has been demonstrated for a single value of the modulation index. However, in reality, this parameter is variable as a consequence of the voltage (amplitude) variability. In this context, a simple interpolation method for calculating the modified sine carrier signal is proposed. The method allows a continuous adjustment in both amplitude and frequency of the fundamental. To assess the performance of the proposed method, software simulations and hardware experiments have been carried out in the case of a single-phase inverter. Obtained results are very satisfactory.

Keywords: harmonic elimination, Particle Swarm Optimisation (PSO), polynomial interpolation, pulse width modulation, real-time harmonics control, voltage inverter

Procedia PDF Downloads 495
5161 Comparative Assessment of Geocell and Geogrid Reinforcement for Flexible Pavement: Numerical Parametric Study

Authors: Anjana R. Menon, Anjana Bhasi

Abstract:

Development of highways and railways play crucial role in a nation’s economic growth. While rigid concrete pavements are durable with high load bearing characteristics, growing economies mostly rely on flexible pavements which are easier in construction and more economical. The strength of flexible pavement is based on the strength of subgrade and load distribution characteristics of intermediate granular layers. In this scenario, to simultaneously meet economy and strength criteria, it is imperative to strengthen and stabilize the load transferring layers, namely subbase and base. Geosynthetic reinforcement in planar and cellular forms have been proven effective in improving soil stiffness and providing a stable load transfer platform. Studies have proven the relative superiority of cellular form-geocells over planar geosynthetic forms like geogrid, owing to the additional confinement of infill material and pocket effect arising from vertical deformation. Hence, the present study investigates the efficiency of geocells over single/multiple layer geogrid reinforcements by a series of three-dimensional model analyses of a flexible pavement section under a standard repetitive wheel load. The stress transfer mechanism and deformation profiles under various reinforcement configurations are also studied. Geocell reinforcement is observed to take up a higher proportion of stress caused by the traffic loads compared to single and double-layer geogrid reinforcements. The efficiency of single geogrid reinforcement reduces with an increase in embedment depth. The contribution of lower geogrid is insignificant in the case of the double-geogrid reinforced system.

Keywords: Geocell, Geogrid, Flexible Pavement, Repetitive Wheel Load, Numerical Analysis

Procedia PDF Downloads 67
5160 High-Throughput Screening and Selection of Electrogenic Microbial Communities Using Single Chamber Microbial Fuel Cells Based on 96-Well Plate Array

Authors: Lukasz Szydlowski, Jiri Ehlich, Igor Goryanin

Abstract:

We demonstrate a single chamber, 96-well-plated based Microbial Fuel Cell (MFC) with printed, electronic components. This invention is aimed at robust selection of electrogenic microbial community under specific conditions, e.g., electrode potential, pH, nutrient concentration, salt concentration that can be altered within the 96 well plate array. This invention enables robust selection of electrogenic microbial community under the homogeneous reactor, with multiple conditions that can be altered to allow comparative analysis. It can be used as a standalone technique or in conjunction with other selective processes, e.g., flow cytometry, microfluidic-based dielectrophoretic trapping. Mobile conductive elements, like carbon paper, carbon sponge, activated charcoal granules, metal mesh, can be inserted inside to increase the anode surface area in order to collect electrogenic microorganisms and to transfer them into new reactors or for other analytical works. An array of 96-well plate allows this device to be operated by automated pipetting stations.

Keywords: bioengineering, electrochemistry, electromicrobiology, microbial fuel cell

Procedia PDF Downloads 131
5159 Predicting Stem Borer Density in Maize Using RapidEye Data and Generalized Linear Models

Authors: Elfatih M. Abdel-Rahman, Tobias Landmann, Richard Kyalo, George Ong’amo, Bruno Le Ru

Abstract:

Maize (Zea mays L.) is a major staple food crop in Africa, particularly in the eastern region of the continent. The maize growing area in Africa spans over 25 million ha and 84% of rural households in Africa cultivate maize mainly as a means to generate food and income. Average maize yields in Sub Saharan Africa are 1.4 t/ha as compared to global average of 2.5–3.9 t/ha due to biotic and abiotic constraints. Amongst the biotic production constraints in Africa, stem borers are the most injurious. In East Africa, yield losses due to stem borers are currently estimated between 12% to 40% of the total production. The objective of the present study was therefore to predict stem borer larvae density in maize fields using RapidEye reflectance data and generalized linear models (GLMs). RapidEye images were captured for a test site in Kenya (Machakos) in January and in February 2015. Stem borer larva numbers were modeled using GLMs assuming Poisson (Po) and negative binomial (NB) distributions with error with log arithmetic link. Root mean square error (RMSE) and ratio prediction to deviation (RPD) statistics were employed to assess the models performance using a leave one-out cross-validation approach. Results showed that NB models outperformed Po ones in all study sites. RMSE and RPD ranged between 0.95 and 2.70, and between 2.39 and 6.81, respectively. Overall, all models performed similar when used the January and the February image data. We conclude that reflectance data from RapidEye data can be used to estimate stem borer larvae density. The developed models could to improve decision making regarding controlling maize stem borers using various integrated pest management (IPM) protocols.

Keywords: maize, stem borers, density, RapidEye, GLM

Procedia PDF Downloads 485
5158 Maximum Likelihood Estimation Methods on a Two-Parameter Rayleigh Distribution under Progressive Type-Ii Censoring

Authors: Daniel Fundi Murithi

Abstract:

Data from economic, social, clinical, and industrial studies are in some way incomplete or incorrect due to censoring. Such data may have adverse effects if used in the estimation problem. We propose the use of Maximum Likelihood Estimation (MLE) under a progressive type-II censoring scheme to remedy this problem. In particular, maximum likelihood estimates (MLEs) for the location (µ) and scale (λ) parameters of two Parameter Rayleigh distribution are realized under a progressive type-II censoring scheme using the Expectation-Maximization (EM) and the Newton-Raphson (NR) algorithms. These algorithms are used comparatively because they iteratively produce satisfactory results in the estimation problem. The progressively type-II censoring scheme is used because it allows the removal of test units before the termination of the experiment. Approximate asymptotic variances and confidence intervals for the location and scale parameters are derived/constructed. The efficiency of EM and the NR algorithms is compared given root mean squared error (RMSE), bias, and the coverage rate. The simulation study showed that in most sets of simulation cases, the estimates obtained using the Expectation-maximization algorithm had small biases, small variances, narrower/small confidence intervals width, and small root of mean squared error compared to those generated via the Newton-Raphson (NR) algorithm. Further, the analysis of a real-life data set (data from simple experimental trials) showed that the Expectation-Maximization (EM) algorithm performs better compared to Newton-Raphson (NR) algorithm in all simulation cases under the progressive type-II censoring scheme.

Keywords: expectation-maximization algorithm, maximum likelihood estimation, Newton-Raphson method, two-parameter Rayleigh distribution, progressive type-II censoring

Procedia PDF Downloads 152
5157 Modified Lot Quality Assurance Sampling (LQAS) Model for Quality Assessment of Malaria Parasite Microscopy and Rapid Diagnostic Tests in Kano, Nigeria

Authors: F. Sarkinfada, Dabo N. Tukur, Abbas A. Muaz, Adamu A. Yahuza

Abstract:

Appropriate Quality Assurance (QA) of parasite-based diagnosis of malaria to justify Artemisinin-based Combination Therapy (ACT) is essential for Malaria Programmes. In Low and Middle Income Countries (LMIC), resource constrain appears to be a major challenge in implementing the conventional QA system. We designed and implemented a modified LQAS model for QA of malaria parasite (MP) microscopy and RDT in a State Specialist Hospital (SSH) and a University Health Clinic (UHC) in Kano, Nigeria. The capacities of both facilities for MP microscopy and RDT were assessed before implementing a modified LQAS over a period of 3 months. Quality indicators comprising the qualities of blood film and staining, MP positivity rates, concordance rates, error rates (in terms of false positives and false negatives), sensitivity and specificity were monitored and evaluated. Seventy one percent (71%) of the basic requirements for malaria microscopy was available in both facilities, with the absence of certifies microscopists, SOPs and Quality Assurance mechanisms. A daily average of 16 to 32 blood samples were tested with a blood film staining quality of >70% recorded in both facilities. Using microscopy, the MP positivity rates were 50.46% and 19.44% in SSH and UHS respectively, while the MP positivity rates were 45.83% and 22.78% in SSH and UHS when RDT was used. Higher concordance rates of 88.90% and 93.98% were recorded in SSH and UHC respectively using microscopy, while lower rates of 74.07% and 80.58% in SSH and UHC were recorded when RDT was used. In both facilities, error rates were higher when RDT was used than with microscopy. Sensitivity and specificity were higher when microscopy was used (95% and 84% in SSH; 94% in UHC) than when RDT was used (72% and 76% in SSH; 78% and 81% in UHC). It could be feasible to implement an integrated QA model for MP microscopy and RDT using modified LQAS in Malaria Control Programmes in Low and Middle Income Countries that might have resource constrain for parasite-base diagnosis of malaria to justify ACT treatment.

Keywords: malaria, microscopy, quality assurance, RDT

Procedia PDF Downloads 211
5156 Heat Transfer Coefficients of Layers of Greenhouse Thermal Screens

Authors: Vitaly Haslavsky, Helena Vitoshkin

Abstract:

The total energy saving effect of different types of greenhouse thermal/shade screens was determined by measuring and calculating the overall heat transfer coefficients (U-values) for single and several layers of screens. The measurements were carried out using the hot box method, and the calculations were performed according to the ISO Standard 15099. The goal was to examine different types of materials with a wide range of thermal radiation properties used for thermal screens in combination with a dehumidification system in order to improve greenhouse insulation. The experimental results were in good agreement with the calculated heat transfer coefficients. It was shown that a high amount of infra-red (IR) radiation can be blocked by the greenhouse covering material in combination with moveable thermal screens. The aluminum foil screen could be replaced by transparent screens, depending on shading requirements. The results indicated that using a single layer, the U-value was reduced by approximately 70% compared to covering material alone, while the contributions of additional screen layers containing aluminum foil strips could reduce the U-value by approximately 90%. It was shown that three screen layers are sufficient for effective insulation.

Keywords: greenhouse insulation, heat loss, thermal screens, U-value

Procedia PDF Downloads 101
5155 Modeling and Temperature Control of Water-cooled PEMFC System Using Intelligent Algorithm

Authors: Chen Jun-Hong, He Pu, Tao Wen-Quan

Abstract:

Proton exchange membrane fuel cell (PEMFC) is the most promising future energy source owing to its low operating temperature, high energy efficiency, high power density, and environmental friendliness. In this paper, a comprehensive PEMFC system control-oriented model is developed in the Matlab/Simulink environment, which includes the hydrogen supply subsystem, air supply subsystem, and thermal management subsystem. Besides, Improved Artificial Bee Colony (IABC) is used in the parameter identification of PEMFC semi-empirical equations, making the maximum relative error between simulation data and the experimental data less than 0.4%. Operation temperature is essential for PEMFC, both high and low temperatures are disadvantageous. In the thermal management subsystem, water pump and fan are both controlled with the PID controller to maintain the appreciate operation temperature of PEMFC for the requirements of safe and efficient operation. To improve the control effect further, fuzzy control is introduced to optimize the PID controller of the pump, and the Radial Basis Function (RBF) neural network is introduced to optimize the PID controller of the fan. The results demonstrate that Fuzzy-PID and RBF-PID can achieve a better control effect with 22.66% decrease in Integral Absolute Error Criterion (IAE) of T_st (Temperature of PEMFC) and 77.56% decrease in IAE of T_in (Temperature of inlet cooling water) compared with traditional PID. In the end, a novel thermal management structure is proposed, which uses the cooling air passing through the main radiator to continue cooling the secondary radiator. In this thermal management structure, the parasitic power dissipation can be reduced by 69.94%, and the control effect can be improved with a 52.88% decrease in IAE of T_in under the same controller.

Keywords: PEMFC system, parameter identification, temperature control, Fuzzy-PID, RBF-PID, parasitic power

Procedia PDF Downloads 70
5154 Evaluation of a Piecewise Linear Mixed-Effects Model in the Analysis of Randomized Cross-over Trial

Authors: Moses Mwangi, Geert Verbeke, Geert Molenberghs

Abstract:

Cross-over designs are commonly used in randomized clinical trials to estimate efficacy of a new treatment with respect to a reference treatment (placebo or standard). The main advantage of using cross-over design over conventional parallel design is its flexibility, where every subject become its own control, thereby reducing confounding effect. Jones & Kenward, discuss in detail more recent developments in the analysis of cross-over trials. We revisit the simple piecewise linear mixed-effects model, proposed by Mwangi et. al, (in press) for its first application in the analysis of cross-over trials. We compared performance of the proposed piecewise linear mixed-effects model with two commonly cited statistical models namely, (1) Grizzle model; and (2) Jones & Kenward model, used in estimation of the treatment effect, in the analysis of randomized cross-over trial. We estimate two performance measurements (mean square error (MSE) and coverage probability) for the three methods, using data simulated from the proposed piecewise linear mixed-effects model. Piecewise linear mixed-effects model yielded lowest MSE estimates compared to Grizzle and Jones & Kenward models for both small (Nobs=20) and large (Nobs=600) sample sizes. It’s coverage probability were highest compared to Grizzle and Jones & Kenward models for both small and large sample sizes. A piecewise linear mixed-effects model is a better estimator of treatment effect than its two competing estimators (Grizzle and Jones & Kenward models) in the analysis of cross-over trials. The data generating mechanism used in this paper captures two time periods for a simple 2-Treatments x 2-Periods cross-over design. Its application is extendible to more complex cross-over designs with multiple treatments and periods. In addition, it is important to note that, even for single response models, adding more random effects increases the complexity of the model and thus may be difficult or impossible to fit in some cases.

Keywords: Evaluation, Grizzle model, Jones & Kenward model, Performance measures, Simulation

Procedia PDF Downloads 113
5153 The Analysis of Gizmos Online Program as Mathematics Diagnostic Program: A Story from an Indonesian Private School

Authors: Shofiayuningtyas Luftiani

Abstract:

Some private schools in Indonesia started integrating the online program Gizmos in the teaching-learning process. Gizmos was developed to supplement the existing curriculum by integrating it into the instructional programs. The program has some features using an inquiry-based simulation, in which students conduct exploration by using a worksheet while teachers use the teacher guidelines to direct and assess students’ performance In this study, the discussion about Gizmos highlights its features as the assessment media of mathematics learning for secondary school students. The discussion is based on the case study and literature review from the Indonesian context. The purpose of applying Gizmos as an assessment media refers to the diagnostic assessment. As a part of the diagnostic assessment, the teachers review the student exploration sheet, analyze particularly in the students’ difficulties and consider findings in planning future learning process. This assessment becomes important since the teacher needs the data about students’ persistent weaknesses. Additionally, this program also helps to build student’ understanding by its interactive simulation. Currently, the assessment over-emphasizes the students’ answers in the worksheet based on the provided answer keys while students perform their skill in translating the question, doing the simulation and answering the question. Whereas, the assessment should involve the multiple perspectives and sources of students’ performance since teacher should adjust the instructional programs with the complexity of students’ learning needs and styles. Consequently, the approach to improving the assessment components is selected to challenge the current assessment. The purpose of this challenge is to involve not only the cognitive diagnosis but also the analysis of skills and error. Concerning the selected setting for this diagnostic assessment that develops the combination of cognitive diagnosis, skills analysis and error analysis, the teachers should create an assessment rubric. The rubric plays the important role as the guide to provide a set of criteria for the assessment. Without the precise rubric, the teacher potentially ineffectively documents and follows up the data about students at risk of failure. Furthermore, the teachers who employ the program of Gizmos as the diagnostic assessment might encounter some obstacles. Based on the condition of assessment in the selected setting, the obstacles involve the time constrain, the reluctance of higher teaching burden and the students’ behavior. Consequently, the teacher who chooses the Gizmos with those approaches has to plan, implement and evaluate the assessment. The main point of this assessment is not in the result of students’ worksheet. However, the diagnostic assessment has the two-stage process; the process to prompt and effectively follow-up both individual weaknesses and those of the learning process. Ultimately, the discussion of Gizmos as the media of the diagnostic assessment refers to the effort to improve the mathematical learning process.

Keywords: diagnostic assessment, error analysis, Gizmos online program, skills analysis

Procedia PDF Downloads 172
5152 Nano-Sensors: Search for New Features

Authors: I. Filikhin, B. Vlahovic

Abstract:

We focus on a novel type of detection based on electron tunneling properties of double nanoscale structures in semiconductor materials. Semiconductor heterostructures as quantum wells (QWs), quantum dots (QDs), and quantum rings (QRs) may have energy level structure of several hundred of electron confinement states. The single electron spectra of the double quantum objects (DQW, DQD, and DQR) were studied in our previous works with relation to the electron localization and tunneling between the objects. The wave function of electron may be localized in one of the QDs or be delocalized when it is spread over the whole system. The localizing-delocalizing tunneling occurs when an electron transition between both states is possible. The tunneling properties of spectra differ strongly for “regular” and “chaotic” systems. We have shown that a small violation of the geometry drastically affects localization of electron. In particular, such violations lead to the elimination of the delocalized states of the system. The same symmetry violation effect happens if electrical or magnetic fields are applied. These phenomena could be used to propose a new type of detection based on the high sensitivity of charge transport between double nanostructures and small violations of the shapes. It may have significant technological implications.

Keywords: double quantum dots, single electron levels, tunneling, electron localizations

Procedia PDF Downloads 493
5151 Salt Tolerance of Potato: Genetically Engineered with Atriplex canescens BADH Gene Driven by 3 Copies of CAMV35s Promoter

Authors: Arfan Ali, Muhammad Shahzad Iqbal, Idrees Ahmad Nasir

Abstract:

Potato (Solanum tuberosum L.) is ranked among the top leading staple foods in the world. Salinity adversely affects potato crop yield and quality. Therefore, increased level of salt tolerance is a key factor to ensure high yield. The present study focused on the Agrobacterium-mediated transformation of Atriplex canescens betaine aldehyde dehydrogenase (BADH) gene, using single, double and triple CAMV35s promoter to improve salt tolerance in potato. Detection of seven potato lines harboring BADH gene, followed by identification of T-DNA insertions, determination of transgenes copies no through Southern Hybridization and quantification of BADH protein through Enzyme Linked Immunosorbent Assay were considered in this study. The results clearly depict that the salt tolerance of potato was found to be promoter-dependent, as the potato transgenic lines with triple promoter showed 4.4 times more glycine betaine production which consequently leads towards high resistance to salt stress as compared to transgenic potato lines with single and double promoters having least production of glycine betaine. Moreover, triple promoter transgenic potato lines have also shown lower levels of H2O2, malondialdehyde (MDA), relative electrical conductivity, high proline and chlorophyll content as compared other two lines having a single and double promoter. Insilco analysis also confirmed that Atriplex canescens BADH has the tendency to interact with sodium ions and water molecules. Taken together these facts it can be concluded that over-expression of BADH under triple CAMV35s promoter with more glycine betaine, chlorophyll & MDA contents, high relative quantities of other metabolites results in an enhanced level of salt tolerance in potato.

Keywords: Atriplex canescens, BADH, CAMV35s promotor, potato, Solanum tubersum

Procedia PDF Downloads 263
5150 Design and Simulation of an Inter-Satellite Optical Wireless Communication System Using Diversity Techniques

Authors: Sridhar Rapuru, D. Mallikarjunreddy, Rajanarendra Sai

Abstract:

In this reign of the internet, the access of any multimedia file to the users at any time with a superior quality is needed. To achieve this goal, it is very important to have a good network without any interruptions between the satellites along with various earth stations. For that purpose, a high speed inter-satellite optical wireless communication system (IsOWC) is designed with space and polarization diversity techniques. IsOWC offers a high bandwidth, small size, less power requirement and affordable when compared with the present microwave satellite systems. To improve the efficiency and to reduce the propagation delay, inter-satellite link is established between the satellites. High accurate tracking systems are required to establish the reliable connection between the satellites as they have their own orbits. The only disadvantage of this IsOWC system is laser beam width is narrower than the RF because of this highly accurate tracking system to meet this requirement. The satellite uses the 'ephemerides data' for rough pointing and tracking system for fine pointing to the other satellite. In this proposed IsOWC system, laser light is used as a wireless connectedness between the source and destination and free space acts as the channel to carry the message. The proposed system will be designed, simulated and analyzed for 6000km with an improvement of data rate over previously existing systems. The performance parameters of the system are Q-factor, eye opening, bit error rate, etc., The proposed system for Inter-satellite Optical Wireless Communication System Design Using Diversity Techniques finds huge scope of applications in future generation communication purposes.

Keywords: inter-satellite optical wireless system, space and polarization diversity techniques, line of sight, bit error rate, Q-factor

Procedia PDF Downloads 253
5149 Improving the Accuracy of Stress Intensity Factors Obtained by Scaled Boundary Finite Element Method on Hybrid Quadtree Meshes

Authors: Adrian W. Egger, Savvas P. Triantafyllou, Eleni N. Chatzi

Abstract:

The scaled boundary finite element method (SBFEM) is a semi-analytical numerical method, which introduces a scaling center in each element’s domain, thus transitioning from a Cartesian reference frame to one resembling polar coordinates. Consequently, an analytical solution is achieved in radial direction, implying that only the boundary need be discretized. The only limitation imposed on the resulting polygonal elements is that they remain star-convex. Further arbitrary p- or h-refinement may be applied locally in a mesh. The polygonal nature of SBFEM elements has been exploited in quadtree meshes to alleviate all issues conventionally associated with hanging nodes. Furthermore, since in 2D this results in only 16 possible cell configurations, these are precomputed in order to accelerate the forward analysis significantly. Any cells, which are clipped to accommodate the domain geometry, must be computed conventionally. However, since SBFEM permits polygonal elements, significantly coarser meshes at comparable accuracy levels are obtained when compared with conventional quadtree analysis, further increasing the computational efficiency of this scheme. The generalized stress intensity factors (gSIFs) are computed by exploiting the semi-analytical solution in radial direction. This is initiated by placing the scaling center of the element containing the crack at the crack tip. Taking an analytical limit of this element’s stress field as it approaches the crack tip, delivers an expression for the singular stress field. By applying the problem specific boundary conditions, the geometry correction factor is obtained, and the gSIFs are then evaluated based on their formal definition. Since the SBFEM solution is constructed as a power series, not unlike mode superposition in FEM, the two modes contributing to the singular response of the element can be easily identified in post-processing. Compared to the extended finite element method (XFEM) this approach is highly convenient, since neither enrichment terms nor a priori knowledge of the singularity is required. Computation of the gSIFs by SBFEM permits exceptional accuracy, however, when combined with hybrid quadtrees employing linear elements, this does not always hold. Nevertheless, it has been shown that crack propagation schemes are highly effective even given very coarse discretization since they only rely on the ratio of mode one to mode two gSIFs. The absolute values of the gSIFs may still be subject to large errors. Hence, we propose a post-processing scheme, which minimizes the error resulting from the approximation space of the cracked element, thus limiting the error in the gSIFs to the discretization error of the quadtree mesh. This is achieved by h- and/or p-refinement of the cracked element, which elevates the amount of modes present in the solution. The resulting numerical description of the element is highly accurate, with the main error source now stemming from its boundary displacement solution. Numerical examples show that this post-processing procedure can significantly improve the accuracy of the computed gSIFs with negligible computational cost even on coarse meshes resulting from hybrid quadtrees.

Keywords: linear elastic fracture mechanics, generalized stress intensity factors, scaled finite element method, hybrid quadtrees

Procedia PDF Downloads 131
5148 Use of Anti-Stick to Reduce Bitterness in Ultra Filtrated Chees-es(Single Packaged)

Authors: B. Khorram, M. Taslikh, R. Sattarzadeh, M. Ghazanfari

Abstract:

Bitterness is one of the most important problems in cheese processing industry all over the world. There are several reasons that bitterness may develop in cheese. With a few exceptions bitterness is generally associated with proteolysis. In this investigation, anti-stick as a neutral substance in proteolysis were considered and studied for reducing the problem. This vast survey was conducted in a big cheese production factory (in Neyshabur) and in the same procedure anti-stick as interested factor in cheeses packaging compared to standard cheeses production, one line productions (65200 packs with anti-stick were tested by 2953 persons for bitterness and another line was included the same procedure with standard cheese. In this investigate: 83% of standard packaging cheeses, compared with only28% of consumers cheese with anti-stick which confirmed bitterness. Although bitterness is generally associated with proteolysis and Microbial factors, Somatic cell, Starters play important role in generating bitterness in ultra filtrated cheeses, but based on the results the other factors such as anti-stick in packaging can be effective methods for reducing and removing unfavorable bitterness in cheese production.

Keywords: bitterness, uf cheese, anti-stick, single packaged

Procedia PDF Downloads 460