Search results for: Data Reduction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8605

Search results for: Data Reduction

8455 Diagnosis of the Heart Rhythm Disorders by Using Hybrid Classifiers

Authors: Sule Yucelbas, Gulay Tezel, Cuneyt Yucelbas, Seral Ozsen

Abstract:

In this study, it was tried to identify some heart rhythm disorders by electrocardiography (ECG) data that is taken from MIT-BIH arrhythmia database by subtracting the required features, presenting to artificial neural networks (ANN), artificial immune systems (AIS), artificial neural network based on artificial immune system (AIS-ANN) and particle swarm optimization based artificial neural network (PSO-NN) classifier systems. The main purpose of this study is to evaluate the performance of hybrid AIS-ANN and PSO-ANN classifiers with regard to the ANN and AIS. For this purpose, the normal sinus rhythm (NSR), atrial premature contraction (APC), sinus arrhythmia (SA), ventricular trigeminy (VTI), ventricular tachycardia (VTK) and atrial fibrillation (AF) data for each of the RR intervals were found. Then these data in the form of pairs (NSR-APC, NSR-SA, NSR-VTI, NSR-VTK and NSR-AF) is created by combining discrete wavelet transform which is applied to each of these two groups of data and two different data sets with 9 and 27 features were obtained from each of them after data reduction. Afterwards, the data randomly was firstly mixed within themselves, and then 4-fold cross validation method was applied to create the training and testing data. The training and testing accuracy rates and training time are compared with each other.

As a result, performances of the hybrid classification systems, AIS-ANN and PSO-ANN were seen to be close to the performance of the ANN system. Also, the results of the hybrid systems were much better than AIS, too. However, ANN had much shorter period of training time than other systems. In terms of training times, ANN was followed by PSO-ANN, AIS-ANN and AIS systems respectively. Also, the features that extracted from the data affected the classification results significantly.

Keywords: AIS, ANN, ECG, hybrid classifiers, PSO.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1876
8454 Aerodynamic Analysis of Dimple Effect on Aircraft Wing

Authors: E. Livya, G. Anitha, P. Valli

Abstract:

The main objective of aircraft aerodynamics is to enhance the aerodynamic characteristics and maneuverability of the aircraft. This enhancement includes the reduction in drag and stall phenomenon. The airfoil which contains dimples will have comparatively less drag than the plain airfoil. Introducing dimples on the aircraft wing will create turbulence by creating vortices which delays the boundary layer separation resulting in decrease of pressure drag and also increase in the angle of stall. In addition, wake reduction leads to reduction in acoustic emission. The overall objective of this paper is to improve the aircraft maneuverability by delaying the flow separation point at stall and thereby reducing the drag by applying the dimple effect over the aircraft wing. This project includes both computational and experimental analysis of dimple effect on aircraft wing, using NACA 0018 airfoil. Dimple shapes of Semi-sphere, hexagon, cylinder, square are selected for the analysis; airfoil is tested under the inlet velocity of 30m/s and 60m/s at different angle of attack (5˚, 10˚, 15˚, 20˚, and 25˚). This analysis favors the dimple effect by increasing L/D ratio and thereby providing the maximum aerodynamic efficiency, which provides the enhanced performance for the aircraft.

Keywords: Airfoil, Boundary layer, Dimple effect, Flow separation, Stall reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6102
8453 A Novel Fuzzy Technique for Image Noise Reduction

Authors: Hamed Vahdat Nejad, Hameed Reza Pourreza, Hasan Ebrahimi

Abstract:

A new fuzzy filter is presented for noise reduction of images corrupted with additive noise. The filter consists of two stages. In the first stage, all the pixels of image are processed for determining noisy pixels. For this, a fuzzy rule based system associates a degree to each pixel. The degree of a pixel is a real number in the range [0,1], which denotes a probability that the pixel is not considered as a noisy pixel. In the second stage, another fuzzy rule based system is employed. It uses the output of the previous fuzzy system to perform fuzzy smoothing by weighting the contributions of neighboring pixel values. Experimental results are obtained to show the feasibility of the proposed filter. These results are also compared to other filters by numerical measure and visual inspection.

Keywords: Additive noise, Fuzzy logic, Image processing, Noise reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2081
8452 Complexity Reduction Approach with Jacobi Iterative Method for Solving Composite Trapezoidal Algebraic Equations

Authors: Mohana Sundaram Muthuvalu, Jumat Sulaiman

Abstract:

In this paper, application of the complexity reduction approach based on half- and quarter-sweep iteration concepts with Jacobi iterative method for solving composite trapezoidal (CT) algebraic equations is discussed. The performances of the methods for CT algebraic equations are comparatively studied by their application in solving linear Fredholm integral equations of the second kind. Furthermore, computational complexity analysis and numerical results for three test problems are also included in order to verify performance of the methods.

Keywords: Complexity reduction approach, Composite trapezoidal scheme, Jacobi method, Linear Fredholm integral equations

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1559
8451 GPS Signal Correction to Improve Vehicle Location during Experimental Campaign

Authors: L. Della Ragione, G. Meccariello

Abstract:

In recent years in Italy the progress of the automobile industry, in the field of reduction of emissions values, is very remarkable. Nevertheless their evaluation and reduction is a key problem, especially in the cities, that account for more than 50% of world population. In this paper we dealt with the problem of describing a quantitatively approach for the reconstruction of GPS coordinates and altitude, in the context of correlation study between driving cycles / emission / geographical location, during an experimental campaign realized with some instrumented cars.

Keywords: Air pollution, Driving cycles, GPS signal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1875
8450 The Automated Selective Acquisition System

Authors: Atisthan Wuttimanop, Suchada Rianmora

Abstract:

To support design process for launching the product on time, reverse engineering (RE) process has been introduced for quickly generating 3D CAD model from its physical object. The accuracy of the 3D CAD model depends upon the data acquisition technique selected, contact or non-contact methods. In order to reduce times used for acquiring surface and eliminating noises, the automated selective acquisition system has been developed and presented in this research as the alternative channel for non-contact acquisition technique where the data is selectively and locally scanned contour by contour without performing data reduction process. The results present as the organized contour points which are directly used to generate 3D virtual model. The comparison between the proposed technique and another non-contact scanning technique has been presented and discussed.

Keywords: Automated selective acquisition system, Non-contact acquisition, Reverse engineering, 3D scanners.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1569
8449 Analysis of Heat Exchanger Network of Distillation Unit of Shiraz Oil Refinery

Authors: J. Khorshidi, E. Zare, A.R. Khademi

Abstract:

The reduction of energy consumption through improvements in energy efficiency has become an important goal for all industries, in order to improve the efficiency of the economy, and to reduce the emissions of Co2 caused by power generation. The objective of this paper is to investigate opportunities to increase process energy efficiency at the distillation unit of Shiraz oil refinery in south of Iran. The main aim of the project is to locate energy savings by use of pinch technology and to assess them. At first all the required data of hot and cold streams in preheating section of distillation unit has been extracted from the available flow sheets and then pinch analysis has been conducted. The present case study is a threshold one which does not need any utilities. After running range, targeting several heat exchanger networks were designed with respect to operating conditions and different ΔTmin. The optimal value of ΔTmin was calculated to be 22.3 °C. Based on this optimal value, there will be 5% reduction in annual total cost of heat exchanger network.

Keywords: Pinch technology, heat exchanger network, operating cost.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1642
8448 Fault Detection of Drinking Water Treatment Process Using PCA and Hotelling's T2 Chart

Authors: Joval P George, Dr. Zheng Chen, Philip Shaw

Abstract:

This paper deals with the application of Principal Component Analysis (PCA) and the Hotelling-s T2 Chart, using data collected from a drinking water treatment process. PCA is applied primarily for the dimensional reduction of the collected data. The Hotelling-s T2 control chart was used for the fault detection of the process. The data was taken from a United Utilities Multistage Water Treatment Works downloaded from an Integrated Program Management (IPM) dashboard system. The analysis of the results show that Multivariate Statistical Process Control (MSPC) techniques such as PCA, and control charts such as Hotelling-s T2, can be effectively applied for the early fault detection of continuous multivariable processes such as Drinking Water Treatment. The software package SIMCA-P was used to develop the MSPC models and Hotelling-s T2 Chart from the collected data.

Keywords: Principal component analysis, hotelling's t2 chart, multivariate statistical process control, drinking water treatment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2738
8447 The Role of Heat Pumps for the Decarbonization of European Regions

Authors: D. M. Mongelli, M. De Carli, L. Carnieletto, F. Busato

Abstract:

This research aims to provide a contribution to the reduction of fossil fuels and the consequent reduction of CO2eq emissions for each European region. Simulations have been carried out to replace fossil fuel fired heating boilers with air-to-water heat pumps, when allowed by favorable environmental conditions (outdoor temperature, water temperature in emission systems, etc.). To estimate the potential coverage of high-temperature heat pumps in European regions, the energy profiles of buildings were considered together with the potential coefficient of performance (COP) of heat pumps operating with flow temperature with variable climatic regulation. The electrification potential for heating buildings was estimated by dividing the 38 European countries examined into 179 territorial units. The yields have been calculated in terms of energy savings and CO2eq reduction.

Keywords: Decarbonization, Space heating, Heat pumps, Energy policies.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 122
8446 Reduction of MMP Using Oleophilic Chemicals

Authors: C. L. Voon, M. Awang

Abstract:

CO2 miscible displacement is not feasible in many oil fields due to high reservoir temperature as higher pressure is required to achieve miscibility. The miscibility pressure is far higher than the formation fracture pressure making it impossible to have CO2 miscible displacement. However, by using oleophilic chemicals, minimum miscibility pressure (MMP) could be lowered. The main objective of this research is to find the best oleophilic chemical in MMP reduction using slim-tube test and Vanishing Interfacial Tension (VIT) The chemicals are selected based on the characteristics that it must be oil soluble, low water solubility, have 4 – 8 carbons, semi polar, economical, and safe for human operation. The families of chemicals chosen are carboxylic acid, alcohol, and ketone. The whole experiment would be conducted at 100°C and the best chemical is said to be effective when it is able to lower CO2-crude oil MMP the most. Findings of this research would have great impact to the oil and gas industry in reduction of operation cost for CO2EOR which is applicable to both onshore and offshore operation.

Keywords: Enhanced Oil Recovery, Oleophilic Chemical, Minimum Miscibility Pressure, CO2 Miscible Displacement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2340
8445 Product Ecodesign Approaches in ISO 14001 Certified Companies

Authors: Gregor Radonjič, Aleksandra P. Korda, Damijan Krajnc

Abstract:

The aim of the study was to investigate whether there is the promotion of product ecodesign measures as a result of adopting ISO 14001 certification in manufacturing companies in the Republic of Slovenia. Companies gave the most of their product development attention to waste and energy reduction during manufacturing process and reduction of material consumption per unit of product. Regarding the importance of different ecodesign criteria reduction of material consumption per unit of product was reported as the most important criterion. Less attention is paid to endof- life issues considering recycling or packaging. Most manufacturing enterprises considered ISO 14001 standard as a very useful tool or at least a useful tool helping them to accelerate and establish product ecodesign activities. Two most frequently considered ecodesign drivers are increased competitive advantage and legal requirements and two most important barriers are high development costs and insufficient market demand.

Keywords: ecodesign, environmental management system, ISO 14001, products

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1498
8444 An Evaluation of Carbon Dioxide Emissions Trading among Enterprises -The Tokyo Cap and Trade Program-

Authors: Hiroki Satou, Kayoko Yamamoto

Abstract:

This study aims to propose three evaluation methods to evaluate the Tokyo Cap and Trade Program when emissions trading is performed virtually among enterprises, focusing on carbon dioxide (CO2), which is the only emitted greenhouse gas that tends to increase. The first method clarifies the optimum reduction rate for the highest cost benefit, the second discusses emissions trading among enterprises through market trading, and the third verifies long-term emissions trading during the term of the plan (2010-2019), checking the validity of emissions trading partly using Geographic Information Systems (GIS). The findings of this study can be summarized in the following three points. 1. Since the total cost benefit is the greatest at a 44% reduction rate, it is possible to set it more highly than that of the Tokyo Cap and Trade Program to get more total cost benefit. 2. At a 44% reduction rate, among 320 enterprises, 8 purchasing enterprises and 245 sales enterprises gain profits from emissions trading, and 67 enterprises perform voluntary reduction without conducting emissions trading. Therefore, to further promote emissions trading, it is necessary to increase the sales volumes of emissions trading in addition to sales enterprises by increasing the number of purchasing enterprises. 3. Compared to short-term emissions trading, there are few enterprises which benefit in each year through the long-term emissions trading of the Tokyo Cap and Trade Program. Only 81 enterprises at the most can gain profits from emissions trading in FY 2019. Therefore, by setting the reduction rate more highly, it is necessary to increase the number of enterprises that participate in emissions trading and benefit from the restraint of CO2 emissions.

Keywords: Emissions Trading, Tokyo Cap and Trade Program, Carbon Dioxide (CO2), Global Warming, Geographic Information Systems (GIS)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2136
8443 Adjustment and Scale-Up Strategy of Pilot Liquid Fermentation Process of Azotobacter sp.

Authors: G. Quiroga-Cubides, A. Díaz, M. Gómez

Abstract:

The genus Azotobacter has been widely used as bio-fertilizer due to its significant effects on the stimulation and promotion of plant growth in various agricultural species of commercial interest. In order to obtain significantly viable cellular concentration, a scale-up strategy for a liquid fermentation process (SmF) with two strains of A. chroococcum (named Ac1 and Ac10) was validated and adjusted at laboratory and pilot scale. A batch fermentation process under previously defined conditions was carried out on a biorreactor Infors®, model Minifors of 3.5 L, which served as a baseline for this research. For the purpose of increasing process efficiency, the effect of the reduction of stirring speed was evaluated in combination with a fed-batch-type fermentation laboratory scale. To reproduce the efficiency parameters obtained, a scale-up strategy with geometric and fluid dynamic behavior similarities was evaluated. According to the analysis of variance, this scale-up strategy did not have significant effect on cellular concentration and in laboratory and pilot fermentations (Tukey, p > 0.05). Regarding air consumption, fermentation process at pilot scale showed a reduction of 23% versus the baseline. The percentage of reduction related to energy consumption reduction under laboratory and pilot scale conditions was 96.9% compared with baseline.

Keywords: Azotobacter chroococcum, scale-up, liquid fermentation, fed-batch process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1277
8442 Determining the Maximum Lateral Displacement Due to Sever Earthquakes without Using Nonlinear Analysis

Authors: Mussa Mahmoudi

Abstract:

For Seismic design, it is important to estimate, maximum lateral displacement (inelastic displacement) of the structures due to sever earthquakes for several reasons. Seismic design provisions estimate the maximum roof and storey drifts occurring in major earthquakes by amplifying the drifts of the structures obtained by elastic analysis subjected to seismic design load, with a coefficient named “displacement amplification factor" which is greater than one. Here, this coefficient depends on various parameters, such as ductility and overstrength factors. The present research aims to evaluate the value of the displacement amplification factor in seismic design codes and then tries to propose a value to estimate the maximum lateral structural displacement from sever earthquakes, without using non-linear analysis. In seismic codes, since the displacement amplification is related to “force reduction factor" hence; this aspect has been accepted in the current study. Meanwhile, two methodologies are applied to evaluate the value of displacement amplification factor and its relation with the force reduction factor. In the first methodology, which is applied for all structures, the ratio of displacement amplification and force reduction factors is determined directly. Whereas, in the second methodology that is applicable just for R/C moment resisting frame, the ratio is obtained by calculating both factors, separately. The acquired results of these methodologies are alike and estimate the ratio of two factors from 1 to 1.2. The results indicate that the ratio of the displacement amplification factor and the force reduction factor differs to those proposed by seismic provisions such as NEHRP, IBC and Iranian seismic code (standard no. 2800).

Keywords: Displacement amplification factor, Ductility factor, Force reduction factor, Maximum lateral displacement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2851
8441 Reduction in Population Growth under Various Contraceptive Strategies in Uttar Pradesh, India

Authors: Prashant Verma, K. K. Singh, Anjali Singh, Ujjaval Srivastava

Abstract:

Contraceptive policies have been derived to achieve desired reductions in the growth rate and also, applied to the data of Uttar-Pradesh, India for illustration. Using the Lotka’s integral equation for the stable population, expressions for the proportion of contraceptive users at different ages have been obtained. At the age of 20 years, 42% of contraceptive users is imperative to reduce the present annual growth rate of 0.036 to 0.02, assuming that 40% of the contraceptive users discontinue at the age of 25 years and 30% again continue contraceptive use at age 30 years. Further, presuming that 75% of women start using contraceptives at the age of 23 years, and 50% of the remaining women start using contraceptives at the age of 28 years, while the rest of them start using it at the age of 32 years. If we set a minimum age of marriage as 20 years, a reduction of 0.019 in growth rate will be obtained. This study describes how the level of contraceptive use at different age groups of women reduces the growth rate in the state of Uttar Pradesh. The article also promotes delayed marriage in the region.

Keywords: Child bearing, contraceptive devices, contraceptive policies, population growth, stable population.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 614
8440 The Role of Home Composting in Waste Management Cost Reduction

Authors: Nahid Hassanshahi, Ayoub Karimi-Jashni, Nasser Talebbeydokhti

Abstract:

Due to the economic and environmental benefits of producing less waste, the US Environmental Protection Agency (EPA) introduces source reduction as one of the most important means to deal with the problems caused by increased landfills and pollution. Waste reduction involves all waste management methods, including source reduction, recycling, and composting, which reduce waste flow to landfills or other disposal facilities. Source reduction of waste can be studied from two perspectives: avoiding waste production, or reducing per capita waste production, and waste deviation that indicates the reduction of waste transfer to landfills. The present paper has investigated home composting as a managerial solution for reduction of waste transfer to landfills. Home composting has many benefits. The use of household waste for the production of compost will result in a much smaller amount of waste being sent to landfills, which in turn will reduce the costs of waste collection, transportation and burial. Reducing the volume of waste for disposal and using them for the production of compost and plant fertilizer might help to recycle the material in a shorter time and to use them effectively in order to preserve the environment and reduce contamination. Producing compost in a home-based manner requires very small piece of land for preparation and recycling compared with other methods. The final product of home-made compost is valuable and helps to grow crops and garden plants. It is also used for modifying the soil structure and maintaining its moisture. The food that is transferred to landfills will spoil and produce leachate after a while. It will also release methane and greenhouse gases. But, composting these materials at home is the best way to manage degradable materials, use them efficiently and reduce environmental pollution. Studies have shown that the benefits of the sale of produced compost and the reduced costs of collecting, transporting, and burying waste can well be responsive to the costs of purchasing home compost machine and the cost of related trainings. Moreover, the process of producing home compost may be profitable within 4 to 5 years and as a result, it will have a major role in reducing waste management.

Keywords: Compost, home compost, reducing waste, waste management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 831
8439 Controller Design of Discrete Systems by Order Reduction Technique Employing Differential Evolution Optimization Algorithm

Authors: J. S. Yadav, N. P. Patidar, J. Singhai

Abstract:

One of the main objectives of order reduction is to design a controller of lower order which can effectively control the original high order system so that the overall system is of lower order and easy to understand. In this paper, a simple method is presented for controller design of a higher order discrete system. First the original higher order discrete system in reduced to a lower order model. Then a Proportional Integral Derivative (PID) controller is designed for lower order model. An error minimization technique is employed for both order reduction and controller design. For the error minimization purpose, Differential Evolution (DE) optimization algorithm has been employed. DE method is based on the minimization of the Integral Squared Error (ISE) between the desired response and actual response pertaining to a unit step input. Finally the designed PID controller is connected to the original higher order discrete system to get the desired specification. The validity of the proposed method is illustrated through a numerical example.

Keywords: Discrete System, Model Order Reduction, PIDController, Integral Squared Error, Differential Evolution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1859
8438 Effect of Fuel Lean Reburning Process on NOx Reduction and CO Emission

Authors: Changyeop Lee, Sewon Kim

Abstract:

Reburning is a useful technology in reducing nitric oxide through injection of a secondary hydrocarbon fuel. In this paper, an experimental study has been conducted to evaluate the effect of fuel lean reburning on NOx/CO reduction in LNG flame. Experiments were performed in flames stabilized by a co-flow swirl burner, which was mounted at the bottom of the furnace. Tests were conducted using LNG gas as the reburn fuel as well as the main fuel. The effects of reburn fuel fraction and injection manner of the reburn fuel were studied when the fuel lean reburning system was applied. The paper reports data on flue gas emissions and temperature distribution in the furnace for a wide range of experimental conditions. At steady state, temperature distribution and emission formation in the furnace have been measured and compared. This paper makes clear that in order to decrease both NOx and CO concentrations in the exhaust when the pulsated fuel lean reburning system was adapted, it is important that the control of some factors such as frequency and duty ratio. Also it shows the fuel lean reburning is also effective method to reduce NOx as much as reburning.

Keywords: Fuel lean reburn, NOx, CO, LNG flame.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2167
8437 Preparation and Characterization of Photocatalyst for the Conversion of Carbon Dioxide to Methanol

Authors: D. M. Reddy Prasad, Nur Sabrina Binti Rahmat, Huei Ruey Ong, Chin Kui Cheng, Maksudur Rahman Khan, D. Sathiyamoorthy

Abstract:

Carbon dioxide (CO2) emission to the environment is inevitable which is responsible for global warming. Photocatalytic reduction of CO2 to fuel, such as methanol, methane etc. is a promising way to reduce greenhouse gas CO2 emission. In the present work, Bi2S3/CdS was synthesized as an effective visible light responsive photocatalyst for CO2 reduction into methanol. The Bi2S3/CdS photocatalyst was prepared by hydrothermal reaction. The catalyst was characterized by X-ray diffraction (XRD) instrument. The photocatalytic activity of the catalyst has been investigated for methanol production as a function of time. Gas chromatograph flame ionization detector (GC-FID) was employed to analyze the product. The yield of methanol was found to increase with higher CdS concentration in Bi2S3/CdS and the maximum yield was obtained for 45 wt% of Bi2S3/CdS under visible light irradiation was 20 μmole/g. The result establishes that Bi2S3/CdS is favorable catalyst to reduce CO2 to methanol.

Keywords: Photocatalyst, Carbon dioxide reduction, visible light, Irradiation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1997
8436 Reduction of Leakage Power in Digital Logic Circuits Using Stacking Technique in 45 Nanometer Regime

Authors: P.K. Sharma, B. Bhargava, S. Akashe

Abstract:

Power dissipation due to leakage current in the digital circuits is a biggest factor which is considered specially while designing nanoscale circuits. This paper is exploring the ideas of reducing leakage current in static CMOS circuits by stacking the transistors in increasing numbers. Clearly it means that the stacking of OFF transistors in large numbers result a significant reduction in power dissipation. Increase in source voltage of NMOS transistor minimizes the leakage current. Thus stacking technique makes circuit with minimum power dissipation losses due to leakage current. Also some of digital circuits such as full adder, D flip flop and 6T SRAM have been simulated in this paper, with the application of reduction technique on ‘cadence virtuoso tool’ using specter at 45nm technology with supply voltage 0.7V.

Keywords: Stack, 6T SRAM cell, low power, threshold voltage

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3379
8435 Studies on Properties of Knowledge Dependency and Reduction Algorithm in Tolerance Rough Set Model

Authors: Chen Wu, Lijuan Wang

Abstract:

Relation between tolerance class and indispensable attribute and knowledge dependency in rough set model with tolerance relation is explored. After giving definitions and concepts of knowledge dependency and knowledge dependency degree for incomplete information system in tolerance rough set model by distinguishing decision attribute containing missing attribute value or not, the result of maintaining reflectivity, transitivity, augmentation, decomposition law and merge law for complete knowledge dependency is proved. Knowledge dependency degrees (not complete knowledge dependency degrees) only satisfy some laws after transitivity, augmentation and decomposition operations. An algorithm to solve attribute reduction in an incomplete decision table is designed. The correctness is checked by an example.

Keywords: Incomplete information system, rough set, tolerance relation, knowledge dependence, attribute reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 685
8434 A 3rd order 3bit Sigma-Delta Modulator with Reduced Delay Time of Data Weighted Averaging

Authors: Soon Jai Yi, Sun-Hong Kim, Hang-Geun Jeong, Seong-Ik Cho

Abstract:

This paper presents a method of reducing the feedback delay time of DWA(Data Weighted Averaging) used in sigma-delta modulators. The delay time reduction results from the elimination of the latch at the quantizer output and also from the falling edge operation. The designed sigma-delta modulator improves the timing margin about 16%. The sub-circuits of sigma-delta modulator such as SC(Switched Capacitor) integrator, 9-level quantizer, comparator, and DWA are designed with the non-ideal characteristics taken into account. The sigma-delta modulator has a maximum SNR (Signal to Noise Ratio) of 84 dB or 13 bit resolution.

Keywords: Sigma-delta modulator, multibit, DWA

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2368
8433 Particle Swarm Optimization with Reduction for Global Optimization Problems

Authors: Michiharu Maeda, Shinya Tsuda

Abstract:

This paper presents an algorithm of particle swarm optimization with reduction for global optimization problems. Particle swarm optimization is an algorithm which refers to the collective motion such as birds or fishes, and a multi-point search algorithm which finds a best solution using multiple particles. Particle swarm optimization is so flexible that it can adapt to a number of optimization problems. When an objective function has a lot of local minimums complicatedly, the particle may fall into a local minimum. For avoiding the local minimum, a number of particles are initially prepared and their positions are updated by particle swarm optimization. Particles sequentially reduce to reach a predetermined number of them grounded in evaluation value and particle swarm optimization continues until the termination condition is met. In order to show the effectiveness of the proposed algorithm, we examine the minimum by using test functions compared to existing algorithms. Furthermore the influence of best value on the initial number of particles for our algorithm is discussed.

Keywords: Particle swarm optimization, Global optimization, Metaheuristics, Reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1575
8432 Effect of Pre-Construction on Construction Schedule and Client Loyalty

Authors: Jong Hoon Kim, Hyun-Soo Lee, Moonseo Park, Min Jeong, Inbeom Lee

Abstract:

Pre-construction is essential in achieving the success of a construction project. Due to the early involvement of project participants in the construction phase, project managers are able to plan ahead and solve issues well in advance leading to the success of the project and the satisfaction of the client. This research utilizes quantitative data derived from construction management projects in order to identify the relationship between pre-construction, construction schedule, and client satisfaction. A total of 65 construction projects and 93 clients were investigated for this research in an attempt to identify (a) the relationship between pre-construction and schedule reduction, and (b) pre-construction and client loyalty. Based on the quantitative analysis, this research was able to establish a negative correlation based on 65 construction projects between pre-construction and project schedule existed. This finding represents that the more pre-construction is performed for a certain project, the overall construction schedule decreased. Then, to determine the relationship between pre-construction and client satisfaction, Net Promoter Score (NPS) of 93 clients from the 65 projects was utilized. Pre-construction and NPS was further analyzed and a positive correlation was found between the two. This infers that clients tend to be more satisfied with projects with higher ratio of pre-construction than those projects with less pre-construction.

Keywords: Client loyalty, NPS, pre-construction, schedule reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3170
8431 TOSOM: A Topic-Oriented Self-Organizing Map for Text Organization

Authors: Hsin-Chang Yang, Chung-Hong Lee, Kuo-Lung Ke

Abstract:

The self-organizing map (SOM) model is a well-known neural network model with wide spread of applications. The main characteristics of SOM are two-fold, namely dimension reduction and topology preservation. Using SOM, a high-dimensional data space will be mapped to some low-dimensional space. Meanwhile, the topological relations among data will be preserved. With such characteristics, the SOM was usually applied on data clustering and visualization tasks. However, the SOM has main disadvantage of the need to know the number and structure of neurons prior to training, which are difficult to be determined. Several schemes have been proposed to tackle such deficiency. Examples are growing/expandable SOM, hierarchical SOM, and growing hierarchical SOM. These schemes could dynamically expand the map, even generate hierarchical maps, during training. Encouraging results were reported. Basically, these schemes adapt the size and structure of the map according to the distribution of training data. That is, they are data-driven or dataoriented SOM schemes. In this work, a topic-oriented SOM scheme which is suitable for document clustering and organization will be developed. The proposed SOM will automatically adapt the number as well as the structure of the map according to identified topics. Unlike other data-oriented SOMs, our approach expands the map and generates the hierarchies both according to the topics and their characteristics of the neurons. The preliminary experiments give promising result and demonstrate the plausibility of the method.

Keywords: Self-organizing map, topic identification, learning algorithm, text clustering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1987
8430 A Study of the Adaptive Reuse for School Land Use Strategy: An Application of the Analytic Network Process and Big Data

Authors: Wann-Ming Wey

Abstract:

In today's popularity and progress of information technology, the big data set and its analysis are no longer a major conundrum. Now, we could not only use the relevant big data to analysis and emulate the possible status of urban development in the near future, but also provide more comprehensive and reasonable policy implementation basis for government units or decision-makers via the analysis and emulation results as mentioned above. In this research, we set Taipei City as the research scope, and use the relevant big data variables (e.g., population, facility utilization and related social policy ratings) and Analytic Network Process (ANP) approach to implement in-depth research and discussion for the possible reduction of land use in primary and secondary schools of Taipei City. In addition to enhance the prosperous urban activities for the urban public facility utilization, the final results of this research could help improve the efficiency of urban land use in the future. Furthermore, the assessment model and research framework established in this research also provide a good reference for schools or other public facilities land use and adaptive reuse strategies in the future.

Keywords: Adaptive reuse, analytic network process, big data, land use strategy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 886
8429 Reduction of Plutonium Production in Heavy Water Research Reactor: A Feasibility Study through Neutronic Analysis Using MCNPX2.6 and CINDER90 Codes

Authors: H. Shamoradifar, B. Teimuri, P. Parvaresh, S. Mohammadi

Abstract:

One of the main characteristics of Heavy Water Moderated Reactors is their high production of plutonium. This article demonstrates the possibility of reduction of plutonium and other actinides in Heavy Water Research Reactor. Among the many ways for reducing plutonium production in a heavy water reactor, in this research, changing the fuel from natural Uranium fuel to Thorium-Uranium mixed fuel was focused. The main fissile nucleus in Thorium-Uranium fuels is U-233 which would be produced after neutron absorption by Th-232, so the Thorium-Uranium fuels have some known advantages compared to the Uranium fuels. Due to this fact, four Thorium-Uranium fuels with different compositions ratios were chosen in our simulations; a) 10% UO2-90% THO2 (enriched= 20%); b) 15% UO2-85% THO2 (enriched= 10%); c) 30% UO2-70% THO2 (enriched= 5%); d) 35% UO2-65% THO2 (enriched= 3.7%). The natural Uranium Oxide (UO2) is considered as the reference fuel, in other words all of the calculated data are compared with the related data from Uranium fuel. Neutronic parameters were calculated and used as the comparison parameters. All calculations were performed by Monte Carol (MCNPX2.6) steady state reaction rate calculation linked to a deterministic depletion calculation (CINDER90). The obtained computational data showed that Thorium-Uranium fuels with four different fissile compositions ratios can satisfy the safety and operating requirements for Heavy Water Research Reactor. Furthermore, Thorium-Uranium fuels have a very good proliferation resistance and consume less fissile material than uranium fuels at the same reactor operation time. Using mixed Thorium-Uranium fuels reduced the long-lived α emitter, high radiotoxic wastes and the radio toxicity level of spent fuel.

Keywords: Burn-up, heavy water reactor, minor actinides, Monte Carlo, proliferation resistance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 960
8428 One-Pot Facile Synthesis of N-Doped Graphene Synthesized from Paraphenylenediamine as Metal-Free Catalysts for the Oxygen Reduction Used for Alkaline Fuel Cells

Authors: Leila Samiee, Amir Yadegari, Saeedeh Tasharrofi

Abstract:

In the work presented here, nitrogen-doped graphene materials were synthesized and used as metal-free electrocatalysts for oxygen reduction reaction (ORR) under alkaline conditions. Paraphenylenediamine was used as N precursor. The N-doped graphene was synthesized under hydrothermal treatment at 200°C. All the materials have been characterized by X-ray diffraction (XRD), Fourier transform infrared spectroscopy (FTIR), Transmission electron microscopy (TEM) and X-ray photo-electron spectroscopy (XPS). Moreover, for electrochemical evaluation of samples, Rotating Disk electrode (RDE) and Cyclic Voltammetry techniques (CV) were employed. The resulting material exhibits an outstanding catalytic activity for the oxygen reduction reaction (ORR) as well as excellent resistance towards methanol crossover effects, indicating their promising potential as ORR electrocatalysts for alkaline fuel cells.

Keywords: Alkaline fuel cell, graphene, metal-free catalyst, paraphenylenediamine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1695
8427 Numerical Investigation of the Evaporation and Mixing of UWS in a Diesel Exhaust Pipe

Authors: Tae Hyun Ahn, Gyo Woo Lee, Man Young Kim

Abstract:

Because of high thermal efficiency and low CO2 emission, diesel engines are being used widely in many industrial fields although it makes many PM and NOx which give both human health and environment a negative effect. NOx regulations for diesel engines, however, are being strengthened and it is impossible to meet the emission standard without NOx reduction devices such as SCR (Selective Catalytic Reduction), LNC (Lean NOx Catalyst), and LNT (Lean NOx Trap). Among the NOx reduction devices, urea-SCR system is known as the most stable and efficient method to solve the problem of NOx emission. But this device has some issues associated with the ammonia slip phenomenon which is occurred by shortage of evaporation and thermolysis time, and that makes it difficult to achieve uniform distribution of the injected urea in front of monolith. Therefore, this study has focused on the mixing enhancement between urea and exhaust gases to enhance the efficiency of the SCR catalyst equipped in catalytic muffler by changing inlet gas temperature and spray conditions to improve the spray uniformity of the urea water solution. Finally, it can be found that various parameters such as inlet gas temperature and injector and injection angles significantly affect the evaporation and mixing of the urea water solution with exhaust gases, and therefore, optimization of these parameters are required.

Keywords: Evaporation, Injection, Selective Catalytic Reduction (SCR), Thermolysis, UWS (Urea-Water-Solution).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2854
8426 A survey Method and new design Lecture Chair for Complied Ergonomics Guideline at Classroom Building 2 Suranaree University of Technology, Thailand

Authors: Sumalee B., Sirinapa L., Jenjira T., Jr., Setasak S.

Abstract:

The paper describes ergonomics problems trend of student at B5101 classroom building 2, Suranaree University of Technology. The objective to survey ergonomics problems and effect from use chairs for sitting in class room. The result from survey method 100 student they use lecture chair for sitting in classroom more than 2 hours/ day by RULA[1]. and Body discomfort survey[2]. The result from Body discomfort survey contribute fatigue problems at neck, lower back, upper back and right shoulder 2.93, 2.91, 2.33, 1.75 respectively and result from RULA contribute fatigue problems at neck, body and right upper arm 4.00, 3.75 and 3.00 respectively are consistent. After that the researcher provide improvement plan for design new chair support student fatigue reduction by prepare data of sample anthropometry and design ergonomics chair prototype 3 unit. Then sample 100 student trial to use new chair and evaluate again by RULA, Body discomfort and satisfaction. The result from trial new chair after improvement by RULA present fatigue reduction average of head and neck from 4.00 to 2.25 , body and trunk from 3.75 to 2.00 and arm force from 1.00 to 0.25 respectively. The result from trial new chair after improvement by Body discomfort present fatigue reduction average of lower back from 2.91 to 0.87, neck from 2.93 to 1.24, upper back 2.33 to 0.84 and right upper arm from 1.75 to 0.74. That statistical of RULA and Body discomfort survey present fatigue reduction after improvement significance with a confidence level of 95% (p-value 0.05). When analyzing the relationship of fatigue as part of the body by Chi – square test during RULA and Body discomfort that before and after improvements were consistent with the significant level of confidence 95% (p-value 0.05) . Moreover the students satisfaction result from trial with a new chair for 30 minutes [3]. 72 percent very satisfied of the folding of the secondary writing simple 66% the width of the writing plate, 64% the suitability of the writing plate, 62% of soft seat cushion and 61% easy to seat the chair.

Keywords: Ergonomics, Work station design, ErgonomicsChair, Student, Fatigue

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3468