Search results for: data reduction
8451 Experimental Technique for Vibration Reduction of a Motor Pumpin Medical Device
Authors: Young Kuen Cho, Dae Won Lee, Young-Jin Jung, Sung Kuk Kim, Dong-Hyun Seo, Chang-Yong Ko, Han Sung Kim
Abstract:
Many medical devices are driven by motor pumps. Some researchers reported that the vibration mainly affected medical devices using a motor pump. The purpose of this study was to examine the effect of stiffness and damping coefficient in a 3-dimensional (3D) model of a motor pump and spring. In the present paper, experimental and mathematical tests for the moments of inertia of the 3D model and the material properties were investigated by an INSTRON machine. The response surfaces could be generated by using 3D multi-body analysis and the design of experiment method. It showed that differences in contours of the response surface were clearly found for the particular area. Displacement of the center of the motor pump was decreased at K≈2000 N/M, C≈12.5 N-sec/M. However, the frequency was increased at K≈2000 N/M, C≈15 N-sec/M. In this study, this study suggested experimental technique for vibration reduction for a motor pump in medical device. The combined method suggested in this study will greatly contribute to design of medical devices concerning vibration and noise intervention.
Keywords: Motor pump, Spring, Vibration reduction, Medicaldevices, Moment of Inertia
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19008450 Reducing Weight and Fuel Consumption of Civil Aircraft by EML
Authors: L. Bertola, T. Cox, P. Wheeler, S. Garvey, H. Morvan
Abstract:
Electromagnetic Launch (EML) systems have been proposed for military applications to accelerate jet planes on aircraft carriers. This paper proposes the implementation of similar technology to aid civil aircraft take-off, which can provide significant economic, environmental and technical benefits. Assisted launch has the potential of reducing on ground noise and emissions near airports and improving overall aircraft efficiency through reducing engine thrust requirements. This paper presents a take-off performance analysis for an Airbus A320-200 taking off with and without the assistance of the electromagnetic catapult. Assisted take-off allows for a significant reduction in take-off field length, giving more capacity with existing airport footprints and reducing the necessary footprint of new airports, which will both reduce costs and increase the number of suitable sites. The electromagnetic catapult may allow the installation of smaller engines with lower rated thrust. The consequent fuel consumption and operational cost reduction is estimated. The potential of reducing the aircraft operational costs and the runway length required make EML system an attractive solution to the air traffic growth in busy airports.
Keywords: EML system, fuel consumption, take-off analysis, weight reduction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10888449 Highlighting of the Factors and Policies Affecting CO2 Emissions Level in Malaysian Transportation Sector
Authors: M. S. Indati, H. A. Bekhet
Abstract:
Global CO2 emission and increasing fuel consumption to meet energy demand has become a threat in recent decades. Effort to reduce the CO2 emission is now a matter of priority in most countries of the world including Malaysia. Transportation has been identified as the most intensive sector of carbon-based fuels and achievement of the voluntary target to meet 40% carbon intensity reduction set at the 15th Conference of the Parties (COP15) means that the emission from the transport sector must be reduced accordingly. This posed a great challenge to Malaysia and effort has to be made to embrace suitable and appropriate energy policy for sustainable energy and emission reduction of this sector. The focus of this paper is to analyze the trends of Malaysia’s energy consumption and emission of four different transport sub-sectors (road, rail, aviation and maritime). Underlying factors influencing the growth of energy consumption and emission trends are discussed. Besides, technology status towards energy efficiency in transportation sub-sectors is presented. By reviewing the existing policies and trends of energy used, the paper highlights prospective policy options towards achieving emission reduction in the transportation sector.
Keywords: CO2 Emission, Energy policy, Fuel consumption, Transportation sector, Malaysia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36838448 Parallel-computing Approach for FFT Implementation on Digital Signal Processor (DSP)
Authors: Yi-Pin Hsu, Shin-Yu Lin
Abstract:
An efficient parallel form in digital signal processor can improve the algorithm performance. The butterfly structure is an important role in fast Fourier transform (FFT), because its symmetry form is suitable for hardware implementation. Although it can perform a symmetric structure, the performance will be reduced under the data-dependent flow characteristic. Even though recent research which call as novel memory reference reduction methods (NMRRM) for FFT focus on reduce memory reference in twiddle factor, the data-dependent property still exists. In this paper, we propose a parallel-computing approach for FFT implementation on digital signal processor (DSP) which is based on data-independent property and still hold the property of low-memory reference. The proposed method combines final two steps in NMRRM FFT to perform a novel data-independent structure, besides it is very suitable for multi-operation-unit digital signal processor and dual-core system. We have applied the proposed method of radix-2 FFT algorithm in low memory reference on TI TMSC320C64x DSP. Experimental results show the method can reduce 33.8% clock cycles comparing with the NMRRM FFT implementation and keep the low-memory reference property.
Keywords: Parallel-computing, FFT, low-memory reference, TIDSP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21978447 Optimal Capacitor Allocation for loss reduction in Distribution System Using Fuzzy and Plant Growth Simulation Algorithm
Authors: R. Srinivasa Rao
Abstract:
This paper presents a new and efficient approach for capacitor placement in radial distribution systems that determine the optimal locations and size of capacitor with an objective of improving the voltage profile and reduction of power loss. The solution methodology has two parts: in part one the loss sensitivity factors are used to select the candidate locations for the capacitor placement and in part two a new algorithm that employs Plant growth Simulation Algorithm (PGSA) is used to estimate the optimal size of capacitors at the optimal buses determined in part one. The main advantage of the proposed method is that it does not require any external control parameters. The other advantage is that it handles the objective function and the constraints separately, avoiding the trouble to determine the barrier factors. The proposed method is applied to 9 and 34 bus radial distribution systems. The solutions obtained by the proposed method are compared with other methods. The proposed method has outperformed the other methods in terms of the quality of solution.Keywords: Distribution systems, Capacitor allocation, Loss reduction, Fuzzy, PGSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22788446 Fast Wavelet Image Denoising Based on Local Variance and Edge Analysis
Authors: Gaoyong Luo
Abstract:
The approach based on the wavelet transform has been widely used for image denoising due to its multi-resolution nature, its ability to produce high levels of noise reduction and the low level of distortion introduced. However, by removing noise, high frequency components belonging to edges are also removed, which leads to blurring the signal features. This paper proposes a new method of image noise reduction based on local variance and edge analysis. The analysis is performed by dividing an image into 32 x 32 pixel blocks, and transforming the data into wavelet domain. Fast lifting wavelet spatial-frequency decomposition and reconstruction is developed with the advantages of being computationally efficient and boundary effects minimized. The adaptive thresholding by local variance estimation and edge strength measurement can effectively reduce image noise while preserve the features of the original image corresponding to the boundaries of the objects. Experimental results demonstrate that the method performs well for images contaminated by natural and artificial noise, and is suitable to be adapted for different class of images and type of noises. The proposed algorithm provides a potential solution with parallel computation for real time or embedded system application.Keywords: Edge strength, Fast lifting wavelet, Image denoising, Local variance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20278445 Degradation in Organic Light Emitting Diodes
Authors: Saba Zare Zardareh, Farhad Akbari Boroumand
Abstract:
The objective is to fabricate organic light emitting diode and to study its degradation process in atmosphere condition in which PFO as an emitting material and PEDOT:PSS as a hole injecting material were used on ITO substrate. Thus degradation process of the OLED was studied upon its current-voltage characteristic. By fabricating this OLED and obtaining blue light and analysis of current-voltage characteristic during the time after fabrication, it was observed that the current of the OLED was exponentially decreased. Current reduction during the initial hours of fabrication was outstanding and after few days its reduction rate was dropped significantly, while the diode was dying.Keywords: OLED, Degradation, Dark spot.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22038444 Innovation in Lean Thinking to Achieve Rapid Construction
Authors: Muhamad Azani Yahya, Vikneswaran Munikanan, Mohammed Alias Yusof
Abstract:
Lean thinking holds the potential for improving the construction sector, and therefore, it is a concept that should be adopted by construction sector players and academicians in the real industry. Bridging from that, a learning process for construction sector players regarding this matter should be the agenda in gaining the knowledge in preparation for their career. Lean principles offer opportunities for reducing lead times, eliminating non-value adding activities, reducing variability, and are facilitated by methods such as pull scheduling, simplified operations and buffer reduction. Thus, the drive for rapid construction, which is a systematic approach in enhancing efficiency to deliver a project using time reduction, while lean is the continuous process of eliminating waste, meeting or exceeding all customer requirements, focusing on the entire value stream and pursuing perfection in the execution of a constructed project. The methodology presented is shown to be valid through literature, interviews and questionnaire. The results show that the majority of construction sector players unfamiliar with lean thinking and they agreed that it can improve the construction process flow. With this background knowledge established and identified, best practices and recommended action are drawn.
Keywords: Construction improvement, rapid construction, time reduction, lean construction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13058443 From Electroencephalogram to Epileptic Seizures Detection by Using Artificial Neural Networks
Authors: Gaetano Zazzaro, Angelo Martone, Roberto V. Montaquila, Luigi Pavone
Abstract:
Seizure is the main factor that affects the quality of life of epileptic patients. The diagnosis of epilepsy, and hence the identification of epileptogenic zone, is commonly made by using continuous Electroencephalogram (EEG) signal monitoring. Seizure identification on EEG signals is made manually by epileptologists and this process is usually very long and error prone. The aim of this paper is to describe an automated method able to detect seizures in EEG signals, using knowledge discovery in database process and data mining methods and algorithms, which can support physicians during the seizure detection process. Our detection method is based on Artificial Neural Network classifier, trained by applying the multilayer perceptron algorithm, and by using a software application, called Training Builder that has been developed for the massive extraction of features from EEG signals. This tool is able to cover all the data preparation steps ranging from signal processing to data analysis techniques, including the sliding window paradigm, the dimensionality reduction algorithms, information theory, and feature selection measures. The final model shows excellent performances, reaching an accuracy of over 99% during tests on data of a single patient retrieved from a publicly available EEG dataset.
Keywords: Artificial Neural Network, Data Mining, Electroencephalogram, Epilepsy, Feature Extraction, Seizure Detection, Signal Processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13128442 A Spatial Hypergraph Based Semi-Supervised Band Selection Method for Hyperspectral Imagery Semantic Interpretation
Authors: Akrem Sellami, Imed Riadh Farah
Abstract:
Hyperspectral imagery (HSI) typically provides a wealth of information captured in a wide range of the electromagnetic spectrum for each pixel in the image. Hence, a pixel in HSI is a high-dimensional vector of intensities with a large spectral range and a high spectral resolution. Therefore, the semantic interpretation is a challenging task of HSI analysis. We focused in this paper on object classification as HSI semantic interpretation. However, HSI classification still faces some issues, among which are the following: The spatial variability of spectral signatures, the high number of spectral bands, and the high cost of true sample labeling. Therefore, the high number of spectral bands and the low number of training samples pose the problem of the curse of dimensionality. In order to resolve this problem, we propose to introduce the process of dimensionality reduction trying to improve the classification of HSI. The presented approach is a semi-supervised band selection method based on spatial hypergraph embedding model to represent higher order relationships with different weights of the spatial neighbors corresponding to the centroid of pixel. This semi-supervised band selection has been developed to select useful bands for object classification. The presented approach is evaluated on AVIRIS and ROSIS HSIs and compared to other dimensionality reduction methods. The experimental results demonstrate the efficacy of our approach compared to many existing dimensionality reduction methods for HSI classification.Keywords: Hyperspectral image, spatial hypergraph, dimensionality reduction, semantic interpretation, band selection, feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12198441 Measuring the Performance of the Accident Reductions: Evidence from Izmir City
Authors: Y. Duvarci, S. Mizokami
Abstract:
Traffic enforcement units (the Police) are partly responsible for the severity and frequency of the traffic accidents via the effectiveness of their safety measures. The Police claims that the reductions in accidents and their severities occur largely by their timely interventions at the black spots, through traffic management or temporary changes in the road design (guiding, reducing speeds and eliminating sight obstructions, etc.). Yet, some other external factors than the Police measures may intervene into which such claims require a statistical confirmation. In order to test the net impact of the Police contribution in the reduction of the number of crashes, Chi square test was applied for 25 spots (streets and intersections) and an average evaluation was achieved for general conclusion in the case study of Izmir city. Separately, the net impact of economic crisis in the reduction of crashes is assessed by the trend analysis for the case of the economic crisis with the probable reduction effects on the trip generation or modal choice. Finally, it was proven that the Police measures were effective to some degree as they claimed, though the economic crisis might have only negligible contribution to the reductions in the same period observed.Keywords: Road Safety, Police, Enforcement units, Chi Squaretest, Economic Impact.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16878440 Comanche – A Compiler-Driven I/O Management System
Authors: Wendy Zhang, Ernst L. Leiss, Huilin Ye
Abstract:
Most scientific programs have large input and output data sets that require out-of-core programming or use virtual memory management (VMM). Out-of-core programming is very error-prone and tedious; as a result, it is generally avoided. However, in many instance, VMM is not an effective approach because it often results in substantial performance reduction. In contrast, compiler driven I/O management will allow a program-s data sets to be retrieved in parts, called blocks or tiles. Comanche (COmpiler MANaged caCHE) is a compiler combined with a user level runtime system that can be used to replace standard VMM for out-of-core programs. We describe Comanche and demonstrate on a number of representative problems that it substantially out-performs VMM. Significantly our system does not require any special services from the operating system and does not require modification of the operating system kernel.Keywords: I/O Management, Out-of-core, Compiler, Tile mapping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13178439 Comparative Analysis of Diverse Collection of Big Data Analytics Tools
Authors: S. Vidhya, S. Sarumathi, N. Shanthi
Abstract:
Over the past era, there have been a lot of efforts and studies are carried out in growing proficient tools for performing various tasks in big data. Recently big data have gotten a lot of publicity for their good reasons. Due to the large and complex collection of datasets it is difficult to process on traditional data processing applications. This concern turns to be further mandatory for producing various tools in big data. Moreover, the main aim of big data analytics is to utilize the advanced analytic techniques besides very huge, different datasets which contain diverse sizes from terabytes to zettabytes and diverse types such as structured or unstructured and batch or streaming. Big data is useful for data sets where their size or type is away from the capability of traditional relational databases for capturing, managing and processing the data with low-latency. Thus the out coming challenges tend to the occurrence of powerful big data tools. In this survey, a various collection of big data tools are illustrated and also compared with the salient features.
Keywords: Big data, Big data analytics, Business analytics, Data analysis, Data visualization, Data discovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37748438 A New Approach to Face Recognition Using Dual Dimension Reduction
Authors: M. Almas Anjum, M. Younus Javed, A. Basit
Abstract:
In this paper a new approach to face recognition is presented that achieves double dimension reduction, making the system computationally efficient with better recognition results and out perform common DCT technique of face recognition. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results change with change in face image resolution and provide optimal results when arriving at a certain resolution level. In the proposed model of face recognition, initially image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to increased computational speed and feature extraction potential of Discrete Cosine Transform (DCT), it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A tradeoff between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL , Yale and EME color database.Keywords: Biometrics, DCT, Face Recognition, Illumination, Computation, Feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16858437 Gypsum Composites with CDW as Raw Material
Authors: R. Santos Jiménez, A. San-Antonio-González, M. Del Río Merino, M. González Cortina, C. Viñas Arrebola
Abstract:
In this study, the feasibility of incorporating ceramic waste from bricks (perforated brick and double hollow brick) and extruded polystyrene (XPS) waste, is analysed. Results show that it is possible to incorporate up to 25% of ceramic waste and 4% of XPS waste over the weight of gypsum in a gypsum matrix. Furthermore, with the addition of ceramic waste an 8% of surface hardness increase and a 25% of capillary water absorption reduction can be obtained. On the other hand, with the addition of XPS, a 26% reduction of density and a 37% improvement of thermal conductivity can be obtained. The obtained results are favorable to use these materials in order to produce prefabricated gypsum and also as material for interior cladding walls.Keywords: CDW, waste materials, ceramic waste, XPS, construction materials, gypsum.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22138436 Robust ANOVA: An Illustrative Study in Horticultural Crop Research
Authors: Dinesh Inamadar, R. Venugopalan, K. Padmini
Abstract:
An attempt has been made in the present communication to elucidate the efficacy of robust ANOVA methods to analyse horticultural field experimental data in the presence of outliers. Results obtained fortify the use of robust ANOVA methods as there was substantiate reduction in error mean square, and hence the probability of committing Type I error, as compared to the regular approach.
Keywords: Outliers, robust ANOVA, horticulture, Cook distance, Type I error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23038435 An Hybrid Approach for Loss Reduction in Distribution Systems using Harmony Search Algorithm
Authors: R. Srinivasa Rao
Abstract:
Individually Network reconfiguration or Capacitor control perform well in minimizing power loss and improving voltage profile of the distribution system. But for heavy reactive power loads network reconfiguration and for heavy active power loads capacitor placement can not effectively reduce power loss and enhance voltage profiles in the system. In this paper, an hybrid approach that combine network reconfiguration and capacitor placement using Harmony Search Algorithm (HSA) is proposed to minimize power loss reduction and improve voltage profile. The proposed approach is tested on standard IEEE 33 and 16 bus systems. Computational results show that the proposed hybrid approach can minimize losses more efficiently than Network reconfiguration or Capacitor control. The results of proposed method are also compared with results obtained by Simulated Annealing (SA). The proposed method has outperformed in terms of the quality of solution compared to SA.Keywords: Capacitor Control, Network Reconfiguration, HarmonySearch Algorithm, Loss Reduction, Voltage Profile.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21678434 Comparison between Torsional Ultrasonic Assisted Drilling and Conventional Drilling of Bone: An in vitro Study
Authors: Nikoo Soleimani
Abstract:
Background: Reducing torque during bone drilling is one of the effective factors in reaching to an optimal drilling process. Methods: 15 bovine femurs were drilled in vitro with a drill bit with a diameter of 4 mm using two methods of torsional ultrasonic assisted drilling (T-UAD) and convent conventional drilling (CD) and the effects of changing the feed rate and rotational speed on the torque were compared in both methods. Results: There was no significant difference in the thrust force measured in both methods due to the direction of vibrations. Results showed that using T-UAD method for bone drilling at feed rates of 0.16, 0.24 and 0.32 mm/rev led for all rotational speeds to a decrease of at least 16.3% in torque compared to the CD method. Further, using T-UAD at rotational speeds of 355~1000 rpm with various feed rates resulted in a torque reduction of 16.3~50.5% compared to CD method. Conclusions: Reducing the feed rate and increasing the rotational speed, except for the rotational speed of 500 rpm and a feed rate of 0.32 mm/rev, resulted generally in torque reduction in both methods. However, T-UAD is a more effective and desirable option for bone drilling considering its significant torque reduction.
Keywords: Torsional ultrasonic assisted drilling, torque, bone drilling, rotational speed, feed rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6448433 Steel Dust as a Coating Agent for Iron Ore Pellets at Ironmaking
Authors: M. Bahgat, H. Hanafy, H. Al-Tassan
Abstract:
Cluster formation is an essential phenomenon during direct reduction processes at shaft furnaces. Decreasing the reducing temperature to avoid this problem can cause a significant drop in throughput. In order to prevent sticking of pellets, a coating material basically inactive under the reducing conditions prevailing in the shaft furnace, should be applied to cover the outer layer of the pellets. In the present work, steel dust is used as coating material for iron ore pellets to explore dust coating effectiveness and determines the best coating conditions. Steel dust coating is applied for iron ore pellets in various concentrations. Dust slurry concentrations of 5.0-30% were used to have a coated steel dust amount of 1.0-5.0 kg per ton iron ore. Coated pellets with various concentrations were reduced isothermally in weight loss technique with simulated gas mixture to the composition of reducing gases at shaft furnaces. The influences of various coating conditions on the reduction behavior and the morphology were studied. The optimum reduced samples were comparatively applied for sticking index measurement. It was found that the optimized steel dust coating condition that achieve higher reducibility with lower sticking index was 30% steel dust slurry concentration with 3.0 kg steel dust/ton ore.Keywords: Ironmaking, coating, steel dust, reduction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9378432 Analysis on Greenhouse Gas Emissions Potential by Deploying the Green Cars in Korean Road Transport Sector
Authors: Sungjun Hong, Yanghon Chung, Nyunbae Park, Sangyong Park
Abstract:
South Korea, as the 7th largest greenhouse gas emitting country in 2011, announced that the national reduction target of greenhouse gas emissions was 30% based on BAU (Business As Usual) by 2020. And the reduction rate of the transport sector is 34.3% which is the highest figure among all sectors. This paper attempts to analyze the environmental effect on deploying the green cars in Korean road transport sector. In order to calculate the greenhouse gas emissions, the LEAP model is applied in this study.Keywords: Green car, greenhouse gas, LEAP model, road transport sector.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11168431 A Study on the Least Squares Reduced Parameter Approximation of FIR Digital Filters
Authors: S. Seyedtabaii, E. Seyedtabaii
Abstract:
Rounding of coefficients is a common practice in hardware implementation of digital filters. Where some coefficients are very close to zero or one, as assumed in this paper, this rounding action also leads to some computation reduction. Furthermore, if the discarded coefficient is of high order, a reduced order filter is obtained, otherwise the order does not change but computation is reduced. In this paper, the Least Squares approximation to rounded (or discarded) coefficient FIR filter is investigated. The result also succinctly extended to general type of FIR filters.Keywords: Digital filter, filter approximation, least squares, model order reduction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15998430 Using PFA in Feature Analysis and Selection for H.264 Adaptation
Authors: Nora A. Naguib, Ahmed E. Hussein, Hesham A. Keshk, Mohamed I. El-Adawy
Abstract:
Classification of video sequences based on their contents is a vital process for adaptation techniques. It helps decide which adaptation technique best fits the resource reduction requested by the client. In this paper we used the principal feature analysis algorithm to select a reduced subset of video features. The main idea is to select only one feature from each class based on the similarities between the features within that class. Our results showed that using this feature reduction technique the source video features can be completely omitted from future classification of video sequences.
Keywords: Adaptation, feature selection, H.264, Principal Feature Analysis (PFA)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16068429 Development of Prediction Tool for Sound Absorption and Sound Insulation for Sound Proof Properties
Authors: Yoshio Kurosawa, Takao Yamaguchi
Abstract:
High frequency automotive interior noise above 500 Hz considerably affects automotive passenger comfort. To reduce this noise, sound insulation material is often laminated on body panels or interior trim panels. For a more effective noise reduction, the sound reduction properties of this laminated structure need to be estimated. We have developed a new calculate tool that can roughly calculate the sound absorption and insulation properties of laminate structure and handy for designers. In this report, the outline of this tool and an analysis example applied to floor mat are introduced.Keywords: Automobile, acoustics, porous material, Transfer Matrix Method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18778428 A 16Kb 10T-SRAM with 4x Read-Power Reduction
Authors: Pardeep Singh, Sanjay Sharma, Parvinder S. Sandhu
Abstract:
This work aims to reduce the read power consumption as well as to enhance the stability of the SRAM cell during the read operation. A new 10-transisor cell is proposed with a new read scheme to minimize the power consumption within the memory core. It has separate read and write ports, thus cell read stability is significantly improved. A 16Kb SRAM macro operating at 1V supply voltage is demonstrated in 65 nm CMOS process. Its read power consumption is reduced to 24% of the conventional design. The new cell also has lower leakage current due to its special bit-line pre-charge scheme. As a result, it is suitable for low-power mobile applications where power supply is restricted by the battery.Keywords: A 16Kb 10T-SRAM, 4x Read-Power Reduction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19458427 Assessing Storage of Stability and Mercury Reduction of Freeze-Dried Pseudomonas putida within Different Types of Lyoprotectant
Authors: A. A. M. Azoddein, Y. Nuratri, A. B. Bustary, F. A. M. Azli, S. C. Sayuti
Abstract:
Pseudomonas putida is a potential strain in biological treatment to remove mercury contained in the effluent of petrochemical industry due to its mercury reductase enzyme that able to reduce ionic mercury to elementary mercury. Freeze-dried P. putida allows easy, inexpensive shipping, handling and high stability of the product. This study was aimed to freeze dry P. putida cells with addition of lyoprotectant. Lyoprotectant was added into the cells suspension prior to freezing. Dried P. putida obtained was then mixed with synthetic mercury. Viability of recovery P. putida after freeze dry was significantly influenced by the type of lyoprotectant. Among the lyoprotectants, tween 80/ sucrose was found to be the best lyoprotectant. Sucrose able to recover more than 78% (6.2E+09 CFU/ml) of the original cells (7.90E+09CFU/ml) after freeze dry and able to retain 5.40E+05 viable cells after 4 weeks storage in 4oC without vacuum. Polyethylene glycol (PEG) pre-treated freeze dry cells and broth pre-treated freeze dry cells after freeze-dry recovered more than 64% (5.0 E+09CFU/ml) and >0.1% (5.60E+07CFU/ml). Freeze-dried P. putida cells in PEG and broth cannot survive after 4 weeks storage. Freeze dry also does not really change the pattern of growth P. putida but extension of lag time was found 1 hour after 3 weeks of storage. Additional time was required for freeze-dried P. putida cells to recover before introduce freeze-dried cells to more complicated condition such as mercury solution. The maximum mercury reduction of PEG pre-treated freeze-dried cells after freeze dry and after storage 3 weeks was 56.78% and 17.91%. The maximum of mercury reduction of tween 80/sucrose pre-treated freeze-dried cells after freeze dry and after storage 3 weeks were 26.35% and 25.03%. Freeze dried P. putida was found to have lower mercury reduction compare to the fresh P. putida that has been growth in agar. Result from this study may be beneficial and useful as initial reference before commercialize freeze-dried P. putida.
Keywords: Pseudomonas putida, freeze-dry, PEG, Tween80/Sucrose, mercury, cell viability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11178426 Multi-labeled Data Expressed by a Set of Labels
Authors: Tetsuya Furukawa, Masahiro Kuzunishi
Abstract:
Collected data must be organized to be utilized efficiently, and hierarchical classification of data is efficient approach to organize data. When data is classified to multiple categories or annotated with a set of labels, users request multi-labeled data by giving a set of labels. There are several interpretations of the data expressed by a set of labels. This paper discusses which data is expressed by a set of labels by introducing orders for sets of labels and shows that there are four types of orders, which are characterized by whether the labels of expressed data includes every label of the given set of labels within the range of the set. Desirable properties of the orders, data is also expressed by the higher set of labels and different sets of labels express different data, are discussed for the orders.
Keywords: Classification Hierarchies, Multi-labeled Data, Multiple Classificaiton, Orders of Sets of Labels
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13038425 Comparison of Different Data Acquisition Techniques for Shape Optimization Problems
Authors: Attila Vámosi, Tamás Mankovits, Dávid Huri, Imre Kocsis, Tamás Szabó
Abstract:
Non-linear FEM calculations are indispensable when important technical information like operating performance of a rubber component is desired. For example rubber bumpers built into air-spring structures may undergo large deformations under load, which in itself shows non-linear behavior. The changing contact range between the parts and the incompressibility of the rubber increases this non-linear behavior further. The material characterization of an elastomeric component is also a demanding engineering task. The shape optimization problem of rubber parts led to the study of FEM based calculation processes. This type of problems was posed and investigated by several authors. In this paper the time demand of certain calculation methods are studied and the possibilities of time reduction is presented.
Keywords: Rubber bumper, data acquisition, finite element analysis, support vector regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21478424 EMD-Based Signal Noise Reduction
Authors: A.O. Boudraa, J.C. Cexus, Z. Saidi
Abstract:
This paper introduces a new signal denoising based on the Empirical mode decomposition (EMD) framework. The method is a fully data driven approach. Noisy signal is decomposed adaptively into oscillatory components called Intrinsic mode functions (IMFs) by means of a process called sifting. The EMD denoising involves filtering or thresholding each IMF and reconstructs the estimated signal using the processed IMFs. The EMD can be combined with a filtering approach or with nonlinear transformation. In this work the Savitzky-Golay filter and shoftthresholding are investigated. For thresholding, IMF samples are shrinked or scaled below a threshold value. The standard deviation of the noise is estimated for every IMF. The threshold is derived for the Gaussian white noise. The method is tested on simulated and real data and compared with averaging, median and wavelet approaches.
Keywords: Empirical mode decomposition, Signal denoisingnonstationary process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40018423 Benchmarking Cleaner Production Performance of Coal-fired Power Plants Using Two-stage Super-efficiency Data Envelopment Analysis
Authors: Shao-lun Zeng, Yu-long Ren
Abstract:
Benchmarking cleaner production performance is an effective way of pollution control and emission reduction in coal-fired power industry. A benchmarking method using two-stage super-efficiency data envelopment analysis for coal-fired power plants is proposed – firstly, to improve the cleaner production performance of DEA-inefficient or weakly DEA-efficient plants, then to select the benchmark from performance-improved power plants. An empirical study is carried out with the survey data of 24 coal-fired power plants. The result shows that in the first stage the performance of 16 plants is DEA-efficient and that of 8 plants is relatively inefficient. The target values for improving DEA-inefficient plants are acquired by projection analysis. The efficient performance of 24 power plants and the benchmarking plant is achieved in the second stage. The two-stage benchmarking method is practical to select the optimal benchmark in the cleaner production of coal-fired power industry and will continuously improve plants- cleaner production performance.Keywords: benchmarking, cleaner production performance, coal-fired power plant, super-efficiency data envelopment analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24308422 An Intelligent Approach of Rough Set in Knowledge Discovery Databases
Authors: Hrudaya Ku. Tripathy, B. K. Tripathy, Pradip K. Das
Abstract:
Knowledge Discovery in Databases (KDD) has evolved into an important and active area of research because of theoretical challenges and practical applications associated with the problem of discovering (or extracting) interesting and previously unknown knowledge from very large real-world databases. Rough Set Theory (RST) is a mathematical formalism for representing uncertainty that can be considered an extension of the classical set theory. It has been used in many different research areas, including those related to inductive machine learning and reduction of knowledge in knowledge-based systems. One important concept related to RST is that of a rough relation. In this paper we presented the current status of research on applying rough set theory to KDD, which will be helpful for handle the characteristics of real-world databases. The main aim is to show how rough set and rough set analysis can be effectively used to extract knowledge from large databases.Keywords: Data mining, Data tables, Knowledge discovery in database (KDD), Rough sets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2335