Search results for: PWM switching technique
2607 Topographical Image Transference Compatibility Generated Through Moiré Technique Applying Parametrical Softwares of Computer Assisted Design
Authors: M. V. G. Silva, J. Gazzola, I. M. Dal Fabbro, A. C. L. Lino
Abstract:
Computer aided design accounts with the support of parametric software in the design of machine components as well as of any other pieces of interest. The complexities of the element under study sometimes offer certain difficulties to computer design, or ever might generate mistakes in the final body conception. Reverse engineering techniques are based on the transformation of already conceived body images into a matrix of points which can be visualized by the design software. The literature exhibits several techniques to obtain machine components dimensional fields, as contact instrument (MMC), calipers and optical methods as laser scanner, holograms as well as moiré methods. The objective of this research work was to analyze the moiré technique as instrument of reverse engineering, applied to bodies of nom complex geometry as simple solid figures, creating matrices of points. These matrices were forwarded to a parametric software named SolidWorks to generate the virtual object. Volume data obtained by mechanical means, i.e., by caliper, the volume obtained through the moiré method and the volume generated by the SolidWorks software were compared and found to be in close agreement. This research work suggests the application of phase shifting moiré methods as instrument of reverse engineering, serving also to support farm machinery element designs.Keywords: Reverse engineering, Moiré technique, three dimensional image generation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34582606 Modern Vibration Signal Processing Techniques for Vehicle Gearbox Fault Diagnosis
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
This paper presents modern vibration signalprocessing techniques for vehicle gearbox fault diagnosis, via the wavelet analysis and the Squared Envelope (SE) technique. The wavelet analysis is regarded as a powerful tool for the detection of sudden changes in non-stationary signals. The Squared Envelope (SE) technique has been extensively used for rolling bearing diagnostics. In the present work a scheme of using the Squared Envelope technique for early detection of gear tooth pit. The pitting defect is manufactured on the tooth side of a fifth speed gear on the intermediate shaft of a vehicle gearbox. The objective is to supplement the current techniques of gearbox fault diagnosis based on using the raw vibration and ordered signals. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers introduce the load on the flanges of output joint shafts. The gearbox used for experimental measurements is the type most commonly used in modern small to mid-sized passenger cars with transversely mounted powertrain and front wheel drive; a five-speed gearbox with final drive gear and front wheel differential. The results show that the approaches methods are effective for detecting and diagnosing localized gear faults in early stage under different operation conditions, and are more sensitive and robust than current gear diagnostic techniques.
Keywords: Wavelet analysis, Squared Envelope, gear faults.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25762605 Loop Back Connected Component Labeling Algorithm and Its Implementation in Detecting Face
Authors: A. Rakhmadi, M. S. M. Rahim, A. Bade, H. Haron, I. M. Amin
Abstract:
In this study, a Loop Back Algorithm for component connected labeling for detecting objects in a digital image is presented. The approach is using loop back connected component labeling algorithm that helps the system to distinguish the object detected according to their label. Deferent than whole window scanning technique, this technique reduces the searching time for locating the object by focusing on the suspected object based on certain features defined. In this study, the approach was also implemented for a face detection system. Face detection system is becoming interesting research since there are many devices or systems that require detecting the face for certain purposes. The input can be from still image or videos, therefore the sub process of this system has to be simple, efficient and accurate to give a good result.Keywords: Image processing, connected components labeling, face detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23002604 Computer Generated Hologram for SemiFragile Watermarking with Encrypted Images
Authors: G. Schirripa Spagnolo, M. De Santis
Abstract:
The protection of the contents of digital products is referred to as content authentication. In some applications, to be able to authenticate a digital product could be extremely essential. For example, if a digital product is used as a piece of evidence in the court, its integrity could mean life or death of the accused. Generally, the problem of content authentication can be solved using semifragile digital watermarking techniques. Recently many authors have proposed Computer Generated Hologram Watermarking (CGHWatermarking) techniques. Starting from these studies, in this paper a semi-fragile Computer Generated Hologram coding technique is proposed, which is able to detect malicious tampering while tolerating some incidental distortions. The proposed technique uses as watermark an encrypted image, and it is well suitable for digital image authentication.Keywords: Asymmetric cryptography, Semi-Fragile watermarking, Image authentication, Hologram watermark, Public- Key Cryptography, RSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16092603 Identification of Lean Implementation Hurdles in Indian Industries
Authors: Bhim Singh
Abstract:
Due to increased pressure from global competitors, manufacturing organizations are switching over to lean philosophies from traditional mass production. Lean manufacturing is a manufacturing philosophy which focuses on elimination of various types of wastes and creates maximum value for the end customers. Lean thinking aims to produce high quality products and services at the lowest possible cost with maximum customer responsiveness. Indian Industry is facing lot of problems in this transformation from traditional mass production to lean production. Through this paper an attempt has been made to identify various lean implementation hurdles in Indian industries with the help of a structured survey. Identified hurdles are grouped with the help of factor analysis and rated by calculating descriptive statistics. To show the effect of lean implementation hurdles a hypothesis “Organizations having higher level of lean implementation hurdles will have poor (negative) performance” has been postulated and tested using correlation matrix between performance parameters of the organizations and identified hurdles. The findings of the paper will be helpful to prepare road map to identify and eradicate the lean implementation hurdles.
Keywords: Factor analysis, global competition, lean implementation and lean hurdles.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20072602 Characterization of Fabricated A 384.1-MgO Based Metal Matrix Composite and Optimization of Tensile Strength using Taguchi Techniques
Authors: Nripjit, Anand K Tyagi, Nirmal Singh
Abstract:
The present work consecutively on synthesis and characterization of composites, Al/Al alloy A 384.1 as matrix in which the main ingredient as Al/Al-5% MgO alloy based metal matrix composite. As practical implications the low cost processing route for the fabrication of Al alloy A 384.1 and operational difficulties of presently available manufacturing processes based in liquid manipulation methods. As all new developments, complete understanding of the influence of processing variables upon the final quality of the product. And the composite is applied comprehensively to the acquaintance for achieving superiority of information concerning the specific heat measurement of a material through the aid of thermographs. Products are evaluated concerning relative particle size and mechanical behavior under tensile strength. Furthermore, Taguchi technique was employed to examine the experimental optimum results are achieved, owing to effectiveness of this approach.Keywords: MMC, Thermographs, Tensile strength, Taguchi technique, Optimal parameters
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16392601 Guided Wave Sensitivity for De-Bond Defects in Aluminum Skin-Honeycomb Core
Authors: A. Satour, F. Boubenider, R. Halimi, A. Badidibouda
Abstract:
Sandwich plates are finding an increasing range of application in the aircraft industry. The inspection of honeycomb composite structure by conventional ultrasonic technique is complex and very time consuming. The present study demonstrates a technique using guided Lamb waves at low frequencies to predict de-bond defects in aluminum skin-honeycomb core sandwich structure used in aeronautics. The numerical method was investigated for drawing the dispersion and displacement curves of ultrasonic Lamb wave propagated in Aluminum plate. An experimental study was carried out to check the theoretical prediction. The detection of unsticking between the skin and the core was tested by the two first modes for a low frequency. It was found that A0 mode is more sensitive to delamination defect compared to S0 mode.
Keywords: Damage detection, delamination, guided waves, Sandwich structure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23242600 Price Quoting Method for Contract Manufacturer
Authors: S. Homrossukon, W. Parinyasart
Abstract:
This is an applied research to propose the method for price quotation for a contract electronics manufacturer. It has had a precise price quoting method but such method could not quickly provide a result as the customer required. This reduces the ability of company to compete in this kind of business. In this case, the cause of long time quotation process was analyzed. A lot of product features have been demanded by customer. By checking routine processes, it was found that high fraction of quoting time was used for production time estimating which has effected to the manufacturing or production cost. Then the historical data of products including types, number of components, assembling method, and their assembling time were used to analyze the key components affecting to production time. The price quoting model then was proposed. The implementation of proposed model was able to remarkably reduce quoting time with an acceptable required precision.Keywords: Price quoting, Contract manufacturer, Stepwise technique, Best subset technique.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 44312599 Generating Class-Based Test Cases for Interface Classes of Object-Oriented Gray-Box Frameworks
Authors: Jehad Al Dallal, Paul Sorenson
Abstract:
An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define Framework Interface Classes (FICs) and their possible specifications, which helps in building reusable test cases for the implementations of these classes. In applications developed using gray-box frameworks, FICs inherit framework classes or use them without inheritance. In this paper, a test-case generation technique is extended to build test cases for FICs built for gray-box frameworks. A tool is developed to automate the introduced technique.Keywords: Class testing, object-oriented framework, reusable test case.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15172598 Concept of Transforaminal Lumbar Interbody Fusion Cage Insertion Device
Authors: Sangram A. Sathe, Neha A. Madgulkar, Shruti S. Raut, S. P. Wadkar
Abstract:
Transforaminal lumbar interbody fusion (TLIF) surgeries have nowadays became popular for treatment of degenerated spinal disorders. The interbody fusion technique like TLIF maintains load bearing capacity of the spine and a suitable disc height. Currently many techniques have been introduced to cure Spondylolisthesis. This surgery provides greater rehabilitation of degenerative spines. While performing this TLIF surgery existing methods use guideway, which is a troublesome surgery technique as the use of two separate instruments is required to perform this surgery. This paper presents a concept which eliminates the use of guideway. This concept also eliminates problems that occur like reverting the cage. The concept discussed in this paper also gives high accuracy while performing surgery.Keywords: Degenerative disc diseases, pedicle screw, spine, spondylolisthesis, transforaminal lumbar interbody fusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13792597 The Optimal Indirect Vector Controller Design via an Adaptive Tabu Search Algorithm
Authors: P. Sawatnatee, S. Udomsuk, K-N. Areerak, K-L. Areerak, A. Srikaew
Abstract:
The paper presents how to design the indirect vector control of three-phase induction motor drive systems using the artificial intelligence technique called the adaptive tabu search. The results from the simulation and the experiment show that the drive system with the controller designed from the proposed method can provide the best output speed response compared with those of the conventional method. The controller design using the proposed technique can be used to create the software package for engineers to achieve the optimal controller design of the induction motor speed control based on the indirect vector concept.
Keywords: Indirect Vector Control, Induction Motor, Adaptive Tabu Search, Control Design, Artificial Intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19332596 Robust Statistics Based Algorithm to Remove Salt and Pepper Noise in Images
Authors: V.R.Vijaykumar, P.T.Vanathi, P.Kanagasabapathy, D.Ebenezer
Abstract:
In this paper, a robust statistics based filter to remove salt and pepper noise in digital images is presented. The function of the algorithm is to detect the corrupted pixels first since the impulse noise only affect certain pixels in the image and the remaining pixels are uncorrupted. The corrupted pixels are replaced by an estimated value using the proposed robust statistics based filter. The proposed method perform well in removing low to medium density impulse noise with detail preservation upto a noise density of 70% compared to standard median filter, weighted median filter, recursive weighted median filter, progressive switching median filter, signal dependent rank ordered mean filter, adaptive median filter and recently proposed decision based algorithm. The visual and quantitative results show the proposed algorithm outperforms in restoring the original image with superior preservation of edges and better suppression of impulse noise
Keywords: Image denoising, Nonlinear filter, Robust Statistics, and Salt and Pepper Noise.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22022595 Solving Directional Overcurrent Relay Coordination Problem Using Artificial Bees Colony
Authors: M. H. Hussain, I. Musirin, A. F. Abidin, S. R. A. Rahim
Abstract:
This paper presents the implementation of Artificial Bees Colony (ABC) algorithm in solving Directional OverCurrent Relays (DOCRs) coordination problem for near-end faults occurring in fixed network topology. The coordination optimization of DOCRs is formulated as linear programming (LP) problem. The objective function is introduced to minimize the operating time of the associated relay which depends on the time multiplier setting. The proposed technique is to taken as a technique for comparison purpose in order to highlight its superiority. The proposed algorithms have been tested successfully on 8 bus test system. The simulation results demonstrated that the ABC algorithm which has been proved to have good search ability is capable in dealing with constraint optimization problems.
Keywords: Artificial bees colony, directional overcurrent relay coordination problem, relay settings, time multiplier setting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35322594 Bi-Criteria Latency Optimization of Intra-and Inter-Autonomous System Traffic Engineering
Authors: K. Vidya, V.Rhymend Uthariaraj
Abstract:
Traffic Engineering (TE) is the process of controlling how traffic flows through a network in order to facilitate efficient and reliable network operations while simultaneously optimizing network resource utilization and traffic performance. TE improves the management of data traffic within a network and provides the better utilization of network resources. Many research works considers intra and inter Traffic Engineering separately. But in reality one influences the other. Hence the effective network performances of both inter and intra Autonomous Systems (AS) are not optimized properly. To achieve a better Joint Optimization of both Intra and Inter AS TE, we propose a joint Optimization technique by considering intra-AS features during inter – AS TE and vice versa. This work considers the important criterion say latency within an AS and between ASes. and proposes a Bi-Criteria Latency optimization model. Hence an overall network performance can be improved by considering this jointoptimization technique in terms of Latency.Keywords: Inter-Domain Routing , Measurement, OptimizationPerformance, Traffic Engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15162593 Kurtosis, Renyi's Entropy and Independent Component Scalp Maps for the Automatic Artifact Rejection from EEG Data
Authors: Antonino Greco, Nadia Mammone, Francesco Carlo Morabito, Mario Versaci
Abstract:
The goal of this work is to improve the efficiency and the reliability of the automatic artifact rejection, in particular from the Electroencephalographic (EEG) recordings. Artifact rejection is a key topic in signal processing. The artifacts are unwelcome signals that may occur during the signal acquisition and that may alter the analysis of the signals themselves. A technique for the automatic artifact rejection, based on the Independent Component Analysis (ICA) for the artifact extraction and on some high order statistics such as kurtosis and Shannon-s entropy, was proposed some years ago in literature. In this paper we enhance this technique introducing the Renyi-s entropy. The performance of our method was tested exploiting the Independent Component scalp maps and it was compared to the performance of the method in literature and it showed to outperform it.
Keywords: Artifact, EEG, Renyi's entropy, independent component analysis, kurtosis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24312592 Optimal Dynamic Economic Load Dispatch Using Artificial Immune System
Authors: I. A. Farhat
Abstract:
The The dynamic economic dispatch (DED) problem is one of the complex constrained optimization problems that have nonlinear, con-convex and non-smooth objective functions. The purpose of the DED is to determine the optimal economic operation of the committed units while meeting the load demand. Associated to this constrained problem there exist highly nonlinear and non-convex practical constraints to be satisfied. Therefore, classical and derivative-based methods are likely not to converge to an optimal or near optimal solution to such a dynamic and large-scale problem. In this paper, an Artificial Immune System technique (AIS) is implemented and applied to solve the DED problem considering the transmission power losses and the valve-point effects in addition to the other operational constraints. To demonstrate the effectiveness of the proposed technique, two case studies are considered. The results obtained using the AIS are compared to those obtained by other methods reported in the literature and found better.
Keywords: Artificial Immune System (AIS), Dynamic Economic Dispatch (DED).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18842591 Numerical Simulation of Progressive Collapse for a Reinforced Concrete Building
Authors: Han-Soo Kim, Jae-Gyun Ahn, Hyo-Seung Ahn
Abstract:
Though nonlinear dynamic analysis using a specialized hydro-code such as AUTODYN is accurate and useful tool for progressive collapse assessment of a multi-story building subjected to blast load, it takes too much time to be applied to a practical simulation of progressive collapse of a tall building. In this paper, blast analysis of a RC frame structure using a simplified model with Reinforcement Contact technique provided in Ansys Workbench was introduced and investigated on its accuracy. Even though the simplified model has a fraction of elements of the detailed model, the simplified model with this modeling technique shows similar structural behavior under the blast load to the detailed model. The proposed modeling method can be effectively applied to blast loading progressive collapse analysis of a RC frame structure.Keywords: Autodyn, Blast Load, Progressive Collapse, Reinforcement Contact.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42662590 Efficient Spectral Analysis of Quasi Stationary Time Series
Authors: Khalid M. Aamir, Mohammad A. Maud
Abstract:
Power Spectral Density (PSD) of quasi-stationary processes can be efficiently estimated using the short time Fourier series (STFT). In this paper, an algorithm has been proposed that computes the PSD of quasi-stationary process efficiently using offline autoregressive model order estimation algorithm, recursive parameter estimation technique and modified sliding window discrete Fourier Transform algorithm. The main difference in this algorithm and STFT is that the sliding window (SW) and window for spectral estimation (WSA) are separately defined. WSA is updated and its PSD is computed only when change in statistics is detected in the SW. The computational complexity of the proposed algorithm is found to be lesser than that for standard STFT technique.
Keywords: Power Spectral Density (PSD), quasi-stationarytime series, short time Fourier Transform, Sliding window DFT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19692589 Rheological and Computational Analysis of Crude Oil Transportation
Authors: Praveen Kumar, Satish Kumar, Jashanpreet Singh
Abstract:
Transportation of unrefined crude oil from the production unit to a refinery or large storage area by a pipeline is difficult due to the different properties of crude in various areas. Thus, the design of a crude oil pipeline is a very complex and time consuming process, when considering all the various parameters. There were three very important parameters that play a significant role in the transportation and processing pipeline design; these are: viscosity profile, temperature profile and the velocity profile of waxy crude oil through the crude oil pipeline. Knowledge of the Rheological computational technique is required for better understanding the flow behavior and predicting the flow profile in a crude oil pipeline. From these profile parameters, the material and the emulsion that is best suited for crude oil transportation can be predicted. Rheological computational fluid dynamic technique is a fast method used for designing flow profile in a crude oil pipeline with the help of computational fluid dynamics and rheological modeling. With this technique, the effect of fluid properties including shear rate range with temperature variation, degree of viscosity, elastic modulus and viscous modulus was evaluated under different conditions in a transport pipeline. In this paper, two crude oil samples was used, as well as a prepared emulsion with natural and synthetic additives, at different concentrations ranging from 1,000 ppm to 3,000 ppm. The rheological properties was then evaluated at a temperature range of 25 to 60 °C and which additive was best suited for transportation of crude oil is determined. Commercial computational fluid dynamics (CFD) has been used to generate the flow, velocity and viscosity profile of the emulsions for flow behavior analysis in crude oil transportation pipeline. This rheological CFD design can be further applied in developing designs of pipeline in the future.
Keywords: Natural surfactant, crude oil, rheology, CFD, viscosity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16762588 Improved Artificial Bee Colony Algorithm for Non-Convex Economic Power Dispatch Problem
Authors: Badr M. Alshammari, T. Guesmi
Abstract:
This study presents a modified version of the artificial bee colony (ABC) algorithm by including a local search technique for solving the non-convex economic power dispatch problem. The local search step is incorporated at the end of each iteration. Total system losses, valve-point loading effects and prohibited operating zones have been incorporated in the problem formulation. Thus, the problem becomes highly nonlinear and with discontinuous objective function. The proposed technique is validated using an IEEE benchmark system with ten thermal units. Simulation results demonstrate that the proposed optimization algorithm has better convergence characteristics in comparison with the original ABC algorithm.
Keywords: Economic power dispatch, artificial bee colony, valve-point loading effects, prohibited operating zones.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7572587 Impact of Weather Conditions on Generalized Frequency Division Multiplexing over Gamma Gamma Channel
Authors: Muhammad Sameer Ahmed, Piotr Remlein, Tansal Gucluoglu
Abstract:
The technique called as Generalized frequency division multiplexing (GFDM) used in the free space optical channel can be a good option for implementation free space optical communication systems. This technique has several strengths e.g. good spectral efficiency, low peak-to-average power ratio (PAPR), adaptability and low co-channel interference. In this paper, the impact of weather conditions such as haze, rain and fog on GFDM over the gamma-gamma channel model is discussed. A Trade off between link distance and system performance under intense weather conditions is also analysed. The symbol error probability (SEP) of GFDM over the gamma-gamma turbulence channel is derived and verified with the computer simulations.
Keywords: Free space optics, generalized frequency division multiplexing, weather conditions, gamma gamma distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6812586 Determining of Threshold Levels of Burst by Burst AQAM/CDMA in Slow Rayleigh Fading Environments
Authors: F. Nejadebrahimi, M. ArdebiliPour
Abstract:
In this paper, we are going to determine the threshold levels of adaptive modulation in a burst by burst CDMA system by a suboptimum method so that the above method attempts to increase the average bit per symbol (BPS) rate of transceiver system by switching between the different modulation modes in variable channel condition. In this method, we choose the minimum values of average bit error rate (BER) and maximum values of average BPS on different values of average channel signal to noise ratio (SNR) and then calculate the relative threshold levels of them, so that when the instantaneous SNR increases, a higher order modulation be employed for increasing throughput and vise-versa when the instantaneous SNR decreases, a lower order modulation be employed for improvement of BER. In transmission step, by this adaptive modulation method, in according to comparison between obtained estimation of pilot symbols and a set of above suboptimum threshold levels, above system chooses one of states no transmission, BPSK, 4QAM and square 16QAM for modulation of data. The expected channel in this paper is a slow Rayleigh fading.
Keywords: AQAM, burst, BER, BPS, CDMA, threshold.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15332585 An Overview of Corroded Pipe Repair Techniques Using Composite Materials
Authors: K. S. Lim, S. N. A. Azraai, N. M. Noor, N. Yahaya
Abstract:
Polymeric composites are being increasingly used as repair material for repairing critical infrastructures such as building, bridge, pressure vessel, piping and pipeline. Technique in repairing damaged pipes is one of the major concerns of pipeline owners. Considerable researches have been carried out on the repair of corroded pipes using composite materials. This article attempts a short review of the subject matter to provide insight into various techniques used in repairing corroded pipes, focusing on a wide range of composite repair systems. These systems including pre-cured layered, flexible wet lay-up, pre-impregnated, split composite sleeve and flexible tape systems. Both advantages and limitations of these repair systems were highlighted. Critical technical aspects have been discussed through the current standards and practices. Research gaps and future study scopes in achieving more effective design philosophy are also presented.Keywords: Composite materials, pipeline, repair technique, polymers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 55432584 Computing the Loop Bound in Iterative Data Flow Graphs Using Natural Token Flow
Authors: Ali Shatnawi
Abstract:
Signal processing applications which are iterative in nature are best represented by data flow graphs (DFG). In these applications, the maximum sampling frequency is dependent on the topology of the DFG, the cyclic dependencies in particular. The determination of the iteration bound, which is the reciprocal of the maximum sampling frequency, is critical in the process of hardware implementation of signal processing applications. In this paper, a novel technique to compute the iteration bound is proposed. This technique is different from all previously proposed techniques, in the sense that it is based on the natural flow of tokens into the DFG rather than the topology of the graph. The proposed algorithm has lower run-time complexity than all known algorithms. The performance of the proposed algorithm is illustrated through analytical analysis of the time complexity, as well as through simulation of some benchmark problems.Keywords: Data flow graph, Iteration period bound, Rateoptimalscheduling, Recursive DSP algorithms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25622583 Automatic LV Segmentation with K-means Clustering and Graph Searching on Cardiac MRI
Authors: Hae-Yeoun Lee
Abstract:
Quantification of cardiac function is performed by calculating blood volume and ejection fraction in routine clinical practice. However, these works have been performed by manual contouring, which requires computational costs and varies on the observer. In this paper, an automatic left ventricle segmentation algorithm on cardiac magnetic resonance images (MRI) is presented. Using knowledge on cardiac MRI, a K-mean clustering technique is applied to segment blood region on a coil-sensitivity corrected image. Then, a graph searching technique is used to correct segmentation errors from coil distortion and noises. Finally, blood volume and ejection fraction are calculated. Using cardiac MRI from 15 subjects, the presented algorithm is tested and compared with manual contouring by experts to show outstanding performance.
Keywords: Cardiac MRI, Graph searching, Left ventricle segmentation, K-means clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20942582 An Advanced Stereo Vision Based Obstacle Detection with a Robust Shadow Removal Technique
Authors: Saeid Fazli, Hajar Mohammadi D., Payman Moallem
Abstract:
This paper presents a robust method to detect obstacles in stereo images using shadow removal technique and color information. Stereo vision based obstacle detection is an algorithm that aims to detect and compute obstacle depth using stereo matching and disparity map. The proposed advanced method is divided into three phases, the first phase is detecting obstacles and removing shadows, the second one is matching and the last phase is depth computing. We propose a robust method for detecting obstacles in stereo images using a shadow removal technique based on color information in HIS space, at the first phase. In this paper we use Normalized Cross Correlation (NCC) function matching with a 5 × 5 window and prepare an empty matching table τ and start growing disparity components by drawing a seed s from S which is computed using canny edge detector, and adding it to τ. In this way we achieve higher performance than the previous works [2,17]. A fast stereo matching algorithm is proposed that visits only a small fraction of disparity space in order to find a semi-dense disparity map. It works by growing from a small set of correspondence seeds. The obstacle identified in phase one which appears in the disparity map of phase two enters to the third phase of depth computing. Finally, experimental results are presented to show the effectiveness of the proposed method.
Keywords: obstacle detection, stereo vision, shadowremoval, color, stereo matching
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20732581 Optimal Dynamic Economic Load Dispatch Using Artificial Immune System
Authors: I. A. Farhat
Abstract:
The dynamic economic dispatch (DED) problem is one of the complex constrained optimization problems that have nonlinear, con-convex and non-smooth objective functions. The purpose of the DED is to determine the optimal economic operation of the committed units while meeting the load demand. Associated to this constrained problem there exist highly nonlinear and non-convex practical constraints to be satisfied. Therefore, classical and derivative-based methods are likely not to converge to an optimal or near optimal solution to such a dynamic and large-scale problem. In this paper, an Artificial Immune System technique (AIS) is implemented and applied to solve the DED problem considering the transmission power losses and the valve-point effects in addition to the other operational constraints. To demonstrate the effectiveness of the proposed technique, two case studies are considered. The results obtained using the AIS are compared to those obtained by other methods reported in the literature and found better.
Keywords: Artificial Immune System (AIS), Dynamic Economic Dispatch (DED).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19912580 Level Shifted Carrier Signal Based Scalar Random Pulse Width Modulation Algorithms for Cascaded Multilevel Inverter Fed Induction Motor Drive
Authors: M. Nayeemuddin, T. Bramhananda Reddy, M. Vijaya Kumar
Abstract:
Acoustic noise becoming ever more obnoxious radiated by voltage source inverter fed induction motor drive in modern and industrial applications. The drive utilized for industrial and modern applications should use “spread spectrum” innovation known as Random pulse width modulation (PWM) algorithms where acoustic noise emanates through the machine should be critically concerned. This paper illustrates three types of random PWM control algorithms with fixed switching frequency namely 1) Random modulating PWM 2) Random carrier PWM and 3) Random modulating-carrier PWM. The spectrum plots of the motor stator current demonstrate the strength and robustness of the proposed PWM algorithms. To affirm the proposed algorithms, experimental tests have been conducted using dSPACE rt1104 control board on a v/f control three phase induction motor drive fed by DC link cascaded multilevel inverter.
Keywords: Multilevel inverter, acoustic noise, CSVPWM, total harmonic distortion, random PWM algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6622579 Color Characteristics of Dried Cocoa Using Shallow Box Fermentation Technique
Authors: Khairul Bariah Sulaiman, Tajul Aris Yang
Abstract:
Fermentation is well known as an essential process to develop chocolate flavor in dried cocoa beans. Besides developing the precursor of cocoa flavor, it also induces the color changes in the beans. The fermentation process is influenced by various factors such as planting material, preconditioning of cocoa pod and fermentation technique. Therefore, this study was conducted to evaluate color of Malaysian cocoa beans and how the duration of pods storage and fermentation technique using shallow box will effect on its color characteristics. There are two factors being studied i.e. duration of cocoa pod storage (0, 2, 4 and 6 days) and duration of cocoa fermentation (0, 1, 2, 3, 4 and 5 days). The experiment is arranged in 4 x 6 factorial designs with 24 treatments and arrangement is in a Completely Randomised Design (CRD). The produced beans are inspected for color changes under artificial light during cut test and divided into four groups of color namely fully brown, purple brown, fully purple and slaty. Cut tests indicated that cocoa beans which are directly dried without undergone fermentation has the highest slaty percentage. However, application of pods storage before fermentation process is found to decrease the slaty percentage. In contrast, the percentages of fully brown beans start to dominate after two days of fermentation, especially from four and six days of pods storage batch. Whereas, almost all batches of cocoa beans have a percentage of fully purple less than 20%. Interestingly, the percentage of purple brown beans are scattered in the entire beans batch regardless any specific trend. Meanwhile, statistical analysis using General Linear Model showed that the pods storage has a significant effect on the color characteristic of the Malaysian dried beans compared to fermentation duration.Keywords: Cocoa beans, color, fermentation, shallow box.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29752578 DWT Based Image Steganalysis
Authors: Indradip Banerjee, Souvik Bhattacharyya, Gautam Sanyal
Abstract:
‘Steganalysis’ is one of the challenging and attractive interests for the researchers with the development of information hiding techniques. It is the procedure to detect the hidden information from the stego created by known steganographic algorithm. In this paper, a novel feature based image steganalysis technique is proposed. Various statistical moments have been used along with some similarity metric. The proposed steganalysis technique has been designed based on transformation in four wavelet domains, which include Haar, Daubechies, Symlets and Biorthogonal. Each domain is being subjected to various classifiers, namely K-nearest-neighbor, K* Classifier, Locally weighted learning, Naive Bayes classifier, Neural networks, Decision trees and Support vector machines. The experiments are performed on a large set of pictures which are available freely in image database. The system also predicts the different message length definitions.
Keywords: Steganalysis, Moments, Wavelet Domain, KNN, K*, LWL, Naive Bayes Classifier, Neural networks, Decision trees, SVM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2572