Search results for: Maximum likelihood detection (MLD).
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3154

Search results for: Maximum likelihood detection (MLD).

1894 Investigation of I/Q Imbalance in Coherent Optical OFDM System

Authors: R. S. Fyath, Mustafa A. B. Al-Qadi

Abstract:

The inphase/quadrature (I/Q) amplitude and phase imbalance effects are studied in coherent optical orthogonal frequency division multiplexing (CO-OFDM) systems. An analytical model for the I/Q imbalance is developed and supported by simulation results. The results indicate that the I/Q imbalance degrades the BER performance considerably.

Keywords: Coherent detection, I/Q imbalance, OFDM, optical communications

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2565
1893 Production Throughput Modeling under Five Uncertain Variables Using Bayesian Inference

Authors: Amir Azizi, Amir Yazid B. Ali, Loh Wei Ping

Abstract:

Throughput is an important measure of performance of production system. Analyzing and modeling of production throughput is complex in today-s dynamic production systems due to uncertainties of production system. The main reasons are that uncertainties are materialized when the production line faces changes in setup time, machinery break down, lead time of manufacturing, and scraps. Besides, demand changes are fluctuating from time to time for each product type. These uncertainties affect the production performance. This paper proposes Bayesian inference for throughput modeling under five production uncertainties. Bayesian model utilized prior distributions related to previous information about the uncertainties where likelihood distributions are associated to the observed data. Gibbs sampling algorithm as the robust procedure of Monte Carlo Markov chain was employed for sampling unknown parameters and estimating the posterior mean of uncertainties. The Bayesian model was validated with respect to convergence and efficiency of its outputs. The results presented that the proposed Bayesian models were capable to predict the production throughput with accuracy of 98.3%.

Keywords: Bayesian inference, Uncertainty modeling, Monte Carlo Markov chain, Gibbs sampling, Production throughput

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2141
1892 Error Detection and Correction for Onboard Satellite Computers Using Hamming Code

Authors: Rafsan Al Mamun, Md. Motaharul Islam, Rabana Tajrin, Nabiha Noor, Shafinaz Qader

Abstract:

In an attempt to enrich the lives of billions of people by providing proper information, security and a way of communicating with others, the need for efficient and improved satellites is constantly growing. Thus, there is an increasing demand for better error detection and correction (EDAC) schemes, which are capable of protecting the data onboard the satellites. The paper is aimed towards detecting and correcting such errors using a special algorithm called the Hamming Code, which uses the concept of parity and parity bits to prevent single-bit errors onboard a satellite in Low Earth Orbit. This paper focuses on the study of Low Earth Orbit satellites and the process of generating the Hamming Code matrix to be used for EDAC using computer programs. The most effective version of Hamming Code generated was the Hamming (16, 11, 4) version using MATLAB, and the paper compares this particular scheme with other EDAC mechanisms, including other versions of Hamming Codes and Cyclic Redundancy Check (CRC), and the limitations of this scheme. This particular version of the Hamming Code guarantees single-bit error corrections as well as double-bit error detections. Furthermore, this version of Hamming Code has proved to be fast with a checking time of 5.669 nanoseconds, that has a relatively higher code rate and lower bit overhead compared to the other versions and can detect a greater percentage of errors per length of code than other EDAC schemes with similar capabilities. In conclusion, with the proper implementation of the system, it is quite possible to ensure a relatively uncorrupted satellite storage system.

Keywords: Bit-flips, Hamming code, low earth orbit, parity bits, satellite, single error upset.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 903
1891 TBC for Protection of Al Alloy Aerospace Component

Authors: P. Niranatlumpong, H. Koiprasert, C. Sukhonket, K. Ninon, N. Coompreedee

Abstract:

The use of a conventional air plasma-sprayed thermal barrier coating (TBC) and a porous, functionally graded TBC as a thermal insulator for Al7075 alloy was explored. A quench test at 1200°C employing fast heating and cooling rates was setup to represent a dynamic thermal condition of an aerospace component. During the test, coated samples were subjected the ambient temperature of 1200°C for a very short time. This was followed by a rapid drop in temperature resulting in cracking of the coatings. For the conventional TBC, it was found that the temperature of the Al7075 substrate decreases with the increase in the ZrO2 topcoat thickness. However, at the topcoat thickness of 1100 µm, large horizontal cracks can be observed in the topcoat and at the topcoat thickness of 1600 µm, the topcoat delaminate during cooling after the quench test. The porous, functionally graded TBC with 600 µm thick topcoat, on the other hand, was found to be as effective at reducing the substrate temperature as the conventional TBC with 1100 µm thick topcoat. The maximum substrate temperature is about 213°C for the former and 208°C for the latter when a heating rate of 38°C/s was used. When the quench tests were conducted with a faster heating rate of 128°C/s, the Al7075 substrate heat up faster with a reduction in the maximum substrate temperatures. The substrate temperatures dropped from 297 to 212°C for the conventional TBC and from 213 to 155°C for the porous TBC, both with 600 µm thick topcoat. Segmentation cracks were observed in both coating after the quench test.

Keywords: Thermal barrier coating, Al7075, porous TBC, Quenching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2454
1890 Fault Classification of Double Circuit Transmission Line Using Artificial Neural Network

Authors: Anamika Jain, A. S. Thoke, R. N. Patel

Abstract:

This paper addresses the problems encountered by conventional distance relays when protecting double-circuit transmission lines. The problems arise principally as a result of the mutual coupling between the two circuits under different fault conditions; this mutual coupling is highly nonlinear in nature. An adaptive protection scheme is proposed for such lines based on application of artificial neural network (ANN). ANN has the ability to classify the nonlinear relationship between measured signals by identifying different patterns of the associated signals. One of the key points of the present work is that only current signals measured at local end have been used to detect and classify the faults in the double circuit transmission line with double end infeed. The adaptive protection scheme is tested under a specific fault type, but varying fault location, fault resistance, fault inception angle and with remote end infeed. An improved performance is experienced once the neural network is trained adequately, which performs precisely when faced with different system parameters and conditions. The entire test results clearly show that the fault is detected and classified within a quarter cycle; thus the proposed adaptive protection technique is well suited for double circuit transmission line fault detection & classification. Results of performance studies show that the proposed neural network-based module can improve the performance of conventional fault selection algorithms.

Keywords: Double circuit transmission line, Fault detection and classification, High impedance fault and Artificial Neural Network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3182
1889 Thresholding Approach for Automatic Detection of Pseudomonas aeruginosa Biofilms from Fluorescence in situ Hybridization Images

Authors: Zonglin Yang, Tatsuya Akiyama, Kerry S. Williamson, Michael J. Franklin, Thiruvarangan Ramaraj

Abstract:

Pseudomonas aeruginosa is an opportunistic pathogen that forms surface-associated microbial communities (biofilms) on artificial implant devices and on human tissue. Biofilm infections are difficult to treat with antibiotics, in part, because the bacteria in biofilms are physiologically heterogeneous. One measure of biological heterogeneity in a population of cells is to quantify the cellular concentrations of ribosomes, which can be probed with fluorescently labeled nucleic acids. The fluorescent signal intensity following fluorescence in situ hybridization (FISH) analysis correlates to the cellular level of ribosomes. The goals here are to provide computationally and statistically robust approaches to automatically quantify cellular heterogeneity in biofilms from a large library of epifluorescent microscopy FISH images. In this work, the initial steps were developed toward these goals by developing an automated biofilm detection approach for use with FISH images. The approach allows rapid identification of biofilm regions from FISH images that are counterstained with fluorescent dyes. This methodology provides advances over other computational methods, allowing subtraction of spurious signals and non-biological fluorescent substrata. This method will be a robust and user-friendly approach which will enable users to semi-automatically detect biofilm boundaries and extract intensity values from fluorescent images for quantitative analysis of biofilm heterogeneity.

Keywords: Image informatics, Pseudomonas aeruginosa, biofilm, FISH, computer vision, data visualization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1167
1888 Turbo-Coded Mobile Terrestrial Communication Systems in Urban and Suburban Areas for Wireless Multimedia Applications

Authors: F. Mehran

Abstract:

With the rapid popularization of internet services, it is apparent that the next generation terrestrial communication systems must be capable of supporting various applications like voice, video, and data. This paper presents the performance evaluation of turbo- coded mobile terrestrial communication systems, which are capable of providing high quality services for delay sensitive (voice or video) and delay tolerant (text transmission) multimedia applications in urban and suburban areas. Different types of multimedia information require different service qualities, which are generally expressed in terms of a maximum acceptable bit-error-rate (BER) and maximum tolerable latency. The breakthrough discovery of turbo codes allows us to significantly reduce the probability of bit errors with feasible latency. In a turbo-coded system, a trade-off between latency and BER results from the choice of convolutional component codes, interleaver type and size, decoding algorithm, and the number of decoding iterations. This trade-off can be exploited for multimedia applications by using optimal and suboptimal performance parameter amalgamations to achieve different service qualities. The results are therefore proposing an adaptive framework for turbo-coded wireless multimedia communications which incorporate a set of performance parameters that achieve an appropriate set of service qualities, depending on the application's requirements.

Keywords: Mobile communications, Turbo codes, wireless multimedia communication systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1587
1887 Computational Analysis of Cavity Effect over Aircraft Wing

Authors: P. Booma Devi, Dilip A. Shah

Abstract:

This paper seeks the potentials of studying aerodynamic characteristics of inward cavities called dimples, as an alternative to the classical vortex generators. Increasing stalling angle is a greater challenge in wing design. But our examination is primarily focused on increasing lift. In this paper, enhancement of lift is mainly done by introduction of dimple or cavity in a wing. In general, aircraft performance can be enhanced by increasing aerodynamic efficiency that is lift to drag ratio of an aircraft wing. Efficiency improvement can be achieved by improving the maximum lift co-efficient or by reducing the drag co-efficient. At the time of landing aircraft, high angle of attack may lead to stalling of aircraft. To avoid this kind of situation, increase in the stalling angle is warranted. Hence, improved stalling characteristic is the best way to ease landing complexity. Computational analysis is done for the wing segment made of NACA 0012. Simulation is carried out for 30 m/s free stream velocity over plain airfoil and different types of cavities. The wing is modeled in CATIA V5R20 and analyses are carried out using ANSYS CFX. Triangle and square shapes are used as cavities for analysis. Simulations revealed that cavity placed on wing segment shows an increase of maximum lift co-efficient when compared to normal wing configuration. Flow separation is delayed at downstream of the wing by the presence of cavities up to a particular angle of attack.

Keywords: Lift, square and rectangle dimples, enhancement of stall angle, cavity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1866
1886 Multiaxial Fatigue Analysis of a High Performance Nickel-Based Superalloy

Authors: P. Selva, B. Lorrain, J. Alexis, A. Seror, A. Longuet, C. Mary, F. Denard

Abstract:

Over the past four decades, the fatigue behavior of nickel-based alloys has been widely studied. However, in recent years, significant advances in the fabrication process leading to grain size reduction have been made in order to improve fatigue properties of aircraft turbine discs. Indeed, a change in particle size affects the initiation mode of fatigue cracks as well as the fatigue life of the material. The present study aims to investigate the fatigue behavior of a newly developed nickel-based superalloy under biaxial-planar loading. Low Cycle Fatigue (LCF) tests are performed at different stress ratios so as to study the influence of the multiaxial stress state on the fatigue life of the material. Full-field displacement and strain measurements as well as crack initiation detection are obtained using Digital Image Correlation (DIC) techniques. The aim of this presentation is first to provide an in-depth description of both the experimental set-up and protocol: the multiaxial testing machine, the specific design of the cruciform specimen and performances of the DIC code are introduced. Second, results for sixteen specimens related to different load ratios are presented. Crack detection, strain amplitude and number of cycles to crack initiation vs. triaxial stress ratio for each loading case are given. Third, from fractographic investigations by scanning electron microscopy it is found that the mechanism of fatigue crack initiation does not depend on the triaxial stress ratio and that most fatigue cracks initiate from subsurface carbides.

Keywords: Cruciform specimen, multiaxial fatigue, Nickelbased superalloy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2189
1885 Experimental Investigation of Heat Pipe with Annular Fins under Natural Convection at Different Inclinations

Authors: Gangacharyulu Dasaroju, Sumeet Sharma, Sanjay Singh

Abstract:

Heat pipe is characterised as superconductor of heat because of its excellent heat removal ability. The operation of several engineering system results in generation of heat. This may cause several overheating problems and lead to failure of the systems. To overcome this problem and to achieve desired rate of heat dissipation, there is need to study the performance of heat pipe with annular fins under free convection at different inclinations. This study demonstrates the effect of different mass flow rate of hot fluid into evaporator section on the condenser side heat transfer coefficient with annular fins under natural convection at different inclinations. In this study annular fins are used for the experimental work having dimensions of length of fin, thickness of fin and spacing of fin as 10 mm, 1 mm and 6 mm, respectively. The main aim of present study is to discover at what inclination angles the maximum heat transfer coefficient shall be achieved. The heat transfer coefficient on the external surface of heat pipe condenser section is determined by experimental method and then predicted by empirical correlations. The results obtained from experimental and Churchill and Chu relation for laminar are in fair agreement with not more than 22% deviation. It is elucidated the maximum heat transfer coefficient of 31.2 W/(m2-K) at 25˚ tilt angle and minimal condenser heat transfer coefficient of 26.4 W/(m2-K) is seen at 45˚ tilt angle and 200 ml/min mass flow rate. Inclination angle also affects the thermal performance of heat pipe. Beyond 25o inclination, heat transport rate starts to decrease.

Keywords: Annular fins, condenser heat transfer coefficient, heat pipe, natural convection, tilt angle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 840
1884 An Intelligent Transportation System for Safety and Integrated Management of Railway Crossings

Authors: M. Magrini, D. Moroni, G. Palazzese, G. Pieri, D. Azzarelli, A. Spada, L. Fanucci, O. Salvetti

Abstract:

Railway crossings are complex entities whose optimal management cannot be addressed unless with the help of an intelligent transportation system integrating information both on train and vehicular flows. In this paper, we propose an integrated system named SIMPLE (Railway Safety and Infrastructure for Mobility applied at level crossings) that, while providing unparalleled safety in railway level crossings, collects data on rail and road traffic and provides value-added services to citizens and commuters. Such services include for example alerts, via variable message signs to drivers and suggestions for alternative routes, towards a more sustainable, eco-friendly and efficient urban mobility. To achieve these goals, SIMPLE is organized as a System of Systems (SoS), with a modular architecture whose components range from specially-designed radar sensors for obstacle detection to smart ETSI M2M-compliant camera networks for urban traffic monitoring. Computational unit for performing forecast according to adaptive models of train and vehicular traffic are also included. The proposed system has been tested and validated during an extensive trial held in the mid-sized Italian town of Montecatini, a paradigmatic case where the rail network is inextricably linked with the fabric of the city. Results of the tests are reported and discussed.

Keywords: Intelligent Transportation Systems (ITS), railway, railroad crossing, smart camera networks, radar obstacle detection, real-time traffic optimization, IoT, ETSI M2M, transport safety.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1408
1883 Dosimetric Analysis of Intensity Modulated Radiotherapy versus 3D Conformal Radiotherapy in Adult Primary Brain Tumors: Regional Cancer Centre, India

Authors: Ravi Kiran Pothamsetty, Radha Rani Ghosh, Baby Paul Thaliath

Abstract:

Radiation therapy has undergone many advancements and evloved from 2D to 3D. Recently, with rapid pace of drug discoveries, cutting edge technology, and clinical trials has made innovative advancements in computer technology and treatment planning and upgraded to intensity modulated radiotherapy (IMRT) which delivers in homogenous dose to tumor and normal tissues. The present study was a hospital-based experience comparing two different conformal radiotherapy techniques for brain tumors. This analytical study design has been conducted at Regional Cancer Centre, India from January 2014 to January 2015. Ten patients have been selected after inclusion and exclusion criteria. All the patients were treated on Artiste Siemens Linac Accelerator. The tolerance level for maximum dose was 6.0 Gyfor lenses and 54.0 Gy for brain stem, optic chiasm and optical nerves as per RTOG criteria. Mean and standard deviation values of PTV98%, PTV 95% and PTV 2% in IMRT were 93.16±2.9, 95.01±3.4 and 103.1±1.1 respectively; for 3DCRT were 91.4±4.7, 94.17±2.6 and 102.7±0.39 respectively. PTV max dose (%) in IMRT and 3D-CRT were 104.7±0.96 and 103.9±1.0 respectively. Maximum dose to the tumor can be delivered with IMRT with acceptable toxicity limits. Variables such as expertise, location of tumor, patient condition, and TPS influence the outcome of the treatment.

Keywords: IMRT, 3D CRT, Brain, tumors, OARs, RTOG.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 815
1882 Data Compression in Ultrasonic Network Communication via Sparse Signal Processing

Authors: Beata Zima, Octavio A. Márquez Reyes, Masoud Mohammadgholiha, Jochen Moll, Luca De Marchi

Abstract:

This document presents the approach of using compressed sensing in signal encoding and information transferring within a guided wave sensor network, comprised of specially designed frequency steerable acoustic transducers (FSATs). Wave propagation in a damaged plate was simulated using commercial FEM-based software COMSOL. Guided waves were excited by means of FSATs, characterized by the special shape of its electrodes, and modeled using PIC255 piezoelectric material. The special shape of the FSAT, allows for focusing wave energy in a certain direction, accordingly to the frequency components of its actuation signal, which makes a larger monitored area available. The process begins when a FSAT detects and records reflection from damage in the structure, this signal is then encoded and prepared for transmission, using a combined approach, based on Compressed Sensing Matching Pursuit and Quadrature Amplitude Modulation (QAM). After codification of the signal is in binary, the information is transmitted between the nodes in the network. The message reaches the last node, where it is finally decoded and processed, to be used for damage detection and localization purposes. The main aim of the investigation is to determine the location of detected damage using reconstructed signals. The study demonstrates that the special steerable capabilities of FSATs, not only facilitate the detection of damage but also permit transmitting the damage information to a chosen area in a specific direction of the investigated structure.

Keywords: Data compression, ultrasonic communication, guided waves, FEM analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 364
1881 Study of Temperature Changes in Fars Province

Authors: A. Gandomkar, R. Dehghani

Abstract:

Climate change is a phenomenon has been based on the available evidence from a very long time ago and now its existence is very probable. The speed and nature of climate parameters changes at the middle of twentieth century has been different and its quickness more than the before and its trend changed to some extent comparing to the past. Climate change issue now regarded as not only one of the most common scientific topic but also a social political one, is not a new issue. Climate change is a complicated atmospheric oceanic phenomenon on a global scale and long-term. Precipitation pattern change, fast decrease of snowcovered resources and its rapid melting, increased evaporation, the occurrence of destroying floods, water shortage crisis, severe reduction at the rate of harvesting agricultural products and, so on are all the significant of climate change. To cope with this phenomenon, its consequences and events in which public instruction is the most important but it may be climate that no significant cant and effective action has been done so far. The present article is included a part of one surrey about climate change in Fars. The study area having annually mean temperature 14 and precipitation 320 mm .23 stations inside the basin with a common 37 year statistical period have been applied to the meteorology data (1974-2010). Man-kendal and change factor methods are two statistical methods, applying them, the trend of changes and the annual mean average temperature and the annual minimum mean temperature were studied by using them. Based on time series for each parameter, the annual mean average temperature and the mean of annual maximum temperature have a rising trend so that this trend is clearer to the mean of annual maximum temperature.

Keywords: Climate change, Coefficient Variation, Fars province, Man-Kendal method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1911
1880 Water Quality Trading with Equitable Total Maximum Daily Loads

Authors: S. Jamshidi, E. Feizi Ashtiani, M. Ardestani

Abstract:

Waste Load Allocation (WLA) strategies usually intend to find economic policies for water resource management. Water quality trading (WQT) is an approach that uses discharge permit market to reduce total environmental protection costs. This primarily requires assigning discharge limits known as total maximum daily loads (TMDLs). These are determined by monitoring organizations with respect to the receiving water quality and remediation capabilities. The purpose of this study is to compare two approaches of TMDL assignment for WQT policy in small catchment area of Haraz River, in north of Iran. At first, TMDLs are assigned uniformly for the whole point sources to keep the concentrations of BOD and dissolved oxygen (DO) at the standard level at checkpoint (terminus point). This was simply simulated and controlled by Qual2kw software. In the second scenario, TMDLs are assigned using multi objective particle swarm optimization (MOPSO) method in which the environmental violation at river basin and total treatment costs are minimized simultaneously. In both scenarios, the equity index and the WLA based on trading discharge permits (TDP) are calculated. The comparative results showed that using economically optimized TMDLs (2nd scenario) has slightly more cost savings rather than uniform TMDL approach (1st scenario). The former annually costs about 1 M$ while the latter is 1.15 M$. WQT can decrease these annual costs to 0.9 and 1.1 M$, respectively. In other word, these approaches may save 35 and 45% economically in comparison with command and control policy. It means that using multi objective decision support systems (DSS) may find more economical WLA, however its outcome is not necessarily significant in comparison with uniform TMDLs. This may be due to the similar impact factors of dischargers in small catchments. Conversely, using uniform TMDLs for WQT brings more equity that makes stakeholders not feel that much envious of difference between TMDL and WQT allocation. In addition, for this case, determination of TMDLs uniformly would be much easier for monitoring. Consequently, uniform TMDL for TDP market is recommended as a sustainable approach. However, economical TMDLs can be used for larger watersheds.

Keywords: Waste load allocation (WLA), Water quality trading (WQT), Total maximum daily loads (TMDLs), Haraz River, Multi objective particle swarm optimization (MOPSO), Equity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2057
1879 Piezoelectric Bimorph Harvester Based on Different Lead Zirconate Titanate Materials to Enhance Energy Collection

Authors: Irene Perez-Alfaro, Nieves Murillo, Carlos Bernal, Daniel Gil-Hernandez

Abstract:

Nowadays, the increasing applicability of internet of things (IoT) systems has changed the way that the world around is perceived. The massive interconnection of systems by means of sensing, processing and communication, allows multitude of data to be at our fingertips. In this way, countless advances have been made in different fields such as personal care, predictive maintenance in industry, quality control in production processes, security, and in everything imaginable. However, all these electronic systems have in common the need to be electrically powered. In this context, batteries and wires are the most commonly used solutions, but they are not a definitive solution in some applications, because of the attainability, the serviceability, or the performance requirements. Therefore, the need arises to look for other types of solutions based on energy harvesting and long-life electronics. Energy Harvesting can be defined as the action of capturing energy from the environment and store it for an instantaneous use or later use. Among the materials capable of harvesting energy from the environment, such as thermoelectrics, electromagnetics, photovoltaics or triboelectrics, the most suitable is the piezoelectric material. The phenomenon of piezoelectricity is one of the most powerful sources for energy harvesting, ranging from a few micro wats to hundreds of wats, depending on certain factors such as material type, geometry, excitation frequency, mechanical and electrical configurations, among others. In this research work, an exhaustive study is carried out on how different types of piezoelectric materials and electrical configurations influence the maximum power that a bimorph harvester is able to extract from mechanical vibrations. A series of experiments has been carried out in which the manufactured bimorph specimens are excited under fixed inertial vibrational conditions. In addition, in order to evaluate the dependence of the maximum transferred power, different load resistors are tested. In this way, the pure active power that achieves the maximum power transfer can be approximated. In this paper, we present the design of low-cost energy harvesting solutions based on piezoelectric smart materials with tunable frequency. The results obtained show the differences in energy extraction between the PZT materials studied and their electrical configurations. The aim of this work is to gain a better understanding of the behavior of piezoelectric materials, and the design process of bimorph PZT harvesters to optimize environmental energy extraction.

Keywords: Bimorph harvesters, electrical impedance, energy harvesting, piezoelectric, smart material.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 453
1878 Parametric Study on Dynamic Analysis of Composite Laminated Plate

Authors: Junaid Kameran Ahmed

Abstract:

A laminated plate composite of graphite/epoxy has been analyzed dynamically in the present work by using a quadratic element (8-node diso-parametric), and by depending on 1st order shear deformation theory, every node in this element has 6-degrees of freedom (displacement in x, y, and z axis and twist about x, y, and z axis). The dynamic analysis in the present work covered parametric studies on a composite laminated plate (square plate) to determine its effect on the natural frequency of the plate. The parametric study is represented by set of changes (plate thickness, number of layers, support conditions, layer orientation), and the plates have been simulated by using ANSYS package 12. The boundary conditions considered in this study, at all four edges of the plate, are simply supported and fixed boundary condition. The results obtained from ANSYS program show that the natural frequency for both fixed and simply supported increases with increasing the number of layers, but this increase in the natural frequency for the first five modes will be neglected after 10 layers. And it is observed that the natural frequency of a composite laminated plate will change with the change of ply orientation, the natural frequency increases and it will be at maximum with angle 45 of ply for simply supported laminated plate, and maximum natural frequency will be with cross-ply (0/90) for fixed laminated composite plate. It is also observed that the natural frequency increase is approximately doubled when the thickness is doubled.

Keywords: Laminated plate, orthotropic plate, square plate, natural frequency, graphite/epoxy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 974
1877 Liquid Chromatography Microfluidics for Detection and Quantification of Urine Albumin Using Linear Regression Method

Authors: Patricia B. Cruz, Catrina Jean G. Valenzuela, Analyn N. Yumang

Abstract:

Nearly a hundred per million of the Filipino population is diagnosed with Chronic Kidney Disease (CKD). The early stage of CKD has no symptoms and can only be discovered once the patient undergoes urinalysis. Over the years, different methods were discovered and used for the quantification of the urinary albumin such as the immunochemical assays where most of these methods require large machinery that has a high cost in maintenance and resources, and a dipstick test which is yet to be proven and is still debated as a reliable method in detecting early stages of microalbuminuria. This research study involves the use of the liquid chromatography concept in microfluidic instruments with biosensor as a means of separation and detection respectively, and linear regression to quantify human urinary albumin. The researchers’ main objective was to create a miniature system that quantifies and detect patients’ urinary albumin while reducing the amount of volume used per five test samples. For this study, 30 urine samples of unknown albumin concentrations were tested using VITROS Analyzer and the microfluidic system for comparison. Based on the data shared by both methods, the actual vs. predicted regression were able to create a positive linear relationship with an R2 of 0.9995 and a linear equation of y = 1.09x + 0.07, indicating that the predicted values and actual values are approximately equal. Furthermore, the microfluidic instrument uses 75% less in total volume – sample and reagents combined, compared to the VITROS Analyzer per five test samples.

Keywords: Chronic kidney disease, microfluidics, linear regression, VITROS analyzer, urinary albumin.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 860
1876 Architecture, Implementation and Application of Tools for Experimental Analysis

Authors: Tom Dowling, Adam Duffy

Abstract:

This paper presents an architecture to assist in the development of tools to perform experimental analysis. Existing implementations of tools based on this architecture are also described in this paper. These tools are applied to the real world problem of fault attack emulation and detection in cryptographic algorithms.

Keywords: Software Architectures and Design, Software Componentsand Reuse, Engineering Secure Software.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1397
1875 High Temperature Deformation Behavior of Cr-containing Superplastic Iron Aluminide

Authors: Seok Hong Min, Woo Young Jung, Tae Kwon Ha

Abstract:

Superplastic deformation and high temperature load relaxation behavior of coarse-grained iron aluminides with the composition of Fe-28 at.% Al have been investigated. A series of load relaxation and tensile tests were conducted at temperatures ranging from 600 to 850oC. The flow curves obtained from load relaxation tests were found to have a sigmoidal shape and to exhibit stress vs. strain rate data in a very wide strain rate range from 10-7/s to 10-2/s. Tensile tests have been conducted at various initial strain rates ranging from 3×10-5/s to 1×10-2/s. Maximum elongation of ~500 % was obtained at the initial strain rate of 3×10-5/s and the maximum strain rate sensitivity was found to be 0.68 at 850oC in binary Fe-28Al alloy. Microstructure observation through the optical microscopy (OM) and the electron back-scattered diffraction (EBSD) technique has been carried out on the deformed specimens and it has revealed the evidences for grain boundary migration and grain refinement to occur during superplastic deformation, suggesting the dynamic recrystallization mechanism. The addition of Cr by the amount of 5 at.% appeared to deteriorate the superplasticity of the binary iron aluminide. By applying the internal variable theory of structural superplasticity, the addition of Cr has been revealed to lower the contribution of the frictional resistance to dislocation glide during high temperature deformation of the Fe3Al alloy.

Keywords: Iron aluminide (Fe3Al), large grain size, structural superplasticity, dynamic recrystallization, chromium (Cr).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1783
1874 Simultaneous Saccharification and Fermentation(SSF) of Sugarcane Bagasse - Kinetics and Modeling

Authors: E.Sasikumar, T.Viruthagiri

Abstract:

Simultaneous Saccharification and Fermentation (SSF) of sugarcane bagasse by cellulase and Pachysolen tannophilus MTCC *1077 were investigated in the present study. Important process variables for ethanol production form pretreated bagasse were optimized using Response Surface Methodology (RSM) based on central composite design (CCD) experiments. A 23 five level CCD experiments with central and axial points was used to develop a statistical model for the optimization of process variables such as incubation temperature (25–45°) X1, pH (5.0–7.0) X2 and fermentation time (24–120 h) X3. Data obtained from RSM on ethanol production were subjected to the analysis of variance (ANOVA) and analyzed using a second order polynomial equation and contour plots were used to study the interactions among three relevant variables of the fermentation process. The fermentation experiments were carried out using an online monitored modular fermenter 2L capacity. The processing parameters setup for reaching a maximum response for ethanol production was obtained when applying the optimum values for temperature (32°C), pH (5.6) and fermentation time (110 h). Maximum ethanol concentration (3.36 g/l) was obtained from 50 g/l pretreated sugarcane bagasse at the optimized process conditions in aerobic batch fermentation. Kinetic models such as Monod, Modified Logistic model, Modified Logistic incorporated Leudeking – Piret model and Modified Logistic incorporated Modified Leudeking – Piret model have been evaluated and the constants were predicted.

Keywords: Sugarcane bagasse, ethanol, optimization, Pachysolen tannophilus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2297
1873 Implementation of an On-Line PD Measurement System Using HFCT

Authors: F. Haghjoo, M. Sarlak, S.M. Shahrtash

Abstract:

In order to perform on-line measuring and detection of PD signals, a total solution composing of an HFCT, A/D converter and a complete software package is proposed. The software package includes compensation of HFCT contribution, filtering and noise reduction using wavelet transform and soft calibration routines. The results have shown good performance and high accuracy.

Keywords: Partial Discharge, Measurement, On-line, HFCT

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1814
1872 Novel Adaptive Channel Equalization Algorithms by Statistical Sampling

Authors: János Levendovszky, András Oláh

Abstract:

In this paper, novel statistical sampling based equalization techniques and CNN based detection are proposed to increase the spectral efficiency of multiuser communication systems over fading channels. Multiuser communication combined with selective fading can result in interferences which severely deteriorate the quality of service in wireless data transmission (e.g. CDMA in mobile communication). The paper introduces new equalization methods to combat interferences by minimizing the Bit Error Rate (BER) as a function of the equalizer coefficients. This provides higher performance than the traditional Minimum Mean Square Error equalization. Since the calculation of BER as a function of the equalizer coefficients is of exponential complexity, statistical sampling methods are proposed to approximate the gradient which yields fast equalization and superior performance to the traditional algorithms. Efficient estimation of the gradient is achieved by using stratified sampling and the Li-Silvester bounds. A simple mechanism is derived to identify the dominant samples in real-time, for the sake of efficient estimation. The equalizer weights are adapted recursively by minimizing the estimated BER. The near-optimal performance of the new algorithms is also demonstrated by extensive simulations. The paper has also developed a (Cellular Neural Network) CNN based approach to detection. In this case fast quadratic optimization has been carried out by t, whereas the task of equalizer is to ensure the required template structure (sparseness) for the CNN. The performance of the method has also been analyzed by simulations.

Keywords: Cellular Neural Network, channel equalization, communication over fading channels, multiuser communication, spectral efficiency, statistical sampling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1517
1871 Detection of Lard in Binary Animal Fats and Vegetable Oils Mixtures and in Some Commercial Processed Foods

Authors: H. A. Al-Kahtani, A. A. Abou Arab, M. Asif

Abstract:

Animal fats (camel, sheep, goat, rabbit and chicken) and vegetable oils (corn, sunflower, palm oil and olive oil) were substituted with different proportions (1, 5, 10 and 20%) of lard. Fatty acid composition in TG and 2-MG were determined using lipase hydrolysis and gas chromatography before and after adulteration. Results indicated that, genuine lard had a high proportion (60.97%) of the total palmitic acid at 2-MG. However, it was 8.70%, 16.40%, 11.38%, 10.57%, 29.97 and 8.97% for camel, beef, sheep, goat, rabbit and chicken, respectively. It could be noticed also the position-2-MG is mostly occupied by unsaturated fatty acids among all tested fats except lard. Vegetable oils (corn, sunflower, palm oil and olive oil) revealed that the levels of palmitic acid esterifies at 2-MG position was 6.84, 1.43, 9.86 and 1.70%, respectively. It could be observed also the studied oils had a higher level of unsaturated fatty acids in the same position, compared with animal fats under investigation. Moreover, palmitic acid esterifies at 2-MG and PAEF increased gradually as the substituted levels increased among all tested fat and oil samples. Statistical analysis showed that the PAEF correlated well with lard level. The detection of lard in some commercial processed foods (5 French fries, 4 Butter fats, 5 processed meat and 6 candy samples) was carried out. Results revealed that 2 samples of French fries and 4 samples of processed meat contained lard due to their higher PAEF, while butter fat and candy were free of lard.

Keywords: Lard, adulteration, PAEF, goat, triglycerides.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2938
1870 Effect of Architecture and Operating Conditions of Vehicle on Bulb Lifetime in Automotive

Authors: Hatice Özbek, Caner Çil, Ahmet Rodoplu

Abstract:

Automotive lighting is the leading function in the configuration of vehicle architecture. Especially headlights and taillights from external lighting functions are among the structures that determine the stylistic character of the vehicle. At the same time, the fact that lighting functions are related to many other functions brings along difficulties in design. Customers expect maximum quality from the vehicle. In these circumstances, it is necessary to make designs that aim to keep the performance of bulbs with limited working lives at the highest level. With this study, the factors that influence the working lives of filament lamps were examined and bulb explosions that can occur sooner than anticipated in the future were prevented while the vehicle was still in the design phase by determining the relations with electrical, dynamical and static variables. Especially the filaments of the bulbs used in the front lighting of the vehicle are deformed in a shorter time due to the high voltage requirement. In addition to this, rear lighting lamps vibrate as a result of the tailgate opening and closing and cause the filaments to be exposed to high stress. With this study, the findings that cause bulb explosions were evaluated. Among the most important findings: 1. The structure of the cables to the lighting functions of the vehicle and the effect of the voltage values are drawn; 2. The effect of the vibration to bulb throughout the life of the vehicle; 3 The effect of the loads carried to bulb while the vehicle doors are opened and closed. At the end of the study, the maximum performance was established in the bulb lifetimes with the optimum changes made in the vehicle architecture based on the findings obtained.

Keywords: Vehicle architecture, automotive lighting functions, filament lamps, bulb lifetime.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 776
1869 Advanced Stochastic Models for Partially Developed Speckle

Authors: Jihad S. Daba (Jean-Pierre Dubois), Philip Jreije

Abstract:

Speckled images arise when coherent microwave, optical, and acoustic imaging techniques are used to image an object, surface or scene. Examples of coherent imaging systems include synthetic aperture radar, laser imaging systems, imaging sonar systems, and medical ultrasound systems. Speckle noise is a form of object or target induced noise that results when the surface of the object is Rayleigh rough compared to the wavelength of the illuminating radiation. Detection and estimation in images corrupted by speckle noise is complicated by the nature of the noise and is not as straightforward as detection and estimation in additive noise. In this work, we derive stochastic models for speckle noise, with an emphasis on speckle as it arises in medical ultrasound images. The motivation for this work is the problem of segmentation and tissue classification using ultrasound imaging. Modeling of speckle in this context involves partially developed speckle model where an underlying Poisson point process modulates a Gram-Charlier series of Laguerre weighted exponential functions, resulting in a doubly stochastic filtered Poisson point process. The statistical distribution of partially developed speckle is derived in a closed canonical form. It is observed that as the mean number of scatterers in a resolution cell is increased, the probability density function approaches an exponential distribution. This is consistent with fully developed speckle noise as demonstrated by the Central Limit theorem.

Keywords: Doubly stochastic filtered process, Poisson point process, segmentation, speckle, ultrasound

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1737
1868 Double Pass Solar Air Heater with Transvers Fins and without Absorber Plate

Authors: A. J. Mahmood, L. B. Y. Aldabbagh

Abstract:

The counter flow solar air heaters, with four transverse fins and wire mesh layers are constructed and investigated experimentally for thermal efficiency at a geographic location of Cyprus in the city of Famagusta. The absorber plate is replaced by sixteen steel wire mesh layers, 0.18 x 0.18cm in cross section opening and a 0.02cm in diameter. The wire mesh layers arranged in three groups, first and second include 6 layers, while the third include 4 layers. All layers fixed in the duct parallel to the glazing and each group separated from the others by wood frame thickness of 0.5cm to reduce the pressure drop. The transverse fins arranged in a way to force the air to flow through the bed like eight letter path with flow depth 3cm. The proposed design has increased the heat transfer rate, but on other hand causes a high pressure drop. The obtained results show that, for air mass flow rate range between 0.011-0.036kg/s, the thermal efficiency increases with increasing the air mass flow. The maximum efficiency obtained is 65.6% for the mass flow rate of 0.036kg/s. Moreover, the temperature difference between the outlet flow and the ambient temperature, ΔT, reduces as the air mass flow rate increase. The maximum difference between the outlet and ambient temperature obtained was 43°C for double pass for minimum mass flow rate of 0.011kg/s. Comparison with a conventional solar air heater collector shows a significantly development in the thermal efficiency.

Keywords: Counter flow, solar air heater (SAH), Wire mesh, Fins, Thermal efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3174
1867 Geotechnical Properties and Compressibility Behavior of Organic Dredged Soils

Authors: Inci Develioglu, Hasan Firat Pulat

Abstract:

Sustainable development is one of the most important topics in today's world, and it is also an important research topic for geoenvironmental engineering. Dredging process is performed to expand the river and port channel, flood control and accessing harbors. Every year large amount of sediment are dredged for these purposes. Dredged marine soils can be reused as filling materials, road and foundation embankments, construction materials and wildlife habitat developments. In this study, geotechnical engineering properties and compressibility behavior of dredged soil obtained from the Izmir Bay were investigated. The samples with four different organic matter contents were obtained and particle size distributions, consistency limits, pH and specific gravity tests were performed. The consolidation tests were conducted to examine organic matter content (OMC) effects on compressibility behavior of dredged soil. This study has shown that the OMC has an important effect on the engineering properties of dredged soils. The liquid and plastic limits increased with increasing OMC. The lowest specific gravity belonged to sample which has the maximum OMC. The specific gravity values ranged between 2.76 and 2.52. The maximum void ratio difference belongs to sample with the highest OMC (De11% = 0.38). As the organic matter content of the samples increases, the change in the void ratio has also increased. The compression index increases with increasing OMC.

Keywords: Compressibility, consolidation, geotechnical properties, organic matter content, organic soils.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951
1866 Battery Grading Algorithm in 2nd-Life Repurposing Li-ion Battery System

Authors: Ya Lv, Benjamin Ong Wei Lin, Wanli Niu, Benjamin Seah Chin Tat

Abstract:

This article presents a methodology that improves reliability and cyclability of 2nd-life Li-ion battery system repurposed as energy storage system (ESS). Most of the 2nd-life retired battery systems in market have module/pack-level state of health (SOH) indicator, which is utilized for guiding appropriate depth of discharge (DOD) in the application of ESS. Due to the lack of cell-level SOH indication, the different degrading behaviors among various cells cannot be identified upon reaching retired status; in the end, considering end of life (EOL) loss and pack-level DOD, the repurposed ESS has to be oversized by > 1.5 times to complement the application requirement of reliability and cyclability. This proposed battery grading algorithm, using non-invasive methodology, is able to detect outlier cells based on historical voltage data and calculate cell-level historical maximum temperature data using semi-analytic methodology. In this way, the individual battery cell in the 2nd-life battery system can be graded in terms of SOH on basis of the historical voltage fluctuation and estimated historical maximum temperature variation. These grades will have corresponding DOD grades in the application of the repurposed ESS to enhance the system reliability and cyclability. In all, this introduced battery grading algorithm is non-invasive, compatible with all kinds of retired Li-ion battery systems which lack of cell-level SOH indication, as well as potentially being embedded into battery management software for preventive maintenance and real-time cyclability optimization.

Keywords: Battery grading algorithm, 2nd-life repurposing battery system, semi-analytic methodology, reliability and cyclability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 834
1865 The Impact of Recommendation Sources on Online Purchase Intentions: The Moderating Effects of Gender and Perceived Risk

Authors: Chiao-Chen Chang, Yang-Chieh Chin

Abstract:

This study examines the issue of recommendation sources from the perspectives of gender and consumers- perceived risk, and validates a model for the antecedents of consumer online purchases. The method of obtaining quantitative data was that of the instrument of a survey questionnaire. Data were collected via questionnaires from 396 undergraduate students aged 18-24, and a multiple regression analysis was conducted to identify causal relationships. Empirical findings established the link between recommendation sources (word-of-mouth, advertising, and recommendation systems) and the likelihood of making online purchases and demonstrated the role of gender and perceived risk as moderators in this context. The results showed that the effects of word-of-mouth on online purchase intentions were stronger than those of advertising and recommendation systems. In addition, female consumers have less experience with online purchases, so they may be more likely than males to refer to recommendations during the decision-making process. The findings of the study will help marketers to address the recommendation factor which influences consumers- intention to purchase and to improve firm performances to meet consumer needs.

Keywords: Recommendation sources, Online purchaseintentions, Gender differences, Perceived risk.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3015