Search results for: real/complex values.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5686

Search results for: real/complex values.

1096 Selection of Best Band Combination for Soil Salinity Studies using ETM+ Satellite Images (A Case study: Nyshaboor Region,Iran)

Authors: Sanaeinejad, S. H.; A. Astaraei, . P. Mirhoseini.Mousavi, M. Ghaemi,

Abstract:

One of the main environmental problems which affect extensive areas in the world is soil salinity. Traditional data collection methods are neither enough for considering this important environmental problem nor accurate for soil studies. Remote sensing data could overcome most of these problems. Although satellite images are commonly used for these studies, however there are still needs to find the best calibration between the data and real situations in each specified area. Neyshaboor area, North East of Iran was selected as a field study of this research. Landsat satellite images for this area were used in order to prepare suitable learning samples for processing and classifying the images. 300 locations were selected randomly in the area to collect soil samples and finally 273 locations were reselected for further laboratory works and image processing analysis. Electrical conductivity of all samples was measured. Six reflective bands of ETM+ satellite images taken from the study area in 2002 were used for soil salinity classification. The classification was carried out using common algorithms based on the best composition bands. The results showed that the reflective bands 7, 3, 4 and 1 are the best band composition for preparing the color composite images. We also found out, that hybrid classification is a suitable method for identifying and delineation of different salinity classes in the area.

Keywords: Soil salinity, Remote sensing, Image processing, ETM+, Nyshaboor

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2021
1095 Building a Transformative Continuing Professional Development Experience for Educators through a Principle-Based, Technological-Driven Knowledge Building Approach: A Case Study of a Professional Learning Team in Secondary Education

Authors: Melvin Chan, Chew Lee Teo

Abstract:

There has been a growing emphasis in elevating the teachers’ proficiency and competencies through continuing professional development (CPD) opportunities. In this era of a Volatile, Uncertain, Complex, Ambiguous (VUCA) world, teachers are expected to be collaborative designers, critical thinkers and creative builders. However, many of the CPD structures are still revolving in the model of transmission, which stands in contradiction to the cultivation of future-ready teachers for the innovative world of emerging technologies. This article puts forward the framing of CPD through a Principle-Based, Technological-Driven Knowledge Building Approach grounded in the essence of andragogy and progressive learning theories where growth is best exemplified through an authentic immersion in a social/community experience-based setting. Putting this Knowledge Building Professional Development Model (KBPDM) in operation via a Professional Learning Team (PLT) situated in a Secondary School in Singapore, research findings reveal that the intervention has led to a fundamental change in the learning paradigm of the teachers, henceforth equipping and empowering them successfully in their pedagogical design and practices for a 21st century classroom experience. This article concludes with the possibility in leveraging the Learning Analytics to deepen the CPD experiences for educators.

Keywords: Continual professional development, knowledge building, learning paradigm, andragogy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1013
1094 Device for 3D Analysis of Basic Movements of the Lower Extremity

Authors: Jiménez Villanueva Mayra Alejandra, Ortíz Casallas Diana Carolina, Luengas Contreras Lely Adriana

Abstract:

This document details the process of developing a wireless device that captures the basic movements of the foot (plantar flexion, dorsal flexion, abduction, adduction.), and the knee movement (flexion). It implements a motion capture system by using a hardware based on optical fiber sensors, due to the advantages in terms of scope, noise immunity and speed of data transmission and reception. The operating principle used by this system is the detection and transmission of joint movement by mechanical elements and their respective measurement by optical ones (in this case infrared). Likewise, Visual Basic software is used for reception, analysis and signal processing of data acquired by the device, generating a 3D graphical representation in real time of each movement. The result is a boot in charge of capturing the movement, a transmission module (Implementing Xbee Technology) and a receiver module for receiving information and sending it to the PC for their respective processing. The main idea with this device is to help on topics such as bioengineering and medicine, by helping to improve the quality of life and movement analysis.

Keywords: abduction, adduction, A / D converter, Autodesk 3DMax, Infrared Diode, Driver, extension, flexion, Infrared LEDs, Interface, Modeling OPENGL, Optical Fiber, USB CDC(Communications Device Class), Virtual Reality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1694
1093 The High Strength Biocompatible Wires of Commercially Pure Titanium

Authors: J. Palán, M. Zemko

Abstract:

COMTES FHT has been active in a field of research and development of high-strength wires for quite some time. The main material was pure titanium. The primary goal of this effort is to develop a continuous production process for ultrafine and nanostructured materials with the aid of severe plastic deformation (SPD). This article outlines mechanical and microstructural properties of the materials and the options available for testing the components made of these materials. Ti Grade 2 and Grade 4 wires are the key products of interest. Ti Grade 2 with ultrafine to nano-sized grain shows ultimate strength of up to 1050 MPa. Ti Grade 4 reaches ultimate strengths of up to 1250 MPa. These values are twice or three times as higher as those found in the unprocessed material. For those fields of medicine where implantable metallic materials are used, bulk ultrafine to nanostructured titanium is available. It is manufactured by SPD techniques. These processes leave the chemical properties of the initial material unchanged but markedly improve its final mechanical properties, in particular, the strength. Ultrafine to nanostructured titanium retains all the significant and, from the biological viewpoint, desirable properties that are important for its use in medicine, i.e. those properties which made pure titanium the preferred material also for dental implants.

Keywords: CONFORM SPD, ECAP, titanium, rotary swaging.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 987
1092 Wine Grape Residues Gasification in Supercritical Water

Authors: D. Selvi Gökkaya, M. Yüksel, M. Sağlam, T. Güngören Madenoğlu, N. Cengiz, T. Çokkuvvetli, L. Ballice

Abstract:

In this study, production possibilities of hydrogen and/or methane via SCWG from black grape residues have been investigated. For this aim, grape residues which remain as a byproduct of the wine making process have been used. Since utilization from grape residues is limited due to the high moisture content, supercritical water gasification is the most convenient method. The effect of the gasification temperature and type of catalyst on supercritical water gasification have been investigated. Gasification experiments were performed in a batch autoclave at four different temperatures 300, 400, 500 and 600°C. K2CO3 and Trona (NaHCO3.Na2CO3·2H2O) were used as catalyst. Real biomass types of black grape residues have been successfully gasified and the product gas (hydrogen, methane, carbon dioxide, carbon monoxide and a small amount of ethane and ethylene) were identified by using gas chromatography. A TOC analyzer was used to determine total organic carbon (TOC) content of aqueous phase. The amounts of carboxylic acids, aldehydes, ketones, furfurals and phenols present in the aqueous solutions were analyzed by high performance liquid chromatography. When the temperature increased from 300°C to 600°C, mol% of H2 in gas products increased. The presence of catalysts improves the hydrogen yield. Trona showed gasification activity to be similar to that of K2CO3. It may be concluded that the use of Trona instead of commercially produced catalysts, can be preferably used in the gasification of biomass in supercritical water.

Keywords: Biomass, hydrogen, grape residues.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2465
1091 DC Bus Voltage Regulator for Renewable Energy Based Microgrid Application

Authors: Bakari M. M. Mwinyiwiwa

Abstract:

Renewable Energy based microgrids are being considered to provide electricity for the expanding energy demand in the grid distribution network and grid isolated areas. The technical challenges associated with the operation and controls are immense. Electricity generation by Renewable Energy Sources is of stochastic nature such that there is a demand for regulation of voltage output in order to satisfy the standard loads’ requirements. In a renewable energy based microgrid, the energy sources give stochastically variable magnitude AC or DC voltages. AC voltage regulation of micro and mini sources pose practical challenges as well as unbearable costs. It is therefore practically and economically viable to convert the voltage outputs from stochastic AC and DC voltage sources to constant DC voltage to satisfy various DC loads including inverters which ultimately feed AC loads. This paper presents results obtained from SEPIC converter based DC bus voltage regulator as a case study for renewable energy microgrid application. Real-Time Simulation results show that upon appropriate choice of controller parameters for control of the SEPIC converter, the output DC bus voltage can be kept constant regardless of wide range of voltage variations of the source. This feature is particularly important in the situation that multiple renewable sources are to be integrated to supply a microgrid under main grid integration or isolated modes of operation.

Keywords: DC Voltage Regulator, microgrid, multisource, Renewable Energy, SEPIC Converter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4311
1090 Atmospheric Fluid Bed Gasification of Different Biomass Fuels

Authors: Martin Lisý, Marek Baláš, Michal Špiláček, Zdeněk Skála

Abstract:

This paper shortly describes various types of biomass and a growing number of facilities utilizing the biomass in the Czech Republic. The considerable part of this paper deals with energy parameters of the most frequently used types of biomass and results of their gasification testing. Sixteen most used "Czech" woody plants and grasses were selected; raw, element and biochemical analyses were performed and basic calorimetric values, ash composition, and ash characteristic temperatures were identified. Later, each biofuel was tested in a fluidized bed gasifier. The essential part of this paper provides results of the gasification of selected biomass types. Operating conditions are described in detail with a focus on individual fuels properties. Gas composition and impurities content are also identified. In terms of operating conditions and gas quality, the essential difference occurred mainly between woody plants and grasses. The woody plants were evaluated as more suitable fuels for fluidized bed gasifiers. Testing results significantly help with a decision-making process regarding suitability of energy plants for growing and with a selection of optimal biomass-treatment technology.

Keywords: Biomass Growing, Biomass Types, Gasification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1923
1089 Applied Actuator Fault Accommodation in Flight Control Systems Using Fault Reconstruction Based FDD and SMC Reconfiguration

Authors: A. Ghodbane, M. Saad, J.-F. Boland, C. Thibeault

Abstract:

Historically, actuators’ redundancy was used to deal with faults occurring suddenly in flight systems. This technique was generally expensive, time consuming and involves increased weight and space in the system. Therefore, nowadays, the on-line fault diagnosis of actuators and accommodation plays a major role in the design of avionic systems. These approaches, known as Fault Tolerant Flight Control systems (FTFCs) are able to adapt to such sudden faults while keeping avionics systems lighter and less expensive. In this paper, a (FTFC) system based on the Geometric Approach and a Reconfigurable Flight Control (RFC) are presented. The Geometric approach is used for cosmic ray fault reconstruction, while Sliding Mode Control (SMC) based on Lyapunov stability theory is designed for the reconfiguration of the controller in order to compensate the fault effect. Matlab®/Simulink® simulations are performed to illustrate the effectiveness and robustness of the proposed flight control system against actuators’ faulty signal caused by cosmic rays. The results demonstrate the successful real-time implementation of the proposed FTFC system on a non-linear 6 DOF aircraft model.

Keywords: Actuators’ faults, Fault detection and diagnosis, Fault tolerant flight control, Sliding mode control, Geometric approach for fault reconstruction, Lyapunov stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2576
1088 Flow Visualization and Characterization of an Artery Model with Stenosis

Authors: Anis S. Shuib, Peter R. Hoskins, William J. Easson

Abstract:

Cardiovascular diseases, principally atherosclerosis, are responsible for 30% of world deaths. Atherosclerosis is due to the formation of plaque. The fatty plaque may be at risk of rupture, leading typically to stroke and heart attack. The plaque is usually associated with a high degree of lumen reduction, called a stenosis.It is increasingly recognized that the initiation and progression of disease and the occurrence of clinical events is a complex interplay between the local biomechanical environment and the local vascular biology. The aim of this study is to investigate the flow behavior through a stenosed artery. A physical experiment was performed using an artery model and blood analogue fluid. An axisymmetric model constructed consists of contraction and expansion region that follow a mathematical form of cosine function. A 30% diameter reduction was used in this study. The flow field was measured using particle image velocimetry (PIV). Spherical particles with 20μm diameter were seeded in a water-glycerol-NaCl mixture. Steady flow Reynolds numbers are 250. The area of interest is the region after the stenosis where the flow separation occurs. The velocity field was measured and the velocity gradient was investigated. There was high particle concentration in the recirculation zone. High velocity gradient formed immediately after the stenosis throat created a lift force that enhanced particle migration to the flow separation area.

Keywords: Stenosis artery, Biofluid mechanics, PIV

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2003
1087 Effects of Using Gusset Plate Stiffeners on the Seismic Performance of Concentrically Braced Frame

Authors: B. Mohebi, N. Asadi, F. Kazemi

Abstract:

Inelastic deformation of the brace in Special Concentrically Braced Frame (SCBF) creates inelastic damages on gusset plate connections such as buckling at edges. In this study, to improve the seismic performance of SCBFs connections, an analytical study was undertaken. To improve the gusset plate connection, this study proposes using ‎edge’s stiffeners in both sides of gusset plate.‎ For this purpose, in order to examine edge’s stiffeners effect on gusset plate connections, two groups of modeling with and without considering edge’s stiffener and different types of braces were modeled using ABAQUS software. The results show that considering the edge’s stiffener reduces the equivalent plastic strain values at a connection region of gusset plate with beam and column, which can improve the seismic performance of gusset plate. Furthermore, considering the edge’s stiffeners significantly decreases the strain concentration at regions where gusset plates have been connected to beam and column. Moreover, considering 2tpl distance causes reduction in the plastic strain.

Keywords: Special concentrically braced frame, gusset plate, edge’s stiffener, seismic performance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 734
1086 A Hybrid Neural Network and Traditional Approach for Forecasting Lumpy Demand

Authors: A. Nasiri Pour, B. Rostami Tabar, A.Rahimzadeh

Abstract:

Accurate demand forecasting is one of the most key issues in inventory management of spare parts. The problem of modeling future consumption becomes especially difficult for lumpy patterns, which characterized by intervals in which there is no demand and, periods with actual demand occurrences with large variation in demand levels. However, many of the forecasting methods may perform poorly when demand for an item is lumpy. In this study based on the characteristic of lumpy demand patterns of spare parts a hybrid forecasting approach has been developed, which use a multi-layered perceptron neural network and a traditional recursive method for forecasting future demands. In the described approach the multi-layered perceptron are adapted to forecast occurrences of non-zero demands, and then a conventional recursive method is used to estimate the quantity of non-zero demands. In order to evaluate the performance of the proposed approach, their forecasts were compared to those obtained by using Syntetos & Boylan approximation, recently employed multi-layered perceptron neural network, generalized regression neural network and elman recurrent neural network in this area. The models were applied to forecast future demand of spare parts of Arak Petrochemical Company in Iran, using 30 types of real data sets. The results indicate that the forecasts obtained by using our proposed mode are superior to those obtained by using other methods.

Keywords: Lumpy Demand, Neural Network, Forecasting, Hybrid Approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2680
1085 Least Square-SVM Detector for Wireless BPSK in Multi-Environmental Noise

Authors: J. P. Dubois, Omar M. Abdul-Latif

Abstract:

Support Vector Machine (SVM) is a statistical learning tool developed to a more complex concept of structural risk minimization (SRM). In this paper, SVM is applied to signal detection in communication systems in the presence of channel noise in various environments in the form of Rayleigh fading, additive white Gaussian background noise (AWGN), and interference noise generalized as additive color Gaussian noise (ACGN). The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these advanced stochastic noise models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to conventional binary signaling optimal model-based detector driven by binary phase shift keying (BPSK) modulation. We show that the SVM performance is superior to that of conventional matched filter-, innovation filter-, and Wiener filter-driven detectors, even in the presence of random Doppler carrier deviation, especially for low SNR (signal-to-noise ratio) ranges. For large SNR, the performance of the SVM was similar to that of the classical detectors. However, the convergence between SVM and maximum likelihood detection occurred at a higher SNR as the noise environment became more hostile.

Keywords: Colour noise, Doppler shift, innovation filter, least square-support vector machine, matched filter, Rayleigh fading, Wiener filter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1813
1084 Analysis of Meteorological Drought in the Ruhr Basin by Using the Standardized Precipitation Index

Authors: Mosaad Khadr, Gerd Morgenschweis, Andreas Schlenkhoff

Abstract:

Drought is one of the most damaging climate-related hazards, it is generally considered as a prolonged absence of precipitation. This normal and recurring climate phenomenon had plagued civilization throughout history because of the negative impacts on economical, environmental and social sectors. Drought characteristics are thus recognized as important factors in water resources planning and management. The purpose of this study is to detect the changes in drought frequency, persistence and severity in the Ruhr river basin. The frequency of drought events was calculated using the Standardized Precipitation Index (SPI). Used data are daily precipitation records from seven meteorological stations covering the period 1961-2007. The main benefit of the application of this index is its versatility, only rainfall data is required to deliver five major dimensions of a drought : duration, intensity, severity, magnitude, and frequency. Furthermore, drought can be calculated in different time steps. In this study SPI was calculated for 1, 3, 6, 9, 12, and 24 months. Several drought events were detected in the covered period, these events contain mild, moderate and severe droughts. Also positive and negative trends in the SPI values were observed.

Keywords: Drought, Germany, Precipitation, Ruhr River, Standardized Precipitation Index (SPI), Trend Test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2317
1083 The Effect of the Hourly Compensation on the Unemployment Rate: Comparative Analysis of United States, Canada and the United Kingdom Using Panel Data Regression Analysis

Authors: Ashiquer Rahman, Hares Mohammad, Ummey Salma

Abstract:

A country’s hourly compensation and unemployment rates are two of its most crucial components. They are not merely statistics but they have profound effects on individual, families, country, and the economy. They are inversely related to one another. The increased hourly compensation in the manufacturing sector can have a favorable effect on job changing issues. Moreover, the relationship between hourly compensation and unemployment is complex and influenced by broader economic factors. In this paper, in order to determine the effect of hourly compensation on unemployment rate, we use the panel data regression models and evaluate the expected link between hourly compensation and unemployment rate. We estimate the fixed effects model (FEM), evaluate the error components model (ECM), and determine which model (the FEM or ECM) is better through pooling all 60 observations. We then analyze and review the data by comparing countries (United States, Canada and the United Kingdom) using panel data regression models. Finally, we provide result, analysis and a summary of this extensive research on how the hourly compensation affects unemployment rate. Additionally, this paper offers relevant and useful guideline for the government and academic community to use an econometrics and social approach for the hourly compensation on unemployment rate to eliminate the problem.

Keywords: Hourly compensation, unemployment rate, panel data regression models, dummy variables, random effects model, fixed effects model, the linear regression model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 70
1082 Effect of Equal Channel Angular Pressing Process on Impact Property of Pure Copper

Authors: F. Al-Mufadi, F. Djavanroodi

Abstract:

Ultrafine grained (UFG) and nanostructured (NS) materials have experienced a rapid development during the last decade and made profound impact on every field of materials science and engineering. The present work has been undertaken to develop ultrafine grained pure copper by severe plastic deformation method and to examine the impact property by different characterizing tools.

For this aim, equal channel angular pressing die with the channel angle, outer corner angle and channel diameter of 90°, 17° and 20mm had been designed and manufactured. Commercial pure copper billets were ECAPed up to four passes by route BC at the ambient temperature. The results indicated that there is a great improvement at the hardness measurement, yield strength and ultimate tensile strength after ECAP process. It is found that the magnitudes of HV reach 136HV from 52HV after the final pass. Also, about 285% and 125% enhancement at the YS and UTS values have been obtained after the fourth pass as compared to the as-received conditions, respectively. On the other hand, the elongation to failure and impact energy have been reduced by imposing ECAP process and pass numbers. It is needed to say that about 56% reduction in the impact energy have been attained for the samples as contrasted to annealed specimens. 

Keywords: SPD, ECAP, Pure Cu, Impact property.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2837
1081 Predicting Crack Initiation Due to Ratchetting in Rail Heads Using Critical Element Analysis

Authors: I. U. Wickramasinghe, D. J. Hargreaves, D. V. De Pellegrin

Abstract:

This paper presents a strategy to predict the lifetime of rails subjected to large rolling contact loads that induce ratchetting strains in the rail head. A critical element concept is used to calculate the number of loading cycles needed for crack initiation to occur in the rail head surface. In this technique the finite element method (FEM) is used to determine the maximum equivalent ratchetting strain per load cycle, which is calculated by combining longitudinal and shear stains in the critical element. This technique builds on a previously developed critical plane concept that has been used to calculate the number of cycles to crack initiation in rolling contact fatigue under ratchetting failure conditions. The critical element concept simplifies the analytical difficulties of critical plane analysis. Finite element analysis (FEA) is used to identify the critical element in the mesh, and then the strain values of the critical element are used to calculate the ratchetting rate analytically. Finally, a ratchetting criterion is used to calculate the number of cycles to crack initiation from the ratchetting rate calculated.

Keywords: Critical element analysis, finite element modeling (FEM), wheel/rail contact.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2933
1080 A Survey on Data-Centric and Data-Aware Techniques for Large Scale Infrastructures

Authors: Silvina Caíno-Lores, Jesús Carretero

Abstract:

Large scale computing infrastructures have been widely developed with the core objective of providing a suitable platform for high-performance and high-throughput computing. These systems are designed to support resource-intensive and complex applications, which can be found in many scientific and industrial areas. Currently, large scale data-intensive applications are hindered by the high latencies that result from the access to vastly distributed data. Recent works have suggested that improving data locality is key to move towards exascale infrastructures efficiently, as solutions to this problem aim to reduce the bandwidth consumed in data transfers, and the overheads that arise from them. There are several techniques that attempt to move computations closer to the data. In this survey we analyse the different mechanisms that have been proposed to provide data locality for large scale high-performance and high-throughput systems. This survey intends to assist scientific computing community in understanding the various technical aspects and strategies that have been reported in recent literature regarding data locality. As a result, we present an overview of locality-oriented techniques, which are grouped in four main categories: application development, task scheduling, in-memory computing and storage platforms. Finally, the authors include a discussion on future research lines and synergies among the former techniques.

Keywords: Co-scheduling, data-centric, data-intensive, data locality, in-memory storage, large scale.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1492
1079 A Numerical Study of Single-phase Forced Convective Heat Transfer in Tube in Tube Heat Exchangers

Authors: P. Mohajeri Khameneh, I. Mirzaie, N. Pourmahmoud, M. Rahimi, S. Majidyfar

Abstract:

Three dimensional simulations in tube in tube heat exchangers are investigated numerically in this study. In these simulations forced convective heat transfer and laminar flow of single-phase water are considered. In order to measure heat transfer parameters in these heat exchangers, FLUENT CFD Solver is used in this numerical method. For the purpose of creating geometry and exert boundary and initial conditions in the present model, finite volume method in Computational Fluid Dynamics is used in this study. In the present study, at each Z-location, variation of local temperatures, heat flux and Nusselt number at the whole tube is investigated in detail. Thereafter, averaged computational Nusselt number in this model is calculated. In addition, conceivable pressure drops have been obtained at each Z-location in this model. Then, pressure drop values in the present model are explored. Finally, all the numerical results for this kind of heat exchanger will be discussed precisely.

Keywords: Heat exchanger, Laminar flow, CFD, Nusseltnumber, Tube in tube, pressure drop.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2034
1078 Wave-Structure Interaction for Submerged Quarter-Circle Breakwaters of Different Radii - Reflection Characteristics

Authors: Arkal Vittal Hegde, L. Ravikiran

Abstract:

The paper presents the results of a series of experiments conducted on physical models of Quarter-circle breakwater (QBW) in a two dimensional monochromatic wave flume. The purpose of the experiments was to evaluate the reflection coefficient Kr of QBW models of different radii (R) for different submergence ratios (d/hc), where d is the depth of water and hc is the height of the breakwater crest from the sea bed. The radii of the breakwater models studied were 20cm, 22.5cm, 25cm, 27.5cm and submergence ratios used varied from 1.067 to 1.667. The wave climate off the Mangalore coast was used for arriving at the various model wave parameters. The incident wave heights (Hi) used in the flume varied from 3 to 18cm, and wave periods (T) ranged from 1.2 s to 2.2 s. The water depths (d) of 40cm, 45cm and 50cm were used in the experiments. The data collected was analyzed to compute variation of reflection coefficient Kr=Hr/Hi (where Hr=reflected wave height) with the wave steepness Hi/gT2 for various R/Hi (R=breakwater radius) values. It was found that the reflection coefficient increased as incident wave steepness increased. Also as wave height decreases reflection coefficient decreases and as structure radius R increased Kr decreased slightly.

Keywords: Incident wave steepness, Quarter-circle breakwater, Reflection coefficient, Submergence ratio.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1795
1077 Facility Location Selection using Preference Programming

Authors: C. Ardil

Abstract:

This paper presents preference programming technique based multiple criteria decision making analysis for selecting a facility location for a new organization or expansion of an existing facility which is of vital importance for a decision support system and strategic planning process. The implementation of decision support systems is considered crucial to sustain competitive advantage and profitability persistence in turbulent environment. As an effective strategic management and decision making is necessary, multiple criteria decision making analysis supports the decision makers to formulate and implement the right strategy. The investment cost associated with acquiring the property and facility construction makes the facility location selection problem a long-term strategic investment decision, which rationalize the best location selection which results in higher economic benefits through increased productivity and optimal distribution network. Selecting the proper facility location from a given set of alternatives is a difficult task, as many potential qualitative and quantitative multiple conflicting criteria are to be considered. This paper solves a facility location selection problem using preference programming, which is an effective multiple criteria decision making analysis tool applied to deal with complex decision problems in the operational research environment. The ranking results of preference programming are compared with WSM, TOPSIS and VIKOR methods.

Keywords: Facility Location Selection, Multiple Criteria Decision Making, Multiple Criteria Decision Making Analysis, Preference Programming, Location Selection, WSM, TOPSIS, VIKOR

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 541
1076 A Monte Carlo Method to Data Stream Analysis

Authors: Kittisak Kerdprasop, Nittaya Kerdprasop, Pairote Sattayatham

Abstract:

Data stream analysis is the process of computing various summaries and derived values from large amounts of data which are continuously generated at a rapid rate. The nature of a stream does not allow a revisit on each data element. Furthermore, data processing must be fast to produce timely analysis results. These requirements impose constraints on the design of the algorithms to balance correctness against timely responses. Several techniques have been proposed over the past few years to address these challenges. These techniques can be categorized as either dataoriented or task-oriented. The data-oriented approach analyzes a subset of data or a smaller transformed representation, whereas taskoriented scheme solves the problem directly via approximation techniques. We propose a hybrid approach to tackle the data stream analysis problem. The data stream has been both statistically transformed to a smaller size and computationally approximated its characteristics. We adopt a Monte Carlo method in the approximation step. The data reduction has been performed horizontally and vertically through our EMR sampling method. The proposed method is analyzed by a series of experiments. We apply our algorithm on clustering and classification tasks to evaluate the utility of our approach.

Keywords: Data Stream, Monte Carlo, Sampling, DensityEstimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1417
1075 Friction Calculation and Simulation of Column Electric Power Steering System

Authors: Seyed Hamid Mirmohammad Sadeghi, Raffaella Sesana, Daniela Maffiodo

Abstract:

This study presents a procedure for friction calculation of column electric power steering (C-EPS) system which affects handling and comfort in driving. The friction losses estimation is obtained from experimental tests and mathematical calculation. Parts in C-EPS mainly involved in friction losses are bearings and worm gear. In the theoretical approach, the gear geometry and Hertz law were employed to measure the normal load and the sliding velocity and contact areas from the worm gears driving conditions. The viscous friction generated in the worm gear was obtained with a theoretical approach and the result was applied to model the friction in the steering system. Finally, by viscous friction coefficient and Coulomb friction coefficient, values of friction in worm gear were calculated. According to the Bearing Company and the characteristics of each bearing, the friction torques due to load and due to speed were calculated. A MATLAB Simulink model for calculating the friction in bearings and worm gear in C-EPS were done and the total friction value was estimated.

Keywords: Friction, worm gear, column electric power steering system, Simulink, bearing, electric power steering, EPS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2243
1074 Development of Sustainable Farming Compartment with Treated Wastewater in Abu Dhabi

Authors: Jongwan Eun, Sam Helwany, Lakshyana K. C.

Abstract:

The United Arab Emirates (UAE) is significantly dependent on desalinated water and groundwater resource, which is expensive and highly energy intensive. Despite the scarce water resource, stagnates only 54% of the recycled water was reused in 2012, and due to the lack of infrastructure to reuse the recycled water, the portion is expected to decrease with growing water usage. In this study, an “Oasis” complex comprised of Sustainable Farming Compartments (SFC) was proposed for reusing treated wastewater. The wastewater is used to decrease the ambient temperature of the SFC via an evaporative cooler. The SFC prototype was designed, built, and tested in an environmentally controlled laboratory and field site to evaluate the feasibility and effectiveness of the SFC subjected to various climatic conditions in Abu Dhabi. Based on the experimental results, the temperature drop achieved in the SFC in the laboratory and field site were5 ̊C from 22 ̊C and 7- 15 ̊C (from 33-45 ̊C to average 28 ̊C at relative humidity < 50%), respectively. An energy simulation using TRNSYS was performed to extend and validate the results obtained from the experiment. The results from the energy simulation and experiments show statistically close agreement. The total power consumption of the SFC system was approximately three and a half times lower than that of an electrical air conditioner. Therefore, by using treated wastewater, the SFC has a promising prospect to solve Abu Dhabi’s ecological concern related to desertification and wind erosion.

Keywords: Ecological farming system, energy simulation, evaporative cooling system, treated wastewater, temperature, humidity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1315
1073 Pavement Roughness Prediction Systems: A Bump Integrator Approach

Authors: Manish Pal, Rumi Sutradhar

Abstract:

Pavement surface unevenness plays a pivotal role on roughness index of road which affects on riding comfort ability. Comfort ability refers to the degree of protection offered to vehicle occupants from uneven elements in the road surface. So, it is preferable to have a lower roughness index value for a better riding quality of road users. Roughness is generally defined as an expression of irregularities in the pavement surface which can be measured using different equipments like MERLIN, Bump integrator, Profilometer etc. Among them Bump Integrator is quite simple and less time consuming in case of long road sections. A case study is conducted on low volume roads in West District in Tripura to determine roughness index (RI) using Bump Integrator at the standard speed of 32 km/h. But it becomes too tough to maintain the requisite standard speed throughout the road section. The speed of Bump Integrator (BI) has to lower or higher in some distinctive situations. So, it becomes necessary to convert these roughness index values of other speeds to the standard speed of 32 km/h. This paper highlights on that roughness index conversional model. Using SPSS (Statistical Package of Social Sciences) software a generalized equation is derived among the RI value at standard speed of 32 km/h and RI value at other speed conditions.

Keywords: Bump Integrator, Pavement Distresses, Roughness Index, SPSS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6671
1072 Land Use/Land Cover Mapping Using Landsat 8 and Sentinel-2 in a Mediterranean Landscape

Authors: M. Vogiatzis, K. Perakis

Abstract:

Spatial-explicit and up-to-date land use/land cover information is fundamental for spatial planning, land management, sustainable development, and sound decision-making. In the last decade, many satellite-derived land cover products at different spatial, spectral, and temporal resolutions have been developed, such as the European Copernicus Land Cover product. However, more efficient and detailed information for land use/land cover is required at the regional or local scale. A typical Mediterranean basin with a complex landscape comprised of various forest types, crops, artificial surfaces, and wetlands was selected to test and develop our approach. In this study, we investigate the improvement of Copernicus Land Cover product (CLC2018) using Landsat 8 and Sentinel-2 pixel-based classification based on all available existing geospatial data (Forest Maps, LPIS, Natura2000 habitats, cadastral parcels, etc.). We examined and compared the performance of the Random Forest classifier for land use/land cover mapping. In total, 10 land use/land cover categories were recognized in Landsat 8 and 11 in Sentinel-2A. A comparison of the overall classification accuracies for 2018 shows that Landsat 8 classification accuracy was slightly higher than Sentinel-2A (82,99% vs. 80,30%). We concluded that the main land use/land cover types of CLC2018, even within a heterogeneous area, can be successfully mapped and updated according to CLC nomenclature. Future research should be oriented toward integrating spatiotemporal information from seasonal bands and spectral indexes in the classification process.

Keywords: land use/land cover, random forest, Landsat-8 OLI, Sentinel-2A MSI, Corine land cover

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 339
1071 Evaluation of Model and Performance of Fuel Cell Hybrid Electric Vehicle in Different Drive Cycles

Authors: Fathollah Ommi, Golnaz Pourabedin, Koros Nekofa

Abstract:

In recent years fuel cell vehicles are rapidly appearing all over the globe. In less than 10 years, fuel cell vehicles have gone from mere research novelties to operating prototypes and demonstration models. At the same time, government and industry in development countries have teamed up to invest billions of dollars in partnerships intended to commercialize fuel cell vehicles within the early years of the 21st century. The purpose of this study is evaluation of model and performance of fuel cell hybrid electric vehicle in different drive cycles. A fuel cell system model developed in this work is a semi-experimental model that allows users to use the theory and experimental relationships in a fuel cell system. The model can be used as part of a complex fuel cell vehicle model in advanced vehicle simulator (ADVISOR). This work reveals that the fuel consumption and energy efficiency vary in different drive cycles. Arising acceleration and speed in a drive cycle leads to Fuel consumption increase. In addition, energy losses in drive cycle relates to fuel cell system power request. Parasitic power in different parts of fuel cell system will increase when power request increases. Finally, most of energy losses in drive cycle occur in fuel cell system because of producing a lot of energy by fuel cell stack.

Keywords: Drive cycle, Energy efficiency, energy consumption, Fuel cell system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685
1070 Solving an Extended Resource Leveling Problem with Multiobjective Evolutionary Algorithms

Authors: Javier Roca, Etienne Pugnaghi, Gaëtan Libert

Abstract:

We introduce an extended resource leveling model that abstracts real life projects that consider specific work ranges for each resource. Contrary to traditional resource leveling problems this model considers scarce resources and multiple objectives: the minimization of the project makespan and the leveling of each resource usage over time. We formulate this model as a multiobjective optimization problem and we propose a multiobjective genetic algorithm-based solver to optimize it. This solver consists in a two-stage process: a main stage where we obtain non-dominated solutions for all the objectives, and a postprocessing stage where we seek to specifically improve the resource leveling of these solutions. We propose an intelligent encoding for the solver that allows including domain specific knowledge in the solving mechanism. The chosen encoding proves to be effective to solve leveling problems with scarce resources and multiple objectives. The outcome of the proposed solvers represent optimized trade-offs (alternatives) that can be later evaluated by a decision maker, this multi-solution approach represents an advantage over the traditional single solution approach. We compare the proposed solver with state-of-art resource leveling methods and we report competitive and performing results.

Keywords: Intelligent problem encoding, multiobjective decision making, evolutionary computing, RCPSP, resource leveling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4195
1069 Wavelet-Based Classification of Myocardial Ischemia, Arrhythmia, Congestive Heart Failure and Sleep Apnea

Authors: Santanu Chattopadhyay, Gautam Sarkar, Arabinda Das

Abstract:

This paper presents wavelet based classification of various heart diseases. Electrocardiogram signals of different heart patients have been studied. Statistical natures of electrocardiogram signals for different heart diseases have been compared with the statistical nature of electrocardiograms for normal persons. Under this study four different heart diseases have been considered as follows: Myocardial Ischemia (MI), Congestive Heart Failure (CHF), Arrhythmia and Sleep Apnea. Statistical nature of electrocardiograms for each case has been considered in terms of kurtosis values of two types of wavelet coefficients: approximate and detail. Nine wavelet decomposition levels have been considered in each case. Kurtosis corresponding to both approximate and detail coefficients has been considered for decomposition level one to decomposition level nine. Based on significant difference, few decomposition levels have been chosen and then used for classification.

Keywords: Arrhythmia, congestive heart failure, discrete wavelet transform, electrocardiogram, myocardial ischemia, sleep apnea.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 668
1068 Staling and Quality of Iranian Flat Bread Stored at Modified Atmosphere in Different Packaging

Authors: A. Hematian Sourki, F. Tabatabaei Yazdi, M. Ghiafeh Davoodi, S.A. Mortazavi, M. Karimi, S.H. Razavizadegan Jahromi, A. Pourfarzad

Abstract:

This study investigated the use of modified atmosphere packaging (MAP) and different packaging to extend the shelf life of Barbari flat bread. Three atmospheres including 70%CO2 and 30%N2, 50% CO2 and 50%N2 and a normal air as control were used. The bread samples were packaged in three type pouches. The shelf life was determined by appearance of mold and yeast (M +Y) in Barbari bread samples stored at 25 ± 1°C and 38 ± 2% relative humidity. The results showed that it is possible to prolong the shelf life of Barbari bread from four days to about 21 days by using modified atmosphere packaging with high carbon dioxide concentration and high-barrier laminated and vacuum bags packages. However, the hardness of samples kept in MAP increase significantly by increase of carbon dioxide concentration. The correlation coefficient (r) between headspace CO2 concentration and hardness was 0.997, 0.997 and 0.599 for A, B and C packaging respectively. High negative correlation coefficients were found between the crumb moisture and the hardness values in various packaging. There were significant negative correlation coefficients between sensory parameters and hardness of texture.

Keywords: modified atmosphere packaging, flat bread, Iranian bread, staling, correlation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2725
1067 Dynamic Traffic Simulation for Traffic Congestion Problem Using an Enhanced Algorithm

Authors: Wong Poh Lee, Mohd. Azam Osman, Abdullah Zawawi Talib, Ahmad Izani Md. Ismail

Abstract:

Traffic congestion has become a major problem in many countries. One of the main causes of traffic congestion is due to road merges. Vehicles tend to move slower when they reach the merging point. In this paper, an enhanced algorithm for traffic simulation based on the fluid-dynamic algorithm and kinematic wave theory is proposed. The enhanced algorithm is used to study traffic congestion at a road merge. This paper also describes the development of a dynamic traffic simulation tool which is used as a scenario planning and to forecast traffic congestion level in a certain time based on defined parameter values. The tool incorporates the enhanced algorithm as well as the two original algorithms. Output from the three above mentioned algorithms are measured in terms of traffic queue length, travel time and the total number of vehicles passing through the merging point. This paper also suggests an efficient way of reducing traffic congestion at a road merge by analyzing the traffic queue length and travel time.

Keywords: Dynamic, fluid-dynamic, kinematic wave theory, simulation, traffic congestion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3142