Search results for: Short Time Fourier Transform.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7626

Search results for: Short Time Fourier Transform.

4686 Duration Analysis of New Firms in the Banking Industry

Authors: Jesus Orbe, Vicente Nunez-Anton

Abstract:

This paper studies the duration or survival time of commercial banks active in the Moscovian three month Rouble deposits market, during the 1994-1997 period. The privatization process of the Russian commercial banking industry, after the 1988 banking reform, caused a massive entry of new banks followed by a period of high rates of exit. As a consequence, many firms went bankrupt without refunding their deposits. Therefore, both for the banks and for the banks- depositors, it is of interest to analyze which are the significant characteristics that motivate the exit or the closing of the bank. We propose a different methodology based on penalized weighted least squares which represents a very general, flexible and innovative approach for this type of analysis. The more relevant results are that smaller banks exit sooner, banks that enter the market in the last part of the study have shorter durations. As expected, the more experienced banks have a longer duration in the market. In addition, the mean survival time is lower for banks which offer extreme interest rates.

Keywords: Banking, censored, duration, Kaplan-Meier.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1533
4685 Multi-Criteria Decision Analysis in Planning of Asbestos-Containing Waste Management

Authors: E. Bruno, F. Lacarbonara, M. C. Placentino, D. Gramegna

Abstract:

Environmental decision making, particularly about hazardous waste management, is inherently exposed to a high potential conflict, principally because of the trade-off between sociopolitical, environmental, health and economic factors. The need to plan complex contexts has led to an increasing request for decision analytic techniques as support for the decision process. In this work, alternative systems of asbestos-containing waste management (ACW) in Puglia (Southern Italy) were explored by a multi-criteria decision analysis. In particular, through Analytic Hierarchy Process five alternatives management have been compared and ranked according to their performance and efficiency, taking into account environmental, health and socio-economic aspects. A separated valuation has been performed for different temporal scale. For short period results showed a narrow deviation between the disposal alternatives “mono-material landfill in public quarry" and “dedicate cells in existing landfill", with the best performance of the first one. While for long period “treatment plant to eliminate hazard from asbestos-containing waste" was prevalent, although high energy demand required to achieve the change of crystalline structure. A comparison with results from a participative approach in valuation process might be considered as future development of method application to ACW management.

Keywords: Multi-criteria decision analysis, Hazardous wastemanagement, Asbestos.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1846
4684 Analysis of Heuristic Based Hybrid Simulated Annealing Algorithm for Multiprocessor Task Scheduling

Authors: Supriya Arya, Sunita Dhingra

Abstract:

Multiprocessor task scheduling problem for dependent and independent tasks is computationally complex problem. Many methods are proposed to achieve optimal running time. As the multiprocessor task scheduling is NP hard in nature, therefore, many heuristics are proposed which have improved the makespan of the problem. But due to problem specific nature, the heuristic method which provide best results for one problem, might not provide good results for another problem. So, Simulated Annealing which is meta heuristic approach is considered. It can be applied on all types of problems. However, due to many runs, meta heuristic approach takes large computation time. Hence, the hybrid approach is proposed by combining the Duplication Scheduling Heuristic and Simulated Annealing (SA) and the makespan results of Simple Simulated Annealing and Hybrid approach are analyzed.

Keywords: Multiprocessor task scheduling Problem, Makespan, Duplication Scheduling Heuristic, Simulated Annealing, Hybrid Approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2200
4683 Optimization of Material Removal Rate in Electrical Discharge Machining Using Fuzzy Logic

Authors: Amit Kohli, Aashim Wadhwa, Tapan Virmani, Ujjwal Jain

Abstract:

The objective of present work is to stimulate the machining of material by electrical discharge machining (EDM) to give effect of input parameters like discharge current (Ip), pulse on time (Ton), pulse off time (Toff) which can bring about changes in the output parameter, i.e. material removal rate. Experimental data was gathered from die sinking EDM process using copper electrode and Medium Carbon Steel (AISI 1040) as work-piece. The rules of membership function (MF) and the degree of closeness to the optimum value of the MMR are within the upper and lower range of the process parameters. It was found that proposed fuzzy model is in close agreement with the experimental results. By Intelligent, model based design and control of EDM process parameters in this study will help to enable dramatically decreased product and process development cycle times.

Keywords: Electrical discharge Machining (EDM), Fuzzy Logic, Material removal rate (MRR), Membership functions (MF).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2726
4682 Developing a New Vibration Analysis Calculative Method for Esfahan Subway Train and Railways Design, Manufacturing, and Construction

Authors: Omid A. Zargar

Abstract:

The simulated mass and spring method evaluation for subway or railways construction and installation systems have a wide application in rail industries. This kind of design should be optimizing all related parameters to reduce the amount of vibration in cities, homelands, historical zones and other critical locations. Finite element method could help us a lot to analysis such applications with an excellent accuracy but always developing some simple, fast and user friendly evaluation method required in subway industrial applications. In addition, process parameter optimization extremely required in railway industries to achieve some optimal design of railways with maximum safety, reliability and performance. Furthermore, it is important to reduce vibrations and further related maintenance costs as well as possible. In this paper a simple but useful simulated mass and spring evaluation system developed for Esfahan subway construction. Besides, some of related recent patent and innovations in rail world industries like Suspension mass tuned vibration reducer, short sleeper vibration attenuation fastener and Airtight track vibration-noise reducing fastener discussed in details.

Keywords: Subway construction engineering, natural frequency, operation frequency, vibration analysis, polyurethane layer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2332
4681 Scheduling Multiple Workflow Using De-De Dodging Algorithm and PBD Algorithm in Cloud: Detailed Study

Authors: B. Arun Kumar, T. Ravichandran

Abstract:

Workflow scheduling is an important part of cloud computing and based on different criteria it decides cost, execution time, and performances. A cloud workflow system is a platform service facilitating automation of distributed applications based on new cloud infrastructure. An aspect which differentiates cloud workflow system from others is market-oriented business model, an innovation which challenges conventional workflow scheduling strategies. Time and Cost optimization algorithm for scheduling Hybrid Clouds (TCHC) algorithm decides which resource should be chartered from public providers is combined with a new De-De algorithm considering that every instance of single and multiple workflows work without deadlocks. To offset this, two new concepts - De-De Dodging Algorithm and Priority Based Decisive Algorithm - combine with conventional deadlock avoidance issues by proposing one algorithm that maximizes active (not just allocated) resource use and reduces Makespan.

Keywords: Workflow Scheduling, cloud workflow, TCHC algorithm, De-De Dodging Algorithm, Priority Based Decisive Algorithm (PBD), Makespan.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2774
4680 A Pole Radius Varying Notch Filter with Transient Suppression for Electrocardiogram

Authors: Ramesh Rajagopalan, Adam Dahlstrom

Abstract:

Noise removal techniques play a vital role in the performance of electrocardiographic (ECG) signal processing systems. ECG signals can be corrupted by various kinds of noise such as baseline wander noise, electromyographic interference, and powerline interference. One of the significant challenges in ECG signal processing is the degradation caused by additive 50 or 60 Hz powerline interference. This work investigates the removal of power line interference and suppression of transient response for filtering noise corrupted ECG signals. We demonstrate the effectiveness of infinite impulse response (IIR) notch filter with time varying pole radius for improving the transient behavior. The temporary change in the pole radius of the filter diminishes the transient behavior. Simulation results show that the proposed IIR filter with time varying pole radius outperforms traditional IIR notch filters in terms of mean square error and transient suppression.

Keywords: Notch filter, ECG, transient, pole radius.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3174
4679 Prediction of Time to Crack Reinforced Concrete by Chloride Induced Corrosion

Authors: Anuruddha Jayasuriya, Thanakorn Pheeraphan

Abstract:

In this paper, a review of different mathematical models which can be used as prediction tools to assess the time to crack reinforced concrete (RC) due to corrosion is investigated. This investigation leads to an experimental study to validate a selected prediction model. Most of these mathematical models depend upon the mechanical behaviors, chemical behaviors, electrochemical behaviors or geometric aspects of the RC members during a corrosion process. The experimental program is designed to verify the accuracy of a well-selected mathematical model from a rigorous literature study. Fundamentally, the experimental program exemplifies both one-dimensional chloride diffusion using RC squared slab elements of 500 mm by 500 mm and two-dimensional chloride diffusion using RC squared column elements of 225 mm by 225 mm by 500 mm. Each set consists of three water-to-cement ratios (w/c); 0.4, 0.5, 0.6 and two cover depths; 25 mm and 50 mm. 12 mm bars are used for column elements and 16 mm bars are used for slab elements. All the samples are subjected to accelerated chloride corrosion in a chloride bath of 5% (w/w) sodium chloride (NaCl) solution. Based on a pre-screening of different models, it is clear that the well-selected mathematical model had included mechanical properties, chemical and electrochemical properties, nature of corrosion whether it is accelerated or natural, and the amount of porous area that rust products can accommodate before exerting expansive pressure on the surrounding concrete. The experimental results have shown that the selected model for both one-dimensional and two-dimensional chloride diffusion had ±20% and ±10% respective accuracies compared to the experimental output. The half-cell potential readings are also used to see the corrosion probability, and experimental results have shown that the mass loss is proportional to the negative half-cell potential readings that are obtained. Additionally, a statistical analysis is carried out in order to determine the most influential factor that affects the time to corrode the reinforcement in the concrete due to chloride diffusion. The factors considered for this analysis are w/c, bar diameter, and cover depth. The analysis is accomplished by using Minitab statistical software, and it showed that cover depth is the significant effect on the time to crack the concrete from chloride induced corrosion than other factors considered. Thus, the time predictions can be illustrated through the selected mathematical model as it covers a wide range of factors affecting the corrosion process, and it can be used to predetermine the durability concern of RC structures that are vulnerable to chloride exposure. And eventually, it is further concluded that cover thickness plays a vital role in durability in terms of chloride diffusion.

Keywords: Accelerated corrosion, chloride diffusion, corrosion cracks, passivation layer, reinforcement corrosion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 870
4678 Application of Sensory Thermography as Measuring Method to Study Median Nerve Temperatures

Authors: Javier Ordorica Villalvazo, Claudia Camargo Wilson, Jesus Everardo Olguin Tiznado

Abstract:

This paper presents an experimental case using sensory thermography to describe temperatures behavior on median nerve once an activity of repetitive motion was done. Thermography is a noninvasive technique without biological hazard and not harm at all times and has been applied in many experiments to seek for temperature patterns that help to understand diseases like cancer and cumulative trauma disorders (CTD’s). An infrared sensory thermography technology was developed to execute this study. Three women in good shape were selected for the repetitive motion tests for 4 days, two right-handed women and 1 left handed woman, two sensory thermographers were put on both median nerve wrists to get measures. The evaluation time was of 3 hours 30 minutes in a controlled temperature, 20 minutes of stabilization time at the beginning and end of the operation. Temperatures distributions are statistically evaluated and showed similar temperature patterns behavior.

Keywords: Median nerve, temperature, sensory thermography, wrists, CTD’s.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1466
4677 Opportunistic Routing with Secure Coded Wireless Multicast Using MAS Approach

Authors: E. Golden Julie, S. Tamil Selvi, Y. Harold Robinson

Abstract:

Many Wireless Sensor Network (WSN) applications necessitate secure multicast services for the purpose of broadcasting delay sensitive data like video files and live telecast at fixed time-slot. This work provides a novel method to deal with end-to-end delay and drop rate of packets. Opportunistic Routing chooses a link based on the maximum probability of packet delivery ratio. Null Key Generation helps in authenticating packets to the receiver. Markov Decision Process based Adaptive Scheduling algorithm determines the time slot for packet transmission. Both theoretical analysis and simulation results show that the proposed protocol ensures better performance in terms of packet delivery ratio, average end-to-end delay and normalized routing overhead.

Keywords: Delay-sensitive data, Markovian Decision Process based Adaptive Scheduling, Opportunistic Routing, Digital Signature authentication.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1938
4676 Replicating Data Objects in Large-scale Distributed Computing Systems using Extended Vickrey Auction

Authors: Samee Ullah Khan, Ishfaq Ahmad

Abstract:

This paper proposes a novel game theoretical technique to address the problem of data object replication in largescale distributed computing systems. The proposed technique draws inspiration from computational economic theory and employs the extended Vickrey auction. Specifically, players in a non-cooperative environment compete for server-side scarce memory space to replicate data objects so as to minimize the total network object transfer cost, while maintaining object concurrency. Optimization of such a cost in turn leads to load balancing, fault-tolerance and reduced user access time. The method is experimentally evaluated against four well-known techniques from the literature: branch and bound, greedy, bin-packing and genetic algorithms. The experimental results reveal that the proposed approach outperforms the four techniques in both the execution time and solution quality.

Keywords: Auctions, data replication, pricing, static allocation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448
4675 Study on Disaster Prevention Plan for an Electronic Industry in Thailand

Authors: S. Pullteap, M. Pathomsuriyaporn

Abstract:

In this article, a study of employee’s opinion to the factors that affect to the flood preventive and the corrective action plan in an electronic industry at the Sharp Manufacturing (Thailand) Co., Ltd. has been investigated. The surveys data of 175 workers and supervisors have, however, been selected for data analysis. The results is shown that the employees emphasize about the needs in a subsidy at the time of disaster at high levels of 77.8%, as the plan focusing on flood prevention of the rehabilitation equipment is valued at the intermediate level, which is 79.8%. Demonstration of the hypothesis has found that the different education levels has thus been affected to the needs factor at the flood disaster time. Moreover, most respondents give priority to flood disaster risk management factor. Consequently, we found that the flood prevention plan is valued at high level, especially on information monitoring, which is 93.4% for the supervisor item. The respondents largely assume that the flood will have impacts on the industry, up to 80%, thus to focus on flood management plans is enormous.

Keywords: Flood prevention plan, flood event, electronic industrial plant, disaster, risk management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1826
4674 High Performance in Parallel Data Integration: An Empirical Evaluation of the Ratio Between Processing Time and Number of Physical Nodes

Authors: Caspar von Seckendorff, Eldar Sultanow

Abstract:

Many studies have shown that parallelization decreases efficiency [1], [2]. There are many reasons for these decrements. This paper investigates those which appear in the context of parallel data integration. Integration processes generally cannot be allocated to packages of identical size (i. e. tasks of identical complexity). The reason for this is unknown heterogeneous input data which result in variable task lengths. Process delay is defined by the slowest processing node. It leads to a detrimental effect on the total processing time. With a real world example, this study will show that while process delay does initially increase with the introduction of more nodes it ultimately decreases again after a certain point. The example will make use of the cloud computing platform Hadoop and be run inside Amazon-s EC2 compute cloud. A stochastic model will be set up which can explain this effect.

Keywords: Process delay, speedup, efficiency, parallel computing, data integration, E-Commerce, Amazon Elastic Compute Cloud (EC2), Hadoop, Nutch.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1612
4673 Faster Pedestrian Recognition Using Deformable Part Models

Authors: Alessandro Preziosi, Antonio Prioletti, Luca Castangia

Abstract:

Deformable part models achieve high precision in pedestrian recognition, but all publicly available implementations are too slow for real-time applications. We implemented a deformable part model algorithm fast enough for real-time use by exploiting information about the camera position and orientation. This implementation is both faster and more precise than alternative DPM implementations. These results are obtained by computing convolutions in the frequency domain and using lookup tables to speed up feature computation. This approach is almost an order of magnitude faster than the reference DPM implementation, with no loss in precision. Knowing the position of the camera with respect to horizon it is also possible prune many hypotheses based on their size and location. The range of acceptable sizes and positions is set by looking at the statistical distribution of bounding boxes in labelled images. With this approach it is not needed to compute the entire feature pyramid: for example higher resolution features are only needed near the horizon. This results in an increase in mean average precision of 5% and an increase in speed by a factor of two. Furthermore, to reduce misdetections involving small pedestrians near the horizon, input images are supersampled near the horizon. Supersampling the image at 1.5 times the original scale, results in an increase in precision of about 4%. The implementation was tested against the public KITTI dataset, obtaining an 8% improvement in mean average precision over the best performing DPM-based method. By allowing for a small loss in precision computational time can be easily brought down to our target of 100ms per image, reaching a solution that is faster and still more precise than all publicly available DPM implementations.

Keywords: Autonomous vehicles, deformable part model, dpm, pedestrian recognition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371
4672 Volatile Profile of Monofloral Honeys Produced by Stingless Bees from the Brazilian Semiarid Region

Authors: Ana Caroliny Vieira da Costa, Marta Suely Madruga

Abstract:

In Brazil, there is a diverse fauna of social bees, known by Meliponinae or native stingless bees. These bees are important for providing a differentiated product, especially regarding unique sweetness, flavor, and aroma. However, information about the volatile fraction in honey produced by stingless native bees is still lacking. The aim of this work was to characterize the volatile compound profile of monofloral honey produced by jandaíra bees (Melipona subnitida Ducke) which used chanana (Turnera ulmifolia L.), malícia (Mimosa quadrivalvis) and algaroba (Prosopis juliflora (Sw.) DC) as their floral sources; and by uruçu bees (Melipona scutellaris Latrelle), which used chanana (Turnera ulmifolia L.), malícia (Mimosa quadrivalvis) and angico (Anadenanthera colubrina) as their floral sources. The volatiles were extracted using HS-SPME-GC-MS technique. The condition for the extraction was: equilibration time of 15 minutes, extraction time of 45 min and extraction temperature of 45°C. Through the results obtained, it was observed that the floral source had a strong influence on the aroma profile of the honey under evaluation, since the chemical profiles were marked primarily by the classes of terpenes, norisoprenoids, and benzene derivatives. Furthermore, the results obtained suggest the existence of differentiator compounds and potential markers for the botanical sources evaluated, such as linalool, D-sylvestrene, rose oxide and benzenethanol. These reports represent a valuable contribution to certifying the authenticity of those honey and provides for the first time, information intended for the construction of chemical knowledge of the aroma and flavor that characterize these honey produced in Brazil.

Keywords: Aroma, honey, semiarid, stingless, volatiles.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1418
4671 Hydrodynamic Modeling of a Surface Water Treatment Pilot Plant

Authors: C.-M. Militaru, A. Pǎcalǎ, I. Vlaicu, K. Bodor, G.-A. Dumitrel, T. Todinca

Abstract:

A mathematical model for the hydrodynamics of a surface water treatment pilot plant was developed and validated by the determination of the residence time distribution (RTD) for the main equipments of the unit. The well known models of ideal/real mixing, ideal displacement (plug flow) and (one-dimensional axial) dispersion model were combined in order to identify the structure that gives the best fitting of the experimental data for each equipment of the pilot plant. RTD experimental results have shown that pilot plant hydrodynamics can be quite well approximated by a combination of simple mathematical models, structure which is suitable for engineering applications. Validated hydrodynamic models will be further used in the evaluation and selection of the most suitable coagulation-flocculation reagents, optimum operating conditions (injection point, reaction times, etc.), in order to improve the quality of the drinking water.

Keywords: drinking water, hydrodynamic modeling, pilot plant, residence time distribution, surface water.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1650
4670 Optimization and Kinetic Study of Gaharu Oil Extraction

Authors: Muhammad Hazwan H., Azlina M.F., Hasfalina C.M., Zurina Z.A., Hishamuddin J

Abstract:

Gaharu that produced by Aquilaria spp. is classified as one of the most valuable forest products traded internationally as it is very resinous, fragrant and highly valuable heartwood. Gaharu has been widely used in aromatheraphy, medicine, perfume and religious practices. This work aimed to determine the factors affecting solid liquid extraction of gaharu oil using hexane as solvent under experimental condition. The kinetics of extraction was assumed and verified based on a second-order mechanism. The effect of three main factors, which were temperature, reaction time and solvent to solid ratio were investigated to achieve maximum oil yield. The optimum condition were found at temperature 65°C, 9 hours reaction time and solvent to solid ratio of 12:1 with 14.5% oil yield. The kinetics experimental data agrees and well fitted with the second order extraction model. The initial extraction rate (h) was 0.0115 gmL-1min-1; the extraction capacity (Cs) was 1.282gmL-1; the second order extraction constant (k) was 0.007 mLg-1min-1 and coefficient of determination, R2 was 0.945.

Keywords: Gaharu, solid liquid extraction, optimization, kinetics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3246
4669 The New Semi-Experimental Method for Simulation of Turbine Flow Meters Rotation in the Transitional Flow

Authors: J. Tonkonogij, A. Pedišius, A. Stankevičius

Abstract:

The new semi-experimental method for simulation of the turbine flow meters rotation in the transitional flow has been developed. The method is based on the experimentally established exponential low of changing of dimensionless relative turbine gas meter rotation frequency and meter inertia time constant. For experimental evaluation of the meter time constant special facility has been developed. The facility ensures instant switching of turbine meter under test from one channel to the other channel with different flow rate and measuring the meter response. The developed method can be used for evaluation and predication of the turbine meters response and dynamic error in the transitional flow with any arbitrary law of flow rate changing. The examples of the method application are presented.

Keywords: Dynamic error, pulsing flow, numerical simulation, response, turbine gas meters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2170
4668 Automatically Driven Vector for Guidewire Segmentation in 2D and Biplane Fluoroscopy

Authors: Simon Lessard, Pascal Bigras, Caroline Lau, Daniel Roy, Gilles Soulez, Jacques A. de Guise

Abstract:

The segmentation of endovascular tools in fluoroscopy images can be accurately performed automatically or by minimum user intervention, using known modern techniques. It has been proven in literature, but no clinical implementation exists so far because the computational time requirements of such technology have not yet been met. A classical segmentation scheme is composed of edge enhancement filtering, line detection, and segmentation. A new method is presented that consists of a vector that propagates in the image to track an edge as it advances. The filtering is performed progressively in the projected path of the vector, whose orientation allows for oriented edge detection, and a minimal image area is globally filtered. Such an algorithm is rapidly computed and can be implemented in real-time applications. It was tested on medical fluoroscopy images from an endovascular cerebral intervention. Ex- periments showed that the 2D tracking was limited to guidewires without intersection crosspoints, while the 3D implementation was able to cope with such planar difficulties.

Keywords: Edge detection, Line Enhancement, Segmentation, Fluoroscopy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1711
4667 Separation Characteristics of Dissolved Gases from Water Concurrently Variable Mixed with Exhalations for the Hollow Fiber Membrane

Authors: Pil Woo Heo

Abstract:

Water contains dissolved oxygen that a fish needs to breathe. It is important to increase the amounts of separation of dissolved oxygen from water for diverse applications using the separation system. In this paper, a separation system of dissolved gases from water concurrently variable mixed with the exhalations using a compressor is proposed. This system takes use of exhalations to increase the amounts of separation of dissolved oxygen from water. A compressor with variable off-time and on-time is used to control the exhalations mixed with inlet water. Exhalations contain some portion of carbon dioxide, oxygen, and nitrogen. Separation of dissolved gases containing dissolved oxygen is enhanced by using exhalations. The amounts of separation and the compositions of carbon dioxide and oxygen are measured. Higher amounts of separation can make the size of the separation device smaller, and then, application areas are diversified.

Keywords: Concurrently, variable mixed, exhalations, separation, hollow fiber.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1092
4666 GSA-Based Design of Dual Proportional Integral Load Frequency Controllers for Nonlinear Hydrothermal Power System

Authors: M. Elsisi, M. Soliman, M. A. S. Aboelela, W. Mansour

Abstract:

This paper considers the design of Dual Proportional- Integral (DPI) Load Frequency Control (LFC), using gravitational search algorithm (GSA). The design is carried out for nonlinear hydrothermal power system where generation rate constraint (GRC) and governor dead band are considered. Furthermore, time delays imposed by governor-turbine, thermodynamic process, and communication channels are investigated. GSA is utilized to search for optimal controller parameters by minimizing a time-domain based objective function. GSA-based DPI has been compared to Ziegler- Nichols based PI, and Genetic Algorithm (GA) based PI controllers in order to demonstrate the superior efficiency of the proposed design. Simulation results are carried for a wide range of operating conditions and system parameters variations.

Keywords: Gravitational Search Algorithm (GSA), Load Frequency Control (LFC), Dual Proportional-Integral (DPI) controller.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1962
4665 PoPCoRN: A Power-Aware Periodic Surveillance Scheme in Convex Region using Wireless Mobile Sensor Networks

Authors: A. K. Prajapati

Abstract:

In this paper, the periodic surveillance scheme has been proposed for any convex region using mobile wireless sensor nodes. A sensor network typically consists of fixed number of sensor nodes which report the measurements of sensed data such as temperature, pressure, humidity, etc., of its immediate proximity (the area within its sensing range). For the purpose of sensing an area of interest, there are adequate number of fixed sensor nodes required to cover the entire region of interest. It implies that the number of fixed sensor nodes required to cover a given area will depend on the sensing range of the sensor as well as deployment strategies employed. It is assumed that the sensors to be mobile within the region of surveillance, can be mounted on moving bodies like robots or vehicle. Therefore, in our scheme, the surveillance time period determines the number of sensor nodes required to be deployed in the region of interest. The proposed scheme comprises of three algorithms namely: Hexagonalization, Clustering, and Scheduling, The first algorithm partitions the coverage area into fixed sized hexagons that approximate the sensing range (cell) of individual sensor node. The clustering algorithm groups the cells into clusters, each of which will be covered by a single sensor node. The later determines a schedule for each sensor to serve its respective cluster. Each sensor node traverses all the cells belonging to the cluster assigned to it by oscillating between the first and the last cell for the duration of its life time. Simulation results show that our scheme provides full coverage within a given period of time using few sensors with minimum movement, less power consumption, and relatively less infrastructure cost.

Keywords: Sensor Network, Graph Theory, MSN, Communication.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1440
4664 Unreliable Production Lines with Simultaneously Unbalanced Operation Time Means, Breakdown, and Repair Rates

Authors: S. Shaaban, T. McNamara, S. Hudson

Abstract:

This paper investigates the benefits of deliberately unbalancing both operation time means (MTs) and unreliability (failure and repair rates) for non-automated production lines. The lines were simulated with various line lengths, buffer capacities, degrees of imbalance and patterns of MT and unreliability imbalance. Data on two performance measures, namely throughput (TR) and average buffer level (ABL) were gathered, analyzed and compared to a balanced line counterpart. A number of conclusions were made with respect to the ranking of configurations, as well as to the relationships among the independent design parameters and the dependent variables. It was found that the best configurations are a balanced line arrangement and a monotone decreasing MT order, coupled with either a decreasing or a bowl unreliability configuration, with the first generally resulting in a reduced TR and the second leading to a lower ABL than those of a balanced line.

Keywords: Average buffer level, throughput, unbalanced failure and repair rates, unequal mean operation times, unreliable production lines.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2221
4663 Anomaly Detection in a Data Center with a Reconstruction Method Using a Multi-Autoencoders Model

Authors: Victor Breux, Jérôme Boutet, Alain Goret, Viviane Cattin

Abstract:

Early detection of anomalies in data centers is important to reduce downtimes and the costs of periodic maintenance. However, there is little research on this topic and even fewer on the fusion of sensor data for the detection of abnormal events. The goal of this paper is to propose a method for anomaly detection in data centers by combining sensor data (temperature, humidity, power) and deep learning models. The model described in the paper uses one autoencoder per sensor to reconstruct the inputs. The auto-encoders contain Long-Short Term Memory (LSTM) layers and are trained using the normal samples of the relevant sensors selected by correlation analysis. The difference signal between the input and its reconstruction is then used to classify the samples using feature extraction and a random forest classifier. The data measured by the sensors of a data center between January 2019 and May 2020 are used to train the model, while the data between June 2020 and May 2021 are used to assess it. Performances of the model are assessed a posteriori through F1-score by comparing detected anomalies with the data center’s history. The proposed model outperforms the state-of-the-art reconstruction method, which uses only one autoencoder taking multivariate sequences and detects an anomaly with a threshold on the reconstruction error, with an F1-score of 83.60% compared to 24.16%.

Keywords: Anomaly detection, autoencoder, data centers, deep learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 703
4662 Machine Learning-Enabled Classification of Climbing Using Small Data

Authors: Nicholas Milburn, Yu Liang, Dalei Wu

Abstract:

Athlete performance scoring within the climbing domain presents interesting challenges as the sport does not have an objective way to assign skill. Assessing skill levels within any sport is valuable as it can be used to mark progress while training, and it can help an athlete choose appropriate climbs to attempt. Machine learning-based methods are popular for complex problems like this. The dataset available was composed of dynamic force data recorded during climbing; however, this dataset came with challenges such as data scarcity, imbalance, and it was temporally heterogeneous. Investigated solutions to these challenges include data augmentation, temporal normalization, conversion of time series to the spectral domain, and cross validation strategies. The investigated solutions to the classification problem included light weight machine classifiers KNN and SVM as well as the deep learning with CNN. The best performing model had an 80% accuracy. In conclusion, there seems to be enough information within climbing force data to accurately categorize climbers by skill.

Keywords: Classification, climbing, data imbalance, data scarcity, machine learning, time sequence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 528
4661 Green Synthesis of Butyl Acetate, A Pineapple Flavour via Lipase-Catalyzed Reaction

Authors: S. Mat Radzi, W.A.F. Mustafa, S.S Othman, H.M. Noor

Abstract:

Nowadays, butyl acetate, a pineapple flavor has been applied widely in food, beverage, cosmetic and pharmaceutical industries. In this study, Butyl acetate, a flavor ester was successfully synthesized via green synthesis of enzymatic reaction route. Commercial immobilized lipase from Rhizomucor miehei (Lipozyme RMIM) was used as biocatalyst in the esterification reaction between acetic acid and butanol. Various reaction parameters such as reaction time (RT), temperature (T) and amount of enzyme (E) were chosen to optimize the reaction synthesis in solvent-free system. The optimum condition to produce butyl acetate was at reaction time (RT), 18 hours; temperature (T), 37°C and amount of enzyme, 25 % (w/w of total substrate). Analysis of yield showed that at optimum condition, >78 % of butyl acetate was produced. The product was confirmed as butyl acetate from FTIR analysis whereby the presence of an ester group was observed at wavenumber of 1742 cm-1.

Keywords: Butyl acetate, immobilized enyzme, esterification, flavor ester, green synthesis

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3735
4660 Artificial Neural Network Model for a Low Cost Failure Sensor: Performance Assessment in Pipeline Distribution

Authors: Asar Khan, Peter D. Widdop, Andrew J. Day, Aliaster S. Wood, Steve, R. Mounce, John Machell

Abstract:

This paper describes an automated event detection and location system for water distribution pipelines which is based upon low-cost sensor technology and signature analysis by an Artificial Neural Network (ANN). The development of a low cost failure sensor which measures the opacity or cloudiness of the local water flow has been designed, developed and validated, and an ANN based system is then described which uses time series data produced by sensors to construct an empirical model for time series prediction and classification of events. These two components have been installed, tested and verified in an experimental site in a UK water distribution system. Verification of the system has been achieved from a series of simulated burst trials which have provided real data sets. It is concluded that the system has potential in water distribution network management.

Keywords: Detection, leakage, neural networks, sensors, water distribution networks

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1726
4659 Computational Prediction of Complicated Atmospheric Motion for Spinning or non- Spinning Projectiles

Authors: Dimitrios N. Gkritzapis, Elias E. Panagiotopoulos, Dionissios P. Margaris, Dimitrios G. Papanikas

Abstract:

A full six degrees of freedom (6-DOF) flight dynamics model is proposed for the accurate prediction of short and long-range trajectories of high spin and fin-stabilized projectiles via atmospheric flight to final impact point. The projectiles is assumed to be both rigid (non-flexible), and rotationally symmetric about its spin axis launched at low and high pitch angles. The mathematical model is based on the full equations of motion set up in the no-roll body reference frame and is integrated numerically from given initial conditions at the firing site. The projectiles maneuvering motion depends on the most significant force and moment variations, in addition to wind and gravity. The computational flight analysis takes into consideration the Mach number and total angle of attack effects by means of the variable aerodynamic coefficients. For the purposes of the present work, linear interpolation has been applied from the tabulated database of McCoy-s book. The developed computational method gives satisfactory agreement with published data of verified experiments and computational codes on atmospheric projectile trajectory analysis for various initial firing flight conditions.

Keywords: Constant-Variable aerodynamic coefficients, low and high pitch angles, wind.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2402
4658 Integrated Modeling Approach for Energy Planning and Climate Change Mitigation Assessment in the State of Florida

Authors: Kuntal Thakkar, Chaouki Ghenai, Ahmed Hachicha

Abstract:

An integrated modeling approach was used in this study for energy planning and climate change mitigation assessment. The main objective of this study was to develop various green-house gas (GHG) mitigations scenarios in the energy demand and supply sectors for the state of Florida. The Long range energy alternative planning (LEAP) model was used in this study to examine the energy alternative and GHG emissions reduction scenarios for short and long term (2010-2050). One of the energy analysis and GHG mitigation scenarios was developed by taking into account the available renewable energy resources potential for power generation in the state of Florida. This will help to compare and analyze the GHG reduction measure against “Business As Usual” and ‘State of Florida Policy” scenarios. Two master scenarios: “Electrification” and “Energy efficiency and Lifestyle” were developed through combination of various mitigation scenarios: technological changes and energy efficiency and conservation. The results show a net reduction of the energy demand and GHG emissions by adopting these two energy scenarios compared to the business as usual.

Keywords: Integrated modeling, energy planning, climate change mitigation assessment, greenhouse gas emissions, renewable energy, energy efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1763
4657 Hexavalent Chromium Removal from Aqueous Solutions by Adsorption onto Synthetic Nano Size ZeroValent Iron (nZVI)

Authors: A.R. Rahmani, M.T. Samadi, R. Noroozi

Abstract:

The present work was conducted for the synthesis of nano size zerovalent iron (nZVI) and hexavalent chromium (Cr(VI)) removal as a highly toxic pollutant by using this nanoparticles. Batch experiments were performed to investigate the effects of Cr(VI), nZVI concentration, pH of solution and contact time variation on the removal efficiency of Cr(VI). nZVI was synthesized by reduction of ferric chloride using sodium borohydrid. SEM and XRD examinations applied for determination of particle size and characterization of produced nanoparticles. The results showed that the removal efficiency decreased with Cr(VI) concentration and pH of solution and increased with adsorbent dosage and contact time. The Langmuir and Freundlich isotherm models were used for the adsorption equilibrium data and the Langmuir isotherm model was well fitted. Nanoparticle ZVI presented an outstanding ability to remove Cr(VI) due to high surface area, low particle size and high inherent activity.

Keywords: Adsorption, aqueous solution, Chromium, nZVI, removal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2540