Search results for: flood prevention scheme
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1524

Search results for: flood prevention scheme

294 Tracing Quality Cost in a Luggage Manufacturing Industry

Authors: S. B. Jaju, R. R. Lakhe

Abstract:

Quality costs are the costs associated with preventing, finding, and correcting defective work. Since the main language of corporate management is money, quality-related costs act as means of communication between the staff of quality engineering departments and the company managers. The objective of quality engineering is to minimize the total quality cost across the life of product. Quality costs provide a benchmark against which improvement can be measured over time. It provides a rupee-based report on quality improvement efforts. It is an effective tool to identify, prioritize and select quality improvement projects. After reviewing through the literature it was noticed that a simplified methodology for data collection of quality cost in a manufacturing industry was required. The quantified standard methodology is proposed for collecting data of various elements of quality cost categories for manufacturing industry. Also in the light of research carried out so far, it is felt necessary to standardise cost elements in each of the prevention, appraisal, internal failure and external failure costs. . Here an attempt is made to standardise the various cost elements applicable to manufacturing industry and data is collected by using the proposed quantified methodology. This paper discusses the case study carried in luggage manufacturing industry.

Keywords: Quality Costs, PAF model, quantified methodology, Case study.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2254
293 Bandwidth Efficient Diversity Scheme Using STTC Concatenated With STBC: MIMO Systems

Authors: Sameru Sharma, Sanjay Sharma, Derick Engles

Abstract:

Multiple-input multiple-output (MIMO) systems are widely in use to improve quality, reliability of wireless transmission and increase the spectral efficiency. However in MIMO systems, multiple copies of data are received after experiencing various channel effects. The limitations on account of complexity due to number of antennas in case of conventional decoding techniques have been looked into. Accordingly we propose a modified sphere decoder (MSD-1) algorithm with lower complexity and give rise to system with high spectral efficiency. With the aim to increase signal diversity we apply rotated quadrature amplitude modulation (QAM) constellation in multi dimensional space. Finally, we propose a new architecture involving space time trellis code (STTC) concatenated with space time block code (STBC) using MSD-1 at the receiver for improving system performance. The system gains have been verified with channel state information (CSI) errors.

Keywords: Channel State Information , Diversity, Multi-Antenna, Rotated Constellation, Space Time Codes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1666
292 Automated Process Quality Monitoring with Prediction of Fault Condition Using Measurement Data

Authors: Hyun-Woo Cho

Abstract:

Detection of incipient abnormal events is important to improve safety and reliability of machine operations and reduce losses caused by failures. Improper set-ups or aligning of parts often leads to severe problems in many machines. The construction of prediction models for predicting faulty conditions is quite essential in making decisions on when to perform machine maintenance. This paper presents a multivariate calibration monitoring approach based on the statistical analysis of machine measurement data. The calibration model is used to predict two faulty conditions from historical reference data. This approach utilizes genetic algorithms (GA) based variable selection, and we evaluate the predictive performance of several prediction methods using real data. The results shows that the calibration model based on supervised probabilistic principal component analysis (SPPCA) yielded best performance in this work. By adopting a proper variable selection scheme in calibration models, the prediction performance can be improved by excluding non-informative variables from their model building steps.

Keywords: Prediction, operation monitoring, on-line data, nonlinear statistical methods, empirical model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1658
291 Usage of Channel Coding Techniques for Peak-to-Average Power Ratio Reduction in Visible Light Communications Systems

Authors: P.L.D.N.M. de Silva, S.G. Edirisinghe, R. Weerasuriya

Abstract:

High Peak-to-Average Power Ratio (PAPR) is a concern of Orthogonal Frequency Division Multiplexing (OFDM) based Visible Light Communication (VLC) systems. Discrete Fourier Transform spread (DFT-s) OFDM is an alternative single carrier modulation scheme which would address this concern. Employing channel coding techniques is another mechanism to reduce the PAPR. In this study, the improvement which can be harnessed by hybridizing these two techniques for VLC system is being studied. Within the study, efficient techniques such as Hamming coding and Convolutional coding have been studied. Thus, we present the impact of the hybrid of DFT-s OFDM and Channel coding (Hamming coding and Convolutional coding) on PAPR in VLC systems, using MATLAB simulations.

Keywords: Convolutional Coding, Discrete Fourier Transform spread Orthogonal Frequency Division Multiplexing (DFT-s OFDM), Hamming Coding, Peak-to-Average Power Ratio (PAPR), Visible Light Communications (VLC).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 514
290 Improved Automated Classification of Alcoholics and Non-alcoholics

Authors: Ramaswamy Palaniappan

Abstract:

In this paper, several improvements are proposed to previous work of automated classification of alcoholics and nonalcoholics. In the previous paper, multiplayer-perceptron neural network classifying energy of gamma band Visual Evoked Potential (VEP) signals gave the best classification performance using 800 VEP signals from 10 alcoholics and 10 non-alcoholics. Here, the dataset is extended to include 3560 VEP signals from 102 subjects: 62 alcoholics and 40 non-alcoholics. Three modifications are introduced to improve the classification performance: i) increasing the gamma band spectral range by increasing the pass-band width of the used filter ii) the use of Multiple Signal Classification algorithm to obtain the power of the dominant frequency in gamma band VEP signals as features and iii) the use of the simple but effective knearest neighbour classifier. To validate that these two modifications do give improved performance, a 10-fold cross validation classification (CVC) scheme is used. Repeat experiments of the previously used methodology for the extended dataset are performed here and improvement from 94.49% to 98.71% in maximum averaged CVC accuracy is obtained using the modifications. This latest results show that VEP based classification of alcoholics is worth exploring further for system development.

Keywords: Alcoholic, Multilayer-perceptron, Nearest neighbour, Gamma band, MUSIC, Visual evoked potential.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1378
289 Data Envelopment Analysis with Partially Perfect Objects

Authors: Alexander Y. Vaninsky

Abstract:

This paper presents a simplified version of Data Envelopment Analysis (DEA) - a conventional approach to evaluating the performance and ranking of competitive objects characterized by two groups of factors acting in opposite directions: inputs and outputs. DEA with a Perfect Object (DEA PO) augments the group of actual objects with a virtual Perfect Object - the one having greatest outputs and smallest inputs. It allows for obtaining an explicit analytical solution and making a step to an absolute efficiency. This paper develops this approach further and introduces a DEA model with Partially Perfect Objects. DEA PPO consecutively eliminates the smallest relative inputs or greatest relative outputs, and applies DEA PO to the reduced collections of indicators. The partial efficiency scores are combined to get the weighted efficiency score. The computational scheme remains simple, like that of DEA PO, but the advantage of the DEA PPO is taking into account all of the inputs and outputs for each actual object. Firm evaluation is considered as an example.

Keywords: Data Envelopment Analysis, Perfect object, Partially perfect object, Partial efficiency, Explicit solution, Simplified algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1697
288 A Mesh Free Moving Node Method To Analyze Flow Through Spirals of Orbiting Scroll Pump

Authors: I.Banerjee, A.K.Mahendra, T.K.Bera, B.G.Chandresh

Abstract:

The scroll pump belongs to the category of positive displacement pump can be used for continuous pumping of gases at low pressure apart from general vacuum application. The shape of volume occupied by the gas moves and deforms continuously as the spiral orbits. To capture flow features in such domain where mesh deformation varies with time in a complicated manner, mesh less solver was found to be very useful. Least Squares Kinetic Upwind Method (LSKUM) is a kinetic theory based mesh free Euler solver working on arbitrary distribution of points. Here upwind is enforced in molecular level based on kinetic flux vector splitting scheme (KFVS). In the present study we extended the LSKUM to moving node viscous flow application. This new code LSKUM-NS-MN for moving node viscous flow is validated for standard airfoil pitching test case. Simulation performed for flow through scroll pump using LSKUM-NS-MN code agrees well with the experimental pumping speed data.

Keywords: Least Squares, Moving node, Pitching, Spirals.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1905
287 Effect of Atmospheric Pressure on the Flow at the Outlet of a Propellant Nozzle

Authors: R. Haoui

Abstract:

The purpose of this work is to simulate the flow at the exit of Vulcan 1 engine of European launcher Ariane 5. The geometry of the propellant nozzle is already determined using the characteristics method. The pressure in the outlet section of the nozzle is less than atmospheric pressure on the ground, causing the existence of oblique and normal shock waves at the exit. During the rise of the launcher, the atmospheric pressure decreases and the shock wave disappears. The code allows the capture of shock wave at exit of nozzle. The numerical technique uses the Flux Vector Splitting method of Van Leer to ensure convergence and avoid the calculation instabilities. The Courant, Friedrichs and Lewy coefficient (CFL) and mesh size level are selected to ensure the numerical convergence. The nonlinear partial derivative equations system which governs this flow is solved by an explicit unsteady numerical scheme by the finite volume method. The accuracy of the solution depends on the size of the mesh and also the step of time used in the discretized equations. We have chosen in this study the mesh that gives us a stationary solution with good accuracy.

Keywords: Launchers, supersonic flow, finite volume, nozzles, shock wave.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 877
286 Analysis of Three-Dimensional Longitudinal Rolls Induced by Double Diffusive Poiseuille-Rayleigh-Benard Flows in Rectangular Channels

Authors: O. Rahli, N. Mimouni, R. Bennacer, K. Bouhadef

Abstract:

This numerical study investigates the travelling wave’s appearance and the behavior of Poiseuille-Rayleigh-Benard (PRB) flow induced in 3D thermosolutale mixed convection (TSMC) in horizontal rectangular channels. The governing equations are discretized by using a control volume method with third order Quick scheme in approximating the advection terms. Simpler algorithm is used to handle coupling between the momentum and continuity equations. To avoid the excessively high computer time, full approximation storage (FAS) with full multigrid (FMG) method is used to solve the problem. For a broad range of dimensionless controlling parameters, the contribution of this work is to analyzing the flow regimes of the steady longitudinal thermoconvective rolls (noted R//) for both thermal and mass transfer (TSMC). The transition from the opposed volume forces to cooperating ones, considerably affects the birth and the development of the longitudinal rolls. The heat and mass transfers distribution are also examined.

Keywords: Heat and mass transfer, mixed convection, Poiseuille-Rayleigh-Benard flow, rectangular duct.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1087
285 A Robust Al-Hawalees Gaming Automation using Minimax and BPNN Decision

Authors: Ahmad Sharieh, R Bremananth

Abstract:

Artificial Intelligence based gaming is an interesting topic in the state-of-art technology. This paper presents an automation of a tradition Omani game, called Al-Hawalees. Its related issues are resolved and implemented using artificial intelligence approach. An AI approach called mini-max procedure is incorporated to make a diverse budges of the on-line gaming. If number of moves increase, time complexity will be increased in terms of propositionally. In order to tackle the time and space complexities, we have employed a back propagation neural network (BPNN) to train in off-line to make a decision for resources required to fulfill the automation of the game. We have utilized Leverberg- Marquardt training in order to get the rapid response during the gaming. A set of optimal moves is determined by the on-line back propagation training fashioned with alpha-beta pruning. The results and analyses reveal that the proposed scheme will be easily incorporated in the on-line scenario with one player against the system.

Keywords: Artificial neural network, back propagation gaming, Leverberg-Marquardt, minimax procedure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1937
284 A Decision Support System Based on Leprosy Scales

Authors: Dennys Robson Girardi, Hugo Bulegon, Claudia Maria Moro Barra

Abstract:

Leprosy is an infectious disease caused by Mycobacterium Leprae, this disease, generally, compromises the neural fibers, leading to the development of disability. Disabilities are changes that limit daily activities or social life of a normal individual. When comes to leprosy, the study of disability considered the functional limitation (physical disabilities), the limitation of activity and social participation, which are measured respectively by the scales: EHF, SALSA and PARTICIPATION SCALE. The objective of this work is to propose an on-line monitoring of leprosy patients, which is based on information scales EHF, SALSA and PARTICIPATION SCALE. It is expected that the proposed system is applied in monitoring the patient during treatment and after healing therapy of the disease. The correlations that the system is between the scales create a variety of information, presented the state of the patient and full of changes or reductions in disability. The system provides reports with information from each of the scales and the relationships that exist between them. This way, health professionals, with access to patient information, can intervene with techniques for the Prevention of Disability. Through the automated scale, the system shows the level of the patient and allows the patient, or the responsible, to take a preventive measure. With an online system, it is possible take the assessments and monitor patients from anywhere.

Keywords: Leprosy, Medical Informatics, Decision SupportSystem, Disability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2049
283 Mass Transfer Modeling of Nitrate in an Ion Exchange Selective Resin

Authors: A. A. Hekmatzadeh, A. Karimi-Jashani, N. Talebbeydokhti

Abstract:

The rate of nitrate adsorption by a nitrate selective ion exchange resin was investigated in a well-stirred batch experiments. The kinetic experimental data were simulated with diffusion models including external mass transfer, particle diffusion and chemical adsorption. Particle pore volume diffusion and particle surface diffusion were taken into consideration separately and simultaneously in the modeling. The model equations were solved numerically using the Crank-Nicholson scheme. An optimization technique was employed to optimize the model parameters. All nitrate concentration decay data were well described with the all diffusion models. The results indicated that the kinetic process is initially controlled by external mass transfer and then by particle diffusion. The external mass transfer coefficient and the coefficients of pore volume diffusion and surface diffusion in all experiments were close to each other with the average value of 8.3×10-3 cm/S for external mass transfer coefficient. In addition, the models are more sensitive to the mass transfer coefficient in comparison with particle diffusion. Moreover, it seems that surface diffusion is the dominant particle diffusion in comparison with pore volume diffusion.

Keywords: External mass transfer, pore volume diffusion, surface diffusion, mass action law isotherm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2243
282 Reversible Watermarking on Stereo Image Sequences

Authors: John N. Ellinas

Abstract:

In this paper, a new reversible watermarking method is presented that reduces the size of a stereoscopic image sequence while keeping its content visible. The proposed technique embeds the residuals of the right frames to the corresponding frames of the left sequence, halving the total capacity. The residual frames may result in after a disparity compensated procedure between the two video streams or by a joint motion and disparity compensation. The residuals are usually lossy compressed before embedding because of the limited embedding capacity of the left frames. The watermarked frames are visible at a high quality and at any instant the stereoscopic video may be recovered by an inverse process. In fact, the left frames may be exactly recovered whereas the right ones are slightly distorted as the residuals are not embedded intact. The employed embedding method reorders the left frame into an array of consecutive pixel pairs and embeds a number of bits according to their intensity difference. In this way, it hides a number of bits in intensity smooth areas and most of the data in textured areas where resulting distortions are less visible. The experimental evaluation demonstrates that the proposed scheme is quite effective.

Keywords: Stereoscopic video, Reversible watermarking, Disparity compensation, Joint compensation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1426
281 Definition and Core Components of the Role-Partner Allocation Problem in Collaborative Networks

Authors: J. Andrade-Garda, A. Anguera, J. Ares-Casal, M. Hidalgo-Lorenzo, J.-A. Lara, D. Lizcano, S. Suárez-Garaboa

Abstract:

In the current constantly changing economic context, collaborative networks allow partners to undertake projects that would not be possible if attempted by them individually. These projects usually involve the performance of a group of tasks (named roles) that have to be distributed among the partners. Thus, an allocation/matching problem arises that will be referred to as Role-Partner Allocation problem. In real life this situation is addressed by negotiation between partners in order to reach ad hoc agreements. Besides taking a long time and being hard work, both historical evidence and economic analysis show that such approach is not recommended. Instead, the allocation process should be automated by means of a centralized matching scheme. However, as a preliminary step to start the search for such a matching mechanism (or even the development of a new one), the problem and its core components must be specified. To this end, this paper establishes (i) the definition of the problem and its constraints, (ii) the key features of the involved elements (i.e., roles and partners); and (iii) how to create preference lists both for roles and partners. Only this way it will be possible to conduct subsequent methodological research on the solution method.     

Keywords: Collaborative network, matching, partner, preference list, role.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 820
280 Exploration of Least Significant Bit Based Watermarking and Its Robustness against Salt and Pepper Noise

Authors: Kamaldeep Joshi, Rajkumar Yadav, Sachin Allwadhi

Abstract:

Image steganography is the best aspect of information hiding. In this, the information is hidden within an image and the image travels openly on the Internet. The Least Significant Bit (LSB) is one of the most popular methods of image steganography. In this method, the information bit is hidden at the LSB of the image pixel. In one bit LSB steganography method, the total numbers of the pixels and the total number of message bits are equal to each other. In this paper, the LSB method of image steganography is used for watermarking. The watermarking is an application of the steganography. The watermark contains 80*88 pixels and each pixel requirs 8 bits for its binary equivalent form so, the total number of bits required to hide the watermark are 80*88*8(56320). The experiment was performed on standard 256*256 and 512*512 size images. After the watermark insertion, histogram analysis was performed. A noise factor (salt and pepper) of 0.02 was added to the stego image in order to evaluate the robustness of the method. The watermark was successfully retrieved after insertion of noise. An experiment was performed in order to know the imperceptibility of stego and the retrieved watermark. It is clear that the LSB watermarking scheme is robust to the salt and pepper noise.

Keywords: LSB, watermarking, salt and pepper, PSNR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1053
279 Energy Efficient Reliable Cooperative Multipath Routing in Wireless Sensor Networks

Authors: Gergely Treplan, Long Tran-Thanh, Janos Levendovszky

Abstract:

In this paper, a reliable cooperative multipath routing algorithm is proposed for data forwarding in wireless sensor networks (WSNs). In this algorithm, data packets are forwarded towards the base station (BS) through a number of paths, using a set of relay nodes. In addition, the Rayleigh fading model is used to calculate the evaluation metric of links. Here, the quality of reliability is guaranteed by selecting optimal relay set with which the probability of correct packet reception at the BS will exceed a predefined threshold. Therefore, the proposed scheme ensures reliable packet transmission to the BS. Furthermore, in the proposed algorithm, energy efficiency is achieved by energy balancing (i.e. minimizing the energy consumption of the bottleneck node of the routing path) at the same time. This work also demonstrates that the proposed algorithm outperforms existing algorithms in extending longevity of the network, with respect to the quality of reliability. Given this, the obtained results make possible reliable path selection with minimum energy consumption in real time.

Keywords: wireless sensor networks, reliability, cooperativerouting, Rayleigh fading model, energy balancing

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610
278 Influence Analysis of Macroeconomic Parameters on Real Estate Price Variation in Taipei, Taiwan

Authors: Li Li, Kai-Hsuan Chu

Abstract:

It is well known that the real estate price depends on a lot of factors. Each house current value is dependent on the location, room number, transportation, living convenience, year and surrounding environments. Although, there are different experienced models for housing agent to estimate the price, it is a case by case study without overall dynamic variation investigation. However, many economic parameters may more or less influence the real estate price variation. Here, the influences of most macroeconomic parameters on real estate price are investigated individually based on least-square scheme and grey correlation strategy. Then those parameters are classified into leading indices, simultaneous indices and laggard indices. In addition, the leading time period is evaluated based on least square method. The important leading and simultaneous indices can be used to establish an artificial intelligent neural network model for real estate price variation prediction. The real estate price variation of Taipei, Taiwan during 2005 ~ 2017 are chosen for this research data analysis and validation. The results show that the proposed method has reasonable prediction function for real estate business reference.

Keywords: Real estate price, least-square, grey correlation, macroeconomics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 990
277 Water Boundary Layer Flow Over Rotating Sphere with Mass Transfer

Authors: G. Revathi, P. Saikrishnan

Abstract:

An analysis is performed to study the influence of nonuniform double slot suction on a steady laminar boundary layer flow over a rotating sphere when fluid properties such as viscosity and Prandtl number are inverse linear functions of temperature. Nonsimilar solutions have been obtained from the starting point of the streamwise co-ordinate to the exact point of separation. The difficulties arising at the starting point of the streamwise co-ordinate, at the edges of the slot and at the point of separation have been overcome by applying an implicit finite difference scheme in combination with the quasi-linearization technique and an appropriate selection of the finer step sizes along the stream-wise direction. The present investigation shows that the point of ordinary separation can be delayed by nonuniform double slot suction if the mass transfer rate is increased and also if the slots are positioned further downstream. In addition, the investigation reveals that double slot suction is found to be more effective compared to a single slot suction in delaying ordinary separation. As rotation parameter increase the point of separation moves upstream direction.

Keywords: Boundary layer, suction, mass transfer, rotating sphere.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6374
276 A Monte Carlo Method to Data Stream Analysis

Authors: Kittisak Kerdprasop, Nittaya Kerdprasop, Pairote Sattayatham

Abstract:

Data stream analysis is the process of computing various summaries and derived values from large amounts of data which are continuously generated at a rapid rate. The nature of a stream does not allow a revisit on each data element. Furthermore, data processing must be fast to produce timely analysis results. These requirements impose constraints on the design of the algorithms to balance correctness against timely responses. Several techniques have been proposed over the past few years to address these challenges. These techniques can be categorized as either dataoriented or task-oriented. The data-oriented approach analyzes a subset of data or a smaller transformed representation, whereas taskoriented scheme solves the problem directly via approximation techniques. We propose a hybrid approach to tackle the data stream analysis problem. The data stream has been both statistically transformed to a smaller size and computationally approximated its characteristics. We adopt a Monte Carlo method in the approximation step. The data reduction has been performed horizontally and vertically through our EMR sampling method. The proposed method is analyzed by a series of experiments. We apply our algorithm on clustering and classification tasks to evaluate the utility of our approach.

Keywords: Data Stream, Monte Carlo, Sampling, DensityEstimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1417
275 Development of Workplace Environmental Monitoring Systems Using Ubiquitous Sensor Network

Authors: Jung-Min Yun, Jong-Hyun Baek, Byoung Ky Kang, Peom Park

Abstract:

In this study, workplace environmental monitoring systems were established using USN(Ubiquitous Sensor Networks) and LabVIEW. Although existing direct sampling methods enable finding accurate values as of the time points of measurement, those methods are disadvantageous in that continuous management and supervision are difficult and costs for are high when those methods are used. Therefore, the efficiency and reliability of workplace management by supervisors are relatively low when those methods are used. In this study, systems were established so that information on workplace environmental factors such as temperatures, humidity and noises is measured and transmitted to the PC in real time to enable supervisors to monitor workplaces through LabVIEW on the PC. When any accidents have occurred in workplaces, supervisors can immediately respond through the monitoring system and this system enables integrated workplace management and the prevention of safety accidents. By introducing these monitoring systems, safety accidents due to harmful environmental factors in workplaces can be prevented and these monitoring systems will be also helpful in finding out the correlation between safety accidents and occupational diseases by comparing and linking databases established by this monitoring system with existing statistical data.

Keywords: Ubiquitous Sensor Nework, LabVIEW, Environment Monitoring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2545
274 Auto Tuning PID Controller based on Improved Genetic Algorithm for Reverse Osmosis Plant

Authors: Jin-Sung Kim, Jin-Hwan Kim, Ji-Mo Park, Sung-Man Park, Won-Yong Choe, Hoon Heo

Abstract:

An optimal control of Reverse Osmosis (RO) plant is studied in this paper utilizing the auto tuning concept in conjunction with PID controller. A control scheme composing an auto tuning stochastic technique based on an improved Genetic Algorithm (GA) is proposed. For better evaluation of the process in GA, objective function defined newly in sense of root mean square error has been used. Also in order to achieve better performance of GA, more pureness and longer period of random number generation in operation are sought. The main improvement is made by replacing the uniform distribution random number generator in conventional GA technique to newly designed hybrid random generator composed of Cauchy distribution and linear congruential generator, which provides independent and different random numbers at each individual steps in Genetic operation. The performance of newly proposed GA tuned controller is compared with those of conventional ones via simulation.

Keywords: Genetic Algorithm, Auto tuning, Hybrid random number generator, Reverse Osmosis, PID controller

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3129
273 Spread Spectrum Image Watermarking for Secured Multimedia Data Communication

Authors: Tirtha S. Das, Ayan K. Sau, Subir K. Sarkar

Abstract:

Digital watermarking is a way to provide the facility of secure multimedia data communication besides its copyright protection approach. The Spread Spectrum modulation principle is widely used in digital watermarking to satisfy the robustness of multimedia signals against various signal-processing operations. Several SS watermarking algorithms have been proposed for multimedia signals but very few works have discussed on the issues responsible for secure data communication and its robustness improvement. The current paper has critically analyzed few such factors namely properties of spreading codes, proper signal decomposition suitable for data embedding, security provided by the key, successive bit cancellation method applied at decoder which have greater impact on the detection reliability, secure communication of significant signal under camouflage of insignificant signals etc. Based on the analysis, robust SS watermarking scheme for secure data communication is proposed in wavelet domain and improvement in secure communication and robustness performance is reported through experimental results. The reported result also shows improvement in visual and statistical invisibility of the hidden data.

Keywords: Spread spectrum modulation, spreading code, signaldecomposition, security, successive bit cancellation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2781
272 Multiobjective Optimization Solution for Shortest Path Routing Problem

Authors: C. Chitra, P. Subbaraj

Abstract:

The shortest path routing problem is a multiobjective nonlinear optimization problem with constraints. This problem has been addressed by considering Quality of service parameters, delay and cost objectives separately or as a weighted sum of both objectives. Multiobjective evolutionary algorithms can find multiple pareto-optimal solutions in one single run and this ability makes them attractive for solving problems with multiple and conflicting objectives. This paper uses an elitist multiobjective evolutionary algorithm based on the Non-dominated Sorting Genetic Algorithm (NSGA), for solving the dynamic shortest path routing problem in computer networks. A priority-based encoding scheme is proposed for population initialization. Elitism ensures that the best solution does not deteriorate in the next generations. Results for a sample test network have been presented to demonstrate the capabilities of the proposed approach to generate well-distributed pareto-optimal solutions of dynamic routing problem in one single run. The results obtained by NSGA are compared with single objective weighting factor method for which Genetic Algorithm (GA) was applied.

Keywords: Multiobjective optimization, Non-dominated SortingGenetic Algorithm, Routing, Weighted sum.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3267
271 Understanding the Notion between Resiliency and Recovery through a Spatial-Temporal Analysis of Section 404 Wetland Alteration Permits before and after Hurricane Ike

Authors: Md Y. Reja, Samuel D. Brody, Wesley E. Highfield, Galen D. Newman

Abstract:

Historically, wetlands in the United States have been lost due to agriculture, anthropogenic activities, and rapid urbanization along the coast. Such losses of wetlands have resulted in high flooding risk for coastal communities over the period of time. In addition, alteration of wetlands via the Section 404 Clean Water Act permits can increase the flooding risk to future hurricane events, as the cumulative impact of this program is poorly understood and under-accounted. Further, recovery after hurricane events is acting as an encouragement for new development and reconstruction activities by converting wetlands under the wetland alteration permitting program. This study investigates the degree to which hurricane recovery activities in coastal communities are undermining the ability of these places to absorb the impacts of future storm events. Specifically, this work explores how and to what extent wetlands are being affected by the federal permitting program post-Hurricane Ike in 2008. Wetland alteration patterns are examined across three counties (Harris, Galveston, and Chambers County) along the Texas Gulf Coast over a 10-year time period, from 2004-2013 (five years before and after Hurricane Ike) by conducting descriptive spatial analyses. Results indicate that after Hurricane Ike, the number of permits substantially increased in Harris and Chambers County. The vast majority of individual and nationwide type permits were issued within the 100-year floodplain, storm surge zones, and areas damaged by Ike flooding, suggesting that recovery after the hurricane is compromising the ecological resiliency on which coastal communities depend. The authors expect that the findings of this study can increase awareness to policy makers and hazard mitigation planners regarding how to manage wetlands during a long-term recovery process to maintain their natural functions for future flood mitigation.

Keywords: Ecological resiliency, Hurricane Ike, recovery, Section 404 permitting, wetland alteration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 861
270 Memory Estimation of Internet Server Using Queuing Theory: Comparative Study between M/G/1, G/M/1 and G/G/1 Queuing Model

Authors: L. K. Singh, Riktesh Srivastava

Abstract:

How to effectively allocate system resource to process the Client request by Gateway servers is a challenging problem. In this paper, we propose an improved scheme for autonomous performance of Gateway servers under highly dynamic traffic loads. We devise a methodology to calculate Queue Length and Waiting Time utilizing Gateway Server information to reduce response time variance in presence of bursty traffic. The most widespread contemplation is performance, because Gateway Servers must offer cost-effective and high-availability services in the elongated period, thus they have to be scaled to meet the expected load. Performance measurements can be the base for performance modeling and prediction. With the help of performance models, the performance metrics (like buffer estimation, waiting time) can be determined at the development process. This paper describes the possible queue models those can be applied in the estimation of queue length to estimate the final value of the memory size. Both simulation and experimental studies using synthesized workloads and analysis of real-world Gateway Servers demonstrate the effectiveness of the proposed system.

Keywords: M/M/1, M/G/1, G/M/1, G/G/1, Gateway Servers, Buffer Estimation, Waiting Time, Queuing Process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1936
269 Searchable Encryption in Cloud Storage

Authors: Ren-Junn Hwang, Chung-Chien Lu, Jain-Shing Wu

Abstract:

Cloud outsource storage is one of important services in cloud computing. Cloud users upload data to cloud servers to reduce the cost of managing data and maintaining hardware and software. To ensure data confidentiality, users can encrypt their files before uploading them to a cloud system. However, retrieving the target file from the encrypted files exactly is difficult for cloud server. This study proposes a protocol for performing multikeyword searches for encrypted cloud data by applying k-nearest neighbor technology. The protocol ranks the relevance scores of encrypted files and keywords, and prevents cloud servers from learning search keywords submitted by a cloud user. To reduce the costs of file transfer communication, the cloud server returns encrypted files in order of relevance. Moreover, when a cloud user inputs an incorrect keyword and the number of wrong alphabet does not exceed a given threshold; the user still can retrieve the target files from cloud server. In addition, the proposed scheme satisfies security requirements for outsourced data storage.

Keywords: Fault-tolerance search, multi-keywords search, outsource storage, ranked search, searchable encryption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3081
268 Strategic Redesign of Public Spaces with a Sustainable Approach: Case Study of Parque Huancavilca, Guayaquil

Authors: Juan Carlos Briones Macias

Abstract:

Currently, the Huancavilca City Park in Guayaquil is an abandoned public space that is discovering a growing problem of insecurity, where various problems have been perceived, such as the lack of green areas, deteriorating furniture, insufficient lighting, the use of inadequate cladding materials and very sunny areas due to the lack of planning in the design of green areas. The objective of this scientific article is to redesign Huancavilca Park through public space design strategies for more attractive and comfortable areas, becoming a point of interaction in a safe and accessible way. A mixed methodology (qualitative and quantitative) was applied, obtaining information based on surveys, interviews, field observations, and systematizing the data in the traditional weighting of the structuring aspects of the park. The results were obtained from the methodological design scheme of iterative analysis of public spaces by Jan Güell. It is concluded that the use of urban strategies in the structuring elements of the park, such as vegetation, furniture, generating new activities, and security interventions, will specifically solve all the problems of the Huancavilca Park tested in a Pareto 80/20 Diagram.

Keywords: Public space, green areas, vegetation, street furniture, urban analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 331
267 Torque Ripple Minimization in Switched Reluctance Motor Using Passivity-Based Robust Adaptive Control

Authors: M.M. Namazi, S.M. Saghaiannejad, A. Rashidi

Abstract:

In this paper by using the port-controlled Hamiltonian (PCH) systems theory, a full-order nonlinear controlled model is first developed. Then a nonlinear passivity-based robust adaptive control (PBRAC) of switched reluctance motor in the presence of external disturbances for the purpose of torque ripple reduction and characteristic improvement is presented. The proposed controller design is separated into the inner loop and the outer loop controller. In the inner loop, passivity-based control is employed by using energy shaping techniques to produce the proper switching function. The outer loop control is employed by robust adaptive controller to determine the appropriate Torque command. It can also overcome the inherent nonlinear characteristics of the system and make the whole system robust to uncertainties and bounded disturbances. A 4KW 8/6 SRM with experimental characteristics that takes magnetic saturation into account is modeled, simulation results show that the proposed scheme has good performance and practical application prospects.

Keywords: Switched Reluctance Motor, Port HamiltonianSystem, Passivity-Based Control, Torque Ripple Minimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1680
266 Transient Stability Assessment Using Fuzzy SVM and Modified Preventive Control

Authors: B. Dora Arul Selvi, .N. Kamaraj

Abstract:

Transient Stability is an important issue in power systems planning, operation and extension. The objective of transient stability analysis problem is not satisfied with mere transient instability detection or evaluation and it is most important to complement it by defining fast and efficient control measures in order to ensure system security. This paper presents a new Fuzzy Support Vector Machines (FSVM) to investigate the stability status of power systems and a modified generation rescheduling scheme to bring back the identified unstable cases to a more economical and stable operating point. FSVM improves the traditional SVM (Support Vector Machines) by adding fuzzy membership to each training sample to indicate the degree of membership of this sample to different classes. The preventive control based on economic generator rescheduling avoids the instability of the power systems with minimum change in operating cost under disturbed conditions. Numerical results on the New England 39 bus test system show the effectiveness of the proposed method.

Keywords: Fuzzy Support Vector Machine (FSVM), Incremental Cost, Preventive Control, Transient stability

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
265 Fluidized-Bed Combustion of Biomass with Elevated Alkali Content: A Comparative Study between Two Alternative Bed Materials

Authors: P. Ninduangdee, V. I. Kuprianov

Abstract:

Palm kernel shell is an important bioenergy resource in Thailand. However, due to elevated alkali content in biomass ash, this oil palm residue shows high tendency to bed agglomeration in a fluidized-bed combustion system using conventional bed material (silica sand). In this study, palm kernel shell was burned in the conical fluidized-bed combustor (FBC) using alumina and dolomite as alternative bed materials to prevent bed agglomeration. For each bed material, the combustion tests were performed at 45kg/h fuel feed rate with excess air within 20–80%. Experimental results revealed rather weak effects of the bed material type but substantial influence of excess air on the behavior of temperature, O2, CO, CxHy, and NO inside the reactor, as well as on the combustion efficiency and major gaseous emissions of the conical FBC. The optimal level of excess air ensuring high combustion efficiency (about 98.5%) and acceptable level of the emissions was found to be about 40% when using alumina and 60% with dolomite. By using these alternative bed materials, bed agglomeration can be prevented when burning the shell in the proposed conical FBC. However, both bed materials exhibited significant changes in their morphological, physical and chemical properties in the course of the time.

Keywords: Palm kernel shell, fluidized-bed combustion, alternative bed materials, combustion and emission performance, bed agglomeration prevention.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3040