Search results for: a method of problem solving.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10548

Search results for: a method of problem solving.

7608 MAS Simulations of Optical Antenna Structures

Authors: K.Tavzarashvili, G.Ghvedashili

Abstract:

A semi-analytic boundary discretization method, the Method of Auxiliary Sources (MAS) is used to analyze Optical Antennas consisting of metallic parts. In addition to standard dipoletype antennas, consisting of two pieces of metal, a new structure consisting of a single metal piece with a tiny groove in the center is analyzed. It is demonstrated that difficult numerical problems are caused because optical antennas exhibit strong material dispersion, loss, and plasmon-polariton effects that require a very accurate numerical simulation. This structure takes advantage of the Channel Plasmon-Polariton (CPP) effect and exhibits a strong enhancement of the electric field in the groove. Also primitive 3D antenna model with spherical nano particles is analyzed.

Keywords: optical antenna, channel plasmon-polariton, computational physics, Method of Auxiliary Sources

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1914
7607 Estimation of Load Impedance in Presence of Harmonics

Authors: Khaled M. EL-Naggar

Abstract:

This paper presents a fast and efficient on-line technique for estimating impedance of unbalanced loads in power systems. The proposed technique is an application of a discrete timedynamic filter based on stochastic estimation theory which is suitable for estimating parameters in noisy environment. The algorithm uses sets of digital samples of the distorted voltage and current waveforms of the non-linear load to estimate the harmonic contents of these two signal. The non-linear load impedance is then calculated from these contents. The method is tested using practical data. Results are reported and compared with those obtained using the conventional least error squares technique. In addition to the very accurate results obtained, the method can detect and reject bad measurements. This can be considered as a very important advantage over the conventional static estimation methods such as the least error square method.

Keywords: Estimation, identification, Harmonics, Dynamic Filter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2064
7606 An Optimization of the New Die Design of Sheet Hydroforming by Taguchi Method

Authors: M. Hosseinzadeh, S. A. Zamani, A. Taheri

Abstract:

During the last few years, several sheet hydroforming processes have been introduced. Despite the advantages of these methods, they have some limitations. Of the processes, the two main ones are the standard hydroforming and hydromechanical deep drawing. A new sheet hydroforming die set was proposed that has the advantages of both processes and eliminates their limitations. In this method, a polyurethane plate was used as a part of the die-set to control the blank holder force. This paper outlines the Taguchi optimization methodology, which is applied to optimize the effective parameters in forming cylindrical cups by the new die set of sheet hydroforming process. The process parameters evaluated in this research are polyurethane hardness, polyurethane thickness, forming pressure path and polyurethane hole diameter. The design of experiments based upon L9 orthogonal arrays by Taguchi was used and analysis of variance (ANOVA) was employed to analyze the effect of these parameters on the forming pressure. The analysis of the results showed that the optimal combination for low forming pressure is harder polyurethane, bigger diameter of polyurethane hole and thinner polyurethane. Finally, the confirmation test was derived based on the optimal combination of parameters and it was shown that the Taguchi method is suitable to examine the optimization process.

Keywords: Sheet Hydroforming, Optimization, Taguchi Method

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2596
7605 Comparison of Particle Swarm Optimization and Genetic Algorithm for TCSC-based Controller Design

Authors: Sidhartha Panda, N. P. Padhy

Abstract:

Recently, genetic algorithms (GA) and particle swarm optimization (PSO) technique have attracted considerable attention among various modern heuristic optimization techniques. Since the two approaches are supposed to find a solution to a given objective function but employ different strategies and computational effort, it is appropriate to compare their performance. This paper presents the application and performance comparison of PSO and GA optimization techniques, for Thyristor Controlled Series Compensator (TCSC)-based controller design. The design objective is to enhance the power system stability. The design problem of the FACTS-based controller is formulated as an optimization problem and both the PSO and GA optimization techniques are employed to search for optimal controller parameters. The performance of both optimization techniques in terms of computational time and convergence rate is compared. Further, the optimized controllers are tested on a weakly connected power system subjected to different disturbances, and their performance is compared with the conventional power system stabilizer (CPSS). The eigenvalue analysis and non-linear simulation results are presented and compared to show the effectiveness of both the techniques in designing a TCSC-based controller, to enhance power system stability.

Keywords: Thyristor Controlled Series Compensator, geneticalgorithm; particle swarm optimization; Phillips-Heffron model;power system stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3153
7604 BPNN Based Processing for End Effects of HHT

Authors: Chun-Yao Lee, Yao-chen Lee

Abstract:

This paper describes a method of signal process applied on an end effects of Hilbert-Huang transform (HHT) to provide an improvement in the reality of spectrum. The method is based on back-propagation network (BPN). To improve the effect, the end extension of the original signal is obtained by back-propagation network. A full waveform including origin and its extension is decomposed by using empirical mode decomposition (EMD) to obtain intrinsic mode functions (IMFs) of the waveform. Then, the Hilbert transform (HT) is applied to the IMFs to obtain the Hilbert spectrum of the waveform. As a result, the method is superiority of the processing of end effect of HHT to obtain the real frequency spectrum of signals.

Keywords: Neural network, back-propagation network, Hilbert-Huang transform

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1790
7603 Joint Optimization of Pricing and Advertisement for Seasonal Branded Products

Authors: Mohammad Modarres, Shirin Aslani

Abstract:

The goal of this paper is to develop a model to integrate “pricing" and “advertisement" for short life cycle products, such as branded fashion clothing products. To achieve this goal, we apply the concept of “Dynamic Pricing". There are two classes of advertisements, for the brand (regardless of product) and for a particular product. Advertising the brand affects the demand and price of all the products. Thus, the model considers all these products in relation with each other. We develop two different methods to integrate both types of advertisement and pricing. The first model is developed within the framework of dynamic programming. However, due to the complexity of the model, this method cannot be applicable for large size problems. Therefore, we develop another method, called hieratical approach, which is capable of handling the real world problems. Finally, we show the accuracy of this method, both theoretically and also by simulation.

Keywords: Advertising, Dynamic programming, Dynamic pricing, Promotion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1635
7602 A PSO-Based Optimum Design of PID Controller for a Linear Brushless DC Motor

Authors: Mehdi Nasri, Hossein Nezamabadi-pour, Malihe Maghfoori

Abstract:

This Paper presents a particle swarm optimization (PSO) method for determining the optimal proportional-integral-derivative (PID) controller parameters, for speed control of a linear brushless DC motor. The proposed approach has superior features, including easy implementation, stable convergence characteristic and good computational efficiency. The brushless DC motor is modelled in Simulink and the PSO algorithm is implemented in MATLAB. Comparing with Genetic Algorithm (GA) and Linear quadratic regulator (LQR) method, the proposed method was more efficient in improving the step response characteristics such as, reducing the steady-states error; rise time, settling time and maximum overshoot in speed control of a linear brushless DC motor.

Keywords: Brushless DC motor, Particle swarm optimization, PID Controller, Optimal control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4963
7601 Requirements Driven Multiple View Paradigm for Developing Security Architecture

Authors: K. Chandra Sekaran

Abstract:

This paper describes a paradigmatic approach to develop architecture of secure systems by describing the requirements from four different points of view: that of the owner, the administrator, the user, and the network. Deriving requirements and developing architecture implies the joint elicitation and describing the problem and the structure of the solution. The view points proposed in this paper are those we consider as requirements towards their contributions as major parties in the design, implementation, usage and maintenance of secure systems. The dramatic growth of the technology of Internet and the applications deployed in World Wide Web have lead to the situation where the security has become a very important concern in the development of secure systems. Many security approaches are currently being used in organizations. In spite of the widespread use of many different security solutions, the security remains a problem. It is argued that the approach that is described in this paper for the development of secure architecture is practical by all means. The models representing these multiple points of view are termed the requirements model (views of owner and administrator) and the operations model (views of user and network). In this paper, this multiple view paradigm is explained by first describing the specific requirements and or characteristics of secure systems (particularly in the domain of networks) and the secure architecture / system development methodology.

Keywords: Multiple view paradigms, requirements model, operations model, secure system, owner, administrator, user, network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371
7600 Principal Component Analysis-Ranking as a Variable Selection Method for the Simultaneous Spectrophotometric Determination of Phenol, Resorcinol and Catechol in Real Samples

Authors: Nahid Ghasemi, Mohammad Goodarzi, Morteza Khosravi

Abstract:

Simultaneous determination of multicomponents of phenol, resorcinol and catechol with a chemometric technique a PCranking artificial neural network (PCranking-ANN) algorithm is reported in this study. Based on the data correlation coefficient method, 3 representative PCs are selected from the scores of original UV spectral data (35 PCs) as the original input patterns for ANN to build a neural network model. The results obtained by iterating 8000 .The RMSEP for phenol, resorcinol and catechol with PCranking- ANN were 0.6680, 0.0766 and 0.1033, respectively. Calibration matrices were 0.50-21.0, 0.50-15.1 and 0.50-20.0 μg ml-1 for phenol, resorcinol and catechol, respectively. The proposed method was successfully applied for the determination of phenol, resorcinol and catechol in synthetic and water samples.

Keywords: Phenol, Resorcinol, Catechol, Principal componentrankingArtificial Neural Network, Chemometrics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1427
7599 Intelligent Temperature Controller for Water-Bath System

Authors: Om Prakash Verma, Rajesh Singla, Rajesh Kumar

Abstract:

Conventional controller’s usually required a prior knowledge of mathematical modelling of the process. The inaccuracy of mathematical modelling degrades the performance of the process, especially for non-linear and complex control problem. The process used is Water-Bath system, which is most widely used and nonlinear to some extent. For Water-Bath system, it is necessary to attain desired temperature within a specified period of time to avoid the overshoot and absolute error, with better temperature tracking capability, else the process is disturbed.

To overcome above difficulties intelligent controllers, Fuzzy Logic (FL) and Adaptive Neuro-Fuzzy Inference System (ANFIS), are proposed in this paper. The Fuzzy controller is designed to work with knowledge in the form of linguistic control rules. But the translation of these linguistic rules into the framework of fuzzy set theory depends on the choice of certain parameters, for which no formal method is known. To design ANFIS, Fuzzy-Inference-System is combined with learning capability of Neural-Network.

It is analyzed that ANFIS is best suitable for adaptive temperature control of above system. As compared to PID and FLC, ANFIS produces a stable control signal. It has much better temperature tracking capability with almost zero overshoot and minimum absolute error.

Keywords: PID Controller, FLC, ANFIS, Non-Linear Control System, Water-Bath System, MATLAB-7.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5548
7598 Efficiency of Post-Tensioning Method for Seismic Retrofitting of Pre-Cast Cylindrical Concrete Reservoirs

Authors: M.E.Karbaschi, R.Goudarzizadeh, N.Hedayat

Abstract:

Cylindrical concrete reservoirs are appropriate choice for storing liquids as water, oil and etc. By using of the pre-cast concrete reservoirs instead of the in-situ constructed reservoirs, the speed and precision of the construction would considerably increase. In this construction method, wall and roof panels would make in factory with high quality materials and precise controlling. Then, pre-cast wall and roof panels would carry out to the construction site for assembling. This method has a few faults such as: the existing weeks in connection of wall panels together and wall panels to foundation. Therefore, these have to be resisted under applied loads such as seismic load. One of the innovative methods which was successfully applied for seismic retrofitting of numerous pre-cast cylindrical water reservoirs in New Zealand, using of the high tensile cables around the reservoirs and post-tensioning them. In this paper, analytical modeling of wall and roof panels and post-tensioned cables are carried out with finite element method and the effect of height to diameter ratio, post-tensioning force value, liquid level in reservoir, installing position of tendons on seismic response of reservoirs are investigated.

Keywords: Seismic Retrofit, Pre-Cast, Concrete Reservoir, Post-Tensioning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2026
7597 ASLT Method for Beer Accelerated Shelf-Life Determination

Authors: Tatjana Rakcejeva, Valentina Skorina, Daina Karklina, Liga Skudra

Abstract:

The aim of current research was to investigate ASLT method suitability for accelerated beer shelf-life determination. The research was accomplished on popular Latvian beer: light filtrated and unfiltered pasteurized beer with alcohol content 5.2%; dark filtrated pasteurized beer with alcohol content 4.2% with shelf-life five months. Bottled in dark glass bottles beer samples were storage during 20 weeks at several temperature regimes: +10±1 °C, +20±1 °C, +30±1 °C, +40±1 °C. Samples quality parameters as physically-chemical and microbiological was tested every two weeks using standard methods. It is possible to determine beer shelf-life rapidly during storage at +30±1 °C for filtered pasteurized light beer by 2.5 times, unfiltered pasteurized light beer by 1.4 times and for filtered pasteurized dark beer by 1.7 times. During preset experiments it was proved, that it is possible to determine beer shelf-life rapidly using ASLT method if beer storage temperature could be increased by +10±1 °C.

Keywords: Beer, shelf-life, ASLT method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6099
7596 Optimal Capacitor Allocation for loss reduction in Distribution System Using Fuzzy and Plant Growth Simulation Algorithm

Authors: R. Srinivasa Rao

Abstract:

This paper presents a new and efficient approach for capacitor placement in radial distribution systems that determine the optimal locations and size of capacitor with an objective of improving the voltage profile and reduction of power loss. The solution methodology has two parts: in part one the loss sensitivity factors are used to select the candidate locations for the capacitor placement and in part two a new algorithm that employs Plant growth Simulation Algorithm (PGSA) is used to estimate the optimal size of capacitors at the optimal buses determined in part one. The main advantage of the proposed method is that it does not require any external control parameters. The other advantage is that it handles the objective function and the constraints separately, avoiding the trouble to determine the barrier factors. The proposed method is applied to 9 and 34 bus radial distribution systems. The solutions obtained by the proposed method are compared with other methods. The proposed method has outperformed the other methods in terms of the quality of solution.

Keywords: Distribution systems, Capacitor allocation, Loss reduction, Fuzzy, PGSA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2279
7595 Introduction of the Harmfulness of the Seismic Signal in the Assessment of the Performance of Reinforced Concrete Frame Structures

Authors: Kahil Amar, Boukais Said, Kezmane Ali, Hamizi Mohand, Hannachi Naceur Eddine

Abstract:

The principle of the seismic performance evaluation methods is to provide a measure of capability for a building or set of buildings to be damaged by an earthquake. The common objective of many of these methods is to supply classification criteria. The purpose of this study is to present a method for assessing the seismic performance of structures, based on Pushover method; we are particularly interested in reinforced concrete frame structures, which represent a significant percentage of damaged structures after a seismic event. The work is based on the characterization of seismic movement of the various earthquake zones in terms of PGA and PGD that is obtained by means of SIMQK_GR and PRISM software and the correlation between the points of performance and the scalar characterizing the earthquakes will developed.

Keywords: Seismic performance, Pushover method, characterization of seismic motion, harmfulness of the seismic signal

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2055
7594 Well-Being Inequality Using Superimposing Satisfaction Waves: Heisenberg Uncertainty in Behavioural Economics and Econometrics

Authors: Okay Gunes

Abstract:

In this article, a new method is proposed for the measuring of well-being inequality through a model composed of superimposing satisfaction waves. The displacement of households’ satisfactory state (i.e. satisfaction) is defined in a satisfaction string. The duration of the satisfactory state for a given period is measured in order to determine the relationship between utility and total satisfactory time, itself dependent on the density and tension of each satisfaction string. Thus, individual cardinal total satisfaction values are computed by way of a one-dimensional form for scalar sinusoidal (harmonic) moving wave function, using satisfaction waves with varying amplitudes and frequencies which allow us to measure wellbeing inequality. One advantage to using satisfaction waves is the ability to show that individual utility and consumption amounts would probably not commute; hence, it is impossible to measure or to know simultaneously the values of these observables from the dataset. Thus, we crystallize the problem by using a Heisenberg-type uncertainty resolution for self-adjoint economic operators. We propose to eliminate any estimation bias by correlating the standard deviations of selected economic operators; this is achieved by replacing the aforementioned observed uncertainties with households’ perceived uncertainties (i.e. corrected standard deviations) obtained through the logarithmic psychophysical law proposed by Weber and Fechner.

Keywords: Heisenberg Uncertainty Principle, superimposing satisfaction waves, Weber–Fechner law, well-being inequality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2055
7593 An Automation of Check Focusing on CRUD for Requirements Analysis Model in UML

Authors: Shinpei Ogata, Yoshitaka Aoki, Hirotaka Okuda, Saeko Matsuura

Abstract:

A key to success of high quality software development is to define valid and feasible requirements specification. We have proposed a method of model-driven requirements analysis using Unified Modeling Language (UML). The main feature of our method is to automatically generate a Web user interface mock-up from UML requirements analysis model so that we can confirm validity of input/output data for each page and page transition on the system by directly operating the mock-up. This paper proposes a support method to check the validity of a data life cycle by using a model checking tool “UPPAAL" focusing on CRUD (Create, Read, Update and Delete). Exhaustive checking improves the quality of requirements analysis model which are validated by the customers through automatically generated mock-up. The effectiveness of our method is discussed by a case study of requirements modeling of two small projects which are a library management system and a supportive sales system for text books in a university.

Keywords: CRUD, Model Checking, Model Driven Development, Requirements Analysis, Unified Modeling Language, UPPAAL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1673
7592 Electromagnetic Assessment of Submarine Power Cable Degradation Using Finite Element Method and Sensitivity Analysis

Authors: N. Boutra, N. Ravot, J. Benoit, O. Picon

Abstract:

Submarine power cables used for offshore wind farms electric energy distribution and transmission are subject to numerous threats. Some of the risks are associated with transport, installation and operating in harsh marine environment. This paper describes the feasibility of an electromagnetic low frequency sensing technique for submarine power cable failure prediction. The impact of a structural damage shape and material variability on the induced electric field is evaluated. The analysis is performed by modeling the cable using the finite element method, we use sensitivity analysis in order to identify the main damage characteristics affecting electric field variation. Lastly, we discuss the results obtained.

Keywords: Electromagnetism, defect, finite element method, sensitivity analysis, submarine power cables.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1093
7591 Developing Improvements to Multi-Hazard Risk Assessments

Authors: A. Fathianpour, M. B. Jelodar, S. Wilkinson

Abstract:

This paper outlines the approaches taken to assess multi-hazard assessments. There is currently confusion in assessing multi-hazard impacts, and so this study aims to determine which of the available options are the most useful. The paper uses an international literature search, and analysis of current multi-hazard assessments and a case study to illustrate the effectiveness of the chosen method. Findings from this study will help those wanting to assess multi-hazards to undertake a straightforward approach. The paper is significant as it helps to interpret the various approaches and concludes with the preferred method. Many people in the world live in hazardous environments and are susceptible to disasters. Unfortunately, when a disaster strikes it is often compounded by additional cascading hazards, thus people would confront more than one hazard simultaneously. Hazards include natural hazards (earthquakes, floods, etc.) or cascading human-made hazards (for example, Natural Hazard Triggering Technological disasters (Natech) such as fire, explosion, toxic release). Multi-hazards have a more destructive impact on urban areas than one hazard alone. In addition, climate change is creating links between different disasters such as causing landslide dams and debris flows leading to more destructive incidents. Much of the prevailing literature deals with only one hazard at a time. However, recently sophisticated multi-hazard assessments have started to appear. Given that multi-hazards occur, it is essential to take multi-hazard risk assessment under consideration. This paper aims to review the multi-hazard assessment methods through articles published to date and categorize the strengths and disadvantages of using these methods in risk assessment. Napier City is selected as a case study to demonstrate the necessity of using multi-hazard risk assessments. In order to assess multi-hazard risk assessments, first, the current multi-hazard risk assessment methods were described. Next, the drawbacks of these multi-hazard risk assessments were outlined. Finally, the improvements to current multi-hazard risk assessments to date were summarised. Generally, the main problem of multi-hazard risk assessment is to make a valid assumption of risk from the interactions of different hazards. Currently, risk assessment studies have started to assess multi-hazard situations, but drawbacks such as uncertainty and lack of data show the necessity for more precise risk assessment. It should be noted that ignoring or partial considering multi-hazards in risk assessment will lead to an overestimate or overlook in resilient and recovery action managements.

Keywords: Cascading hazards, multi-hazard, risk assessment, risk reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1097
7590 Offset Dependent Uniform Delay Mathematical Optimization Model for Signalized Traffic Network Using Differential Evolution Algorithm

Authors: Tahseen Al-Shaikhli, Halim Ceylan, Jonathan Weaver, Osman Nuri Çelik, Onur Gungor Sahin

Abstract:

A concept of uniform delay offset dependent mathematical optimization problem is derived as the main objective for this study using a differential evolution algorithm. Furthermore, the objectives are to control the coordination problem which mainly depends on offset selection, and to estimate the uniform delay based on the offset choice at each signalized intersection. The assumption is the periodic sinusoidal function for arrival and departure patterns. The cycle time is optimized at the entry links and the optimized value is used in the non-entry links as a common cycle time. The offset optimization algorithm is used to calculate the uniform delay at each link. The results are illustrated by using a case study and compared with the canonical uniform delay model derived by Webster and the highway capacity manual’s model. The findings show that the derived model minimizes the total uniform delay to almost half compared to conventional models; the mathematical objective function is robust; the algorithm convergence time is fast.

Keywords: Area traffic control, differential evolution, offset variable, sinusoidal periodic function, traffic flow, uniform delay.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 368
7589 A Comparative Study of Turbulence Models Performance for Turbulent Flow in a Planar Asymmetric Diffuser

Authors: Samy M. El-Behery, Mofreh H. Hamed

Abstract:

This paper presents a computational study of the separated flow in a planer asymmetric diffuser. The steady RANS equations for turbulent incompressible fluid flow and six turbulence closures are used in the present study. The commercial software code, FLUENT 6.3.26, was used for solving the set of governing equations using various turbulence models. Five of the used turbulence models are available directly in the code while the v2-f turbulence model was implemented via User Defined Scalars (UDS) and User Defined Functions (UDF). A series of computational analysis is performed to assess the performance of turbulence models at different grid density. The results show that the standard k-ω, SST k-ω and v2-f models clearly performed better than other models when an adverse pressure gradient was present. The RSM model shows an acceptable agreement with the velocity and turbulent kinetic energy profiles but it failed to predict the location of separation and attachment points. The standard k-ε and the low-Re k- ε delivered very poor results.

Keywords: Turbulence models, turbulent flow, wall functions, separation, reattachment, diffuser.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3770
7588 Simultaneous Determination of Reference Free-Stream Temperature and Convective Heat Transfer Coefficient

Authors: Giho Jeong, Sooin Jeong, Kuisoon Kim

Abstract:

It is very important to determine reference temperature when convective temperature because it should be used to calculate the temperature potential. This paper deals with the development of a new method that can determine heat transfer coefficient and reference free stream temperature simultaneously, based on transient heat transfer experiments with using two narrow band thermo-tropic liquid crystals (TLC's). The method is validated through error analysis in terms of the random uncertainties in the measured temperatures. It is shown how the uncertainties in heat transfer coefficient and free stream temperature can be reduced. The general method described in this paper is applicable to many heat transfer models with unknown free stream temperature.

Keywords: Heat transfer coefficient, Thermo-tropic LiquidCrystal (TLC), Free stream temperature.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1613
7587 Study on Scheduling of the Planning Method Using the Web-based Visualization System in a Shipbuilding Block Assembly Shop

Authors: A. Eui Koog Ahn, B. Gi-Nam Wang, C. Sang C. Park

Abstract:

Higher productivity and less cost in the ship manufacturing process are required to maintain the international competitiveness of morden manufacturing industries. In shipbuilding, however, the Engineering To Order (ETO) production method and production process is very difficult. Thus, designs change frequently. In accordance with production, planning should be set up according to scene changes. Therefore, fixed production planning is very difficult. Thus, a scheduler must first make sketchy plans, then change the plans based on the work progress and modifications. Thus, data sharing in a shipbuilding block assembly shop is very important. In this paper, we proposed to scheduling method applicable to the shipbuilding industry and decision making support system through web based visualization system.

Keywords: Shipbuilding, Monitoring, Block assembly shop, Visualization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2064
7586 Exploration of Least Significant Bit Based Watermarking and Its Robustness against Salt and Pepper Noise

Authors: Kamaldeep Joshi, Rajkumar Yadav, Sachin Allwadhi

Abstract:

Image steganography is the best aspect of information hiding. In this, the information is hidden within an image and the image travels openly on the Internet. The Least Significant Bit (LSB) is one of the most popular methods of image steganography. In this method, the information bit is hidden at the LSB of the image pixel. In one bit LSB steganography method, the total numbers of the pixels and the total number of message bits are equal to each other. In this paper, the LSB method of image steganography is used for watermarking. The watermarking is an application of the steganography. The watermark contains 80*88 pixels and each pixel requirs 8 bits for its binary equivalent form so, the total number of bits required to hide the watermark are 80*88*8(56320). The experiment was performed on standard 256*256 and 512*512 size images. After the watermark insertion, histogram analysis was performed. A noise factor (salt and pepper) of 0.02 was added to the stego image in order to evaluate the robustness of the method. The watermark was successfully retrieved after insertion of noise. An experiment was performed in order to know the imperceptibility of stego and the retrieved watermark. It is clear that the LSB watermarking scheme is robust to the salt and pepper noise.

Keywords: LSB, watermarking, salt and pepper, PSNR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1053
7585 Traffic Density Measurement by Automatic Detection of Vehicles Using Gradient Vectors from Aerial Images

Authors: Saman Ghaffarian, Ilgın Gökasar

Abstract:

This paper presents a new automatic vehicle detection method from very high resolution aerial images to measure traffic density. The proposed method starts by extracting road regions from image using road vector data. Then, the road image is divided into equal sections considering resolution of the images. Gradient vectors of the road image are computed from edge map of the corresponding image. Gradient vectors on the each boundary of the sections are divided where the gradient vectors significantly change their directions. Finally, number of vehicles in each section is carried out by calculating the standard deviation of the gradient vectors in each group and accepting the group as vehicle that has standard deviation above predefined threshold value. The proposed method was tested in four very high resolution aerial images acquired from Istanbul, Turkey which illustrate roads and vehicles with diverse characteristics. The results show the reliability of the proposed method in detecting vehicles by producing 86% overall F1 accuracy value.

Keywords: Aerial images, intelligent transportation systems, traffic density measurement, vehicle detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2936
7584 Exploring Inter-Relationships between Events to Identify Strategic Technological Competencies: A Combined Approach

Authors: Cláudio Santos, Madalena Araújo, Nuno Correia

Abstract:

The inherent complexity in nowadays- business environments is forcing organizations to be attentive to the dynamics in several fronts. Therefore, the management of technological innovation is continually faced with uncertainty about the future. These issues lead to a need for a systemic perspective, able to analyze the consequences of interactions between different factors. The field of technology foresight has proposed methods and tools to deal with this broader perspective. In an attempt to provide a method to analyze the complex interactions between events in several areas, departing from the identification of the most strategic competencies, this paper presents a methodology based on the Delphi method and Quality Function Deployment. This methodology is applied in a sheet metal processing equipment manufacturer, as a case study.

Keywords: Competencies, Delphi Method, Quality Function Deployment, Technology Foresight.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703
7583 Statistical Reliability Based Modeling of Series and Parallel Operating Systems using Extreme Value Theory

Authors: Mohamad Mahdavi, Mojtaba Mahdavi

Abstract:

This paper tries to represent a new method for computing the reliability of a system which is arranged in series or parallel model. In this method we estimate life distribution function of whole structure using the asymptotic Extreme Value (EV) distribution of Type I, or Gumbel theory. We use EV distribution in minimal mode, for estimate the life distribution function of series structure and maximal mode for parallel system. All parameters also are estimated by Moments method. Reliability function and failure (hazard) rate and p-th percentile point of each function are determined. Other important indexes such as Mean Time to Failure (MTTF), Mean Time to repair (MTTR), for non-repairable and renewal systems in both of series and parallel structure will be computed.

Keywords: Reliability, extreme value, parallel, series, lifedistribution

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2090
7582 Stress and Strain Analysis of Notched Bodies Subject to Non-Proportional Loadings

Authors: A. Ince

Abstract:

In this paper, an analytical simplified method for calculating elasto-plastic stresses strains of notched bodies subject to non-proportional loading paths is discussed. The method was based on the Neuber notch correction, which relates the incremental elastic and elastic-plastic strain energy densities at the notch root and the material constitutive relationship. The validity of the method was presented by comparing computed results of the proposed model against finite element numerical data of notched shaft. The comparison showed that the model estimated notch-root elasto-plastic stresses strains with good accuracy using linear-elastic stresses. The prosed model provides more efficient and simple analysis method preferable to expensive experimental component tests and more complex and time consuming incremental non-linear FE analysis. The model is particularly suitable to perform fatigue life and fatigue damage estimates of notched components subjected to nonproportional loading paths.

Keywords: Elasto-plastic, stress-strain, notch analysis, nonprortional loadings, cyclic plasticity, fatigue.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2557
7581 Determination of Penicillins Residues in Livestock and Marine Products by LC/MS/MS

Authors: Ji Young Song, Soo Jung Hu, Hyunjin Joo, Joung Boon Hwang, Mi Ok Kim, Shin Jung Kang, Dae Hyun Cho

Abstract:

Multi-residue analysis method for penicillins was developed and validated in bovine muscle, chicken, milk, and flatfish. Detection was based on liquid chromatography tandem mass spectrometry (LC/MS/MS). The developed method was validated for specificity, precision, recovery, and linearity. The analytes were extracted with 80% acetonitrile and clean-up by a single reversed-phase solid-phase extraction step. Six penicillins presented recoveries higher than 76% with the exception of Amoxicillin (59.7%). Relative standard deviations (RSDs) were not more than 10%. LOQs values ranged from 0.1 and to 4.5 ug/kg. The method was applied to 128 real samples. Benzylpenicillin was detected in 15 samples and Cloxacillin was detected in 7 samples. Oxacillin was detected in 2 samples. But the detected levels were under the MRL levels for penicillins in samples.

Keywords: Penicillins, livestock product, Multi-residue analysis, LC/MS/MS

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3418
7580 Optimal DG Placement in Distribution systems Using Cost/Worth Analysis

Authors: M Ahmadigorji, A. Abbaspour, A Rajabi-Ghahnavieh, M. Fotuhi- Firuzabad

Abstract:

DG application has received increasing attention during recent years. The impact of DG on various aspects of distribution system operation, such as reliability and energy loss, depend highly on DG location in distribution feeder. Optimal DG placement is an important subject which has not been fully discussed yet. This paper presents an optimization method to determine optimal DG placement, based on a cost/worth analysis approach. This method considers technical and economical factors such as energy loss, load point reliability indices and DG costs, and particularly, portability of DG. The proposed method is applied to a test system and the impacts of different parameters such as load growth rate and load forecast uncertainty (LFU) on optimum DG location are studied.

Keywords: Distributed generation, optimal placement, cost/worthanalysis, customer interruption cost, Dynamic programming

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2975
7579 Studding of Number of Dataset on Precision of Estimated Saturated Hydraulic Conductivity

Authors: M. Siosemarde, M. Byzedi

Abstract:

Saturated hydraulic conductivity of Soil is an important property in processes involving water and solute flow in soils. Saturated hydraulic conductivity of soil is difficult to measure and can be highly variable, requiring a large number of replicate samples. In this study, 60 sets of soil samples were collected at Saqhez region of Kurdistan province-IRAN. The statistics such as Correlation Coefficient (R), Root Mean Square Error (RMSE), Mean Bias Error (MBE) and Mean Absolute Error (MAE) were used to evaluation the multiple linear regression models varied with number of dataset. In this study the multiple linear regression models were evaluated when only percentage of sand, silt, and clay content (SSC) were used as inputs, and when SSC and bulk density, Bd, (SSC+Bd) were used as inputs. The R, RMSE, MBE and MAE values of the 50 dataset for method (SSC), were calculated 0.925, 15.29, -1.03 and 12.51 and for method (SSC+Bd), were calculated 0.927, 15.28,-1.11 and 12.92, respectively, for relationship obtained from multiple linear regressions on data. Also the R, RMSE, MBE and MAE values of the 10 dataset for method (SSC), were calculated 0.725, 19.62, - 9.87 and 18.91 and for method (SSC+Bd), were calculated 0.618, 24.69, -17.37 and 22.16, respectively, which shows when number of dataset increase, precision of estimated saturated hydraulic conductivity, increases.

Keywords: dataset, precision, saturated hydraulic conductivity, soil and statistics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793