Search results for: load contour method
5817 Experimental Design and Performance Analysis in Plasma Arc Surface Hardening
Authors: M.I.S. Ismail, Z. Taha
Abstract:
In this paper, the experimental design of using the Taguchi method is employed to optimize the processing parameters in the plasma arc surface hardening process. The processing parameters evaluated are arc current, scanning velocity and carbon content of steel. In addition, other significant effects such as the relation between processing parameters are also investigated. An orthogonal array, signal-to-noise (S/N) ratio and analysis of variance (ANOVA) are employed to investigate the effects of these processing parameters. Through this study, not only the hardened depth increased and surface roughness improved, but also the parameters that significantly affect the hardening performance are identified. Experimental results are provided to verify the effectiveness of this approach.Keywords: Plasma arc, hardened depth, surface roughness, Taguchi method, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23605816 The Auto-Tuning PID Controller for Interacting Water Level Process
Authors: Satean Tunyasrirut, Tianchai Suksri, Arjin Numsomran, Supan Gulpanich, Kitti Tirasesth
Abstract:
This paper presents the approach to design the Auto- Tuning PID controller for interactive Water Level Process using integral step response. The Integral Step Response (ISR) is the method to model a dynamic process which can be done easily, conveniently and very efficiently. Therefore this method is advantage for design the auto tune PID controller. Our scheme uses the root locus technique to design PID controller. In this paper MATLAB is used for modeling and testing of the control system. The experimental results of the interacting water level process can be satisfyingly illustrated the transient response and the steady state response.Keywords: Coupled-Tank, Interacting water level process, PIDController, Auto-tuning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23055815 Evaluation of Heterogeneity of Paint Coating on Metal Substrate Using Laser Infrared Thermography and Eddy Current
Authors: S. Mezghani, E. Perrin, J. L Bodnar, J. Marthe, B. Cauwe, V. Vrabie
Abstract:
Non contact evaluation of the thickness of paint coatings can be attempted by different destructive and nondestructive methods such as cross-section microscopy, gravimetric mass measurement, magnetic gauges, Eddy current, ultrasound or terahertz. Infrared thermography is a nondestructive and non-invasive method that can be envisaged as a useful tool to measure the surface thickness variations by analyzing the temperature response. In this paper, the thermal quadrupole method for two layered samples heated up with a pulsed excitation is firstly used. By analyzing the thermal responses as a function of thermal properties and thicknesses of both layers, optimal parameters for the excitation source can be identified. Simulations show that a pulsed excitation with duration of ten milliseconds allows obtaining a substrate-independent thermal response. Based on this result, an experimental setup consisting of a near-infrared laser diode and an Infrared camera was next used to evaluate the variation of paint coating thickness between 60 μm and 130 μm on two samples. Results show that the parameters extracted for thermal images are correlated with the estimated thicknesses by the Eddy current methods. The laser pulsed thermography is thus an interesting alternative nondestructive method that can be moreover used for nonconductive substrates.Keywords: Nondestructive, paint coating, thickness, infrared thermography, laser, heterogeneity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20755814 Reversible Watermarking for H.264/AVC Videos
Authors: Yih-Chuan Lin, Jung-Hong Li
Abstract:
In this paper, we propose a reversible watermarking scheme based on histogram shifting (HS) to embed watermark bits into the H.264/AVC standard videos by modifying the last nonzero level in the context adaptive variable length coding (CAVLC) domain. The proposed method collects all of the last nonzero coefficients (or called last level coefficient) of 4×4 sub-macro blocks in a macro block and utilizes predictions for the current last level from the neighbor block-s last levels to embed watermark bits. The feature of the proposed method is low computational and has the ability of reversible recovery. The experimental results have demonstrated that our proposed scheme has acceptable degradation on video quality and output bit-rate for most test videos.Keywords: Reversible data hiding, H.264/AVC standard, CAVLC, Histogram shifting
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20315813 Resilient Modulus and Deformation Responses of Waste Glass in Flexible Pavement System
Authors: M. Al-Saedi, A. Chegenizadeh, H. Nikraz
Abstract:
Experimental investigations are conducted to assess a layered structure of glass (G) - rock (R) blends under the impact of repeated loading. Laboratory tests included sieve analyses, modified compaction test and repeated load triaxial test (RLTT) is conducted on different structures of stratified GR samples to reach the objectives of this study. Waste materials are such essential components in the climate system, and also commonly used in minimising the need for natural materials in many countries. Glass is one of the most widely used groups of waste materials which have been extensively using in road applications. Full range particle size and colours of glass are collected and mixed at different ratios with natural rock material trying to use the blends in pavement layers. Whole subsurface specimen sequentially consists of a single layer of R and a layer of G-R blend. 12G/88R and 45G/55R mix ratios are employed in this research, the thickness of G-R layer was changed, and the results were compared between the pure rock and the layered specimens. The relations between resilient module (Mr) and permanent deformation with sequence number are presented. During the earlier stages of RLTT, the results indicated that the 45G/55R specimen shows higher moduli than R specimen.
Keywords: Rock base course, layered structure, glass, resilient modulus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6255812 Wet Polymeric Precipitation Synthesis for Monophasic Tricalcium Phosphate
Authors: I. Grigoraviciute-Puroniene, K. Tsuru, E. Garskaite, Z. Stankeviciute, A. Beganskiene, K. Ishikawa, A. Kareiva
Abstract:
Tricalcium phosphate (β-Ca3(PO4)2, β-TCP) powders were synthesized using wet polymeric precipitation method for the first time to our best knowledge. The results of X-ray diffraction analysis showed the formation of almost single a Ca-deficient hydroxyapatite (CDHA) phase of a poor crystallinity already at room temperature. With continuously increasing the calcination temperature up to 800 °C, the crystalline β-TCP was obtained as the main phase. It was demonstrated that infrared spectroscopy is very effective method to characterize the formation of β-TCP. The SEM results showed that β-TCP solids were homogeneous having a small particle size distribution. The β-TCP powders consisted of spherical particles varying in size from 100 to 300 nm. Fabricated β-TCP specimens were placed to the bones of the rats and maintained for 1-2 months.
Keywords: β-TCP, bone regeneration, wet chemical processing, polymeric precipitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10645811 Methodology of Estimating Assembly Cost by MODAPTS
Authors: Heung Jae Cho, Jae Il Park
Abstract:
This paper presents the development of an MODAPTS based cost estimating system to help designers in estimating the manufacturing cost of a assembly products which is belonged from the workers in working fields. Competitiveness of manufacturing cost is getting harder because of the development of Information and telecommunication, but also globalization. Therefore, the accuracy of the assembly cost estimation is getting important. DFA and MODAPTS is useful method for measuring the working hour. But these two methods are used just as a timetable. Therefore, in this paper, we suggest the process of measuring the working hours by MODAPTS which includes the working field-s accurate information. In addition, we adduce the estimation method of accuracy assembly cost with the real information. This research could be useful for designers that can estimate the assembly cost more accurately, and also effective for the companies that which are concerned to reduce the product cost.
Keywords: Cost estimation, DFA, MODAPTS, Assembly cost
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39835810 Neutrosophic Multiple Criteria Decision Making Analysis Method for Selecting Stealth Fighter Aircraft
Authors: C. Ardil
Abstract:
In this paper, a neutrosophic multiple criteria decision analysis method is proposed to select stealth fighter aircraft. Neutrosophic multiple criteria decision analysis methods are used to analyze the neutrosophic environment and give results under uncertainty and incompleteness. Neutrosophic numbers are used to evaluate alternatives over a set of evaluation criteria in decision making problems. Finally, the proposed model is applied to a practical decision problem for selecting stealth fighter aircraft.
Keywords: neutrosophic sets, multiple criteria decision making analysis, stealth fighter aircraft, aircraft selection, MCDMA, SVNNs
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5905809 Effects and Mechanization of a High Gradient Magnetic Separation Process for Particulate and Microbe Removal from Ballast Water
Authors: Zhijun Ren, Zhang Lin, Zhao Ye, Zuo Xiangyu, Mei Dongxing
Abstract:
As a pretreatment process of ballast water treatment, the performance of high gradient magnetic separation (HGMS) technology for the removal of particulates and microorganisms was studied. The results showed that HGMS process could effectively remove suspended particles larger than 5 µm and had ability to resist impact load. Microorganism could also be effectively removed by HGMS process, and the removal effect increased with increasing magnetic field strength. The maximum removal rates for Escherichia coli (E. coli) and Staphylococcus aureus (S. aureus) were 4016.1% and 9675.3% higher, respectively, than without the magnetic field. In addition, the superoxide dismutase (SOD) activity of the microbes decreased by 32.2% when the magnetic field strength was 15.4 mT for 72 min. The microstructure of the stainless steel wool was investigated, and the results showed that particle removal by HGMS has common function by the magnetic force of the high-strength, high-gradient magnetic field on weakly magnetic particles in the water, and on the stainless steel wool.Keywords: HGMS, particulates, superoxide dismutase activity, steel wool magnetic medium.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9275808 Fuzzy Wavelet Packet based Feature Extraction Method for Multifunction Myoelectric Control
Authors: Rami N. Khushaba, Adel Al-Jumaily
Abstract:
The myoelectric signal (MES) is one of the Biosignals utilized in helping humans to control equipments. Recent approaches in MES classification to control prosthetic devices employing pattern recognition techniques revealed two problems, first, the classification performance of the system starts degrading when the number of motion classes to be classified increases, second, in order to solve the first problem, additional complicated methods were utilized which increase the computational cost of a multifunction myoelectric control system. In an effort to solve these problems and to achieve a feasible design for real time implementation with high overall accuracy, this paper presents a new method for feature extraction in MES recognition systems. The method works by extracting features using Wavelet Packet Transform (WPT) applied on the MES from multiple channels, and then employs Fuzzy c-means (FCM) algorithm to generate a measure that judges on features suitability for classification. Finally, Principle Component Analysis (PCA) is utilized to reduce the size of the data before computing the classification accuracy with a multilayer perceptron neural network. The proposed system produces powerful classification results (99% accuracy) by using only a small portion of the original feature set.Keywords: Biomedical Signal Processing, Data mining andInformation Extraction, Machine Learning, Rehabilitation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17375807 The Investigation of Crack's Parameters on the V-Notch using Photoelasticity Method
Authors: M. Saravani, M. Azizi,
Abstract:
The V-notches are most possible case for initiation of cracks in parts. The specifications of cracks on the tip of the notch will be influenced via opening angle, tip radius and depth of V-notch. In this study, the effects of V-notch-s opening angle on stress intensity factor and T-stress of crack on the notch has been investigated. The experiment has been done in different opening angles and various crack length in mode (I) loading using Photoelasticity method. The results illustrate that while angle increases in constant crack-s length, SIF and T-stress will decrease. Beside, the effect of V-notch angle in short crack is more than long crack. These V-notch affects are negligible by increasing the length of crack, and the crack-s behavior can be considered as a single-edge crack specimen. Finally, the results have been evaluated with numerical finite element analysis and good agreement was obvious.
Keywords: Photoelasticity, Stress intensity factor, T-stress, V-notch.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18035806 An Integrated DEMATEL-QFD Model for Medical Supplier Selection
Authors: Mehtap Dursun, Zeynep Şener
Abstract:
Supplier selection is considered as one of the most critical issues encountered by operations and purchasing managers to sharpen the company’s competitive advantage. In this paper, a novel fuzzy multi-criteria group decision making approach integrating quality function deployment (QFD) and decision making trial and evaluation laboratory (DEMATEL) method is proposed for supplier selection. The proposed methodology enables to consider the impacts of inner dependence among supplier assessment criteria. A house of quality (HOQ) which translates purchased product features into supplier assessment criteria is built using the weights obtained by DEMATEL approach to determine the desired levels of supplier assessment criteria. Supplier alternatives are ranked by a distance-based method.
Keywords: DEMATEL, Group decision making, QFD, Supplier selection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28245805 An Improved Construction Method for MIHCs on Cycle Composition Networks
Authors: Hsun Su, Yuan-Kang Shih, Shin-Shin Kao
Abstract:
Many well-known interconnection networks, such as kary n-cubes, recursive circulant graphs, generalized recursive circulant graphs, circulant graphs and so on, are shown to belong to the family of cycle composition networks. Recently, various studies about mutually independent hamiltonian cycles, abbreviated as MIHC-s, on interconnection networks are published. In this paper, using an improved construction method, we obtain MIHC-s on cycle composition networks with a much weaker condition than the known result. In fact, we established the existence of MIHC-s in the cycle composition networks and the result is optimal in the sense that the number of MIHC-s we constructed is maximal.
Keywords: Hamiltonian cycle, k-ary n-cube, cycle composition networks, mutually independent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13895804 A P-SPACE Algorithm for Groebner Bases Computation in Boolean Rings
Authors: Quoc-Nam Tran
Abstract:
The theory of Groebner Bases, which has recently been honored with the ACM Paris Kanellakis Theory and Practice Award, has become a crucial building block to computer algebra, and is widely used in science, engineering, and computer science. It is wellknown that Groebner bases computation is EXP-SPACE in a general setting. In this paper, we give an algorithm to show that Groebner bases computation is P-SPACE in Boolean rings. We also show that with this discovery, the Groebner bases method can theoretically be as efficient as other methods for automated verification of hardware and software. Additionally, many useful and interesting properties of Groebner bases including the ability to efficiently convert the bases for different orders of variables making Groebner bases a promising method in automated verification.Keywords: Algorithm, Complexity, Groebner basis, Applications of Computer Science.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18335803 Robotic Arm Control with Neural Networks Using Genetic Algorithm Optimization Approach
Authors: A. Pajaziti, H. Cana
Abstract:
In this paper, the structural genetic algorithm is used to optimize the neural network to control the joint movements of robotic arm. The robotic arm has also been modeled in 3D and simulated in real-time in MATLAB. It is found that Neural Networks provide a simple and effective way to control the robot tasks. Computer simulation examples are given to illustrate the significance of this method. By combining Genetic Algorithm optimization method and Neural Networks for the given robotic arm with 5 D.O.F. the obtained the results shown that the base joint movements overshooting time without controller was about 0.5 seconds, while with Neural Network controller (optimized with Genetic Algorithm) was about 0.2 seconds, and the population size of 150 gave best results.
Keywords: Robotic Arm, Neural Network, Genetic Algorithm, Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35955802 Influence of Technology Parameters on Properties of AA6061/SiC Composites Produced By Kobo Method
Authors: J. Wozniak, M. Kostecki, K. Broniszewski, W. Bochniak, A. Olszyna
Abstract:
The influence of extrusion parameters on surface quality and properties of AA6061+x% vol. SiC (x = 0; 2,5; 5; 7,5;10) composites was discussed in this paper. The averages size of AA6061 and SiC particles were 10.6 μm and 0.42 μm, respectively. Two series of composites (I - compacts were preheated at extrusion temperature through 0.5 h and cooled by water directly after process; II - compacts were preheated through 3 hours and were not cooled) were consolidated via powder metallurgy processing and extruded by KoBo method. High values of density for both series of composites were achieved. Better surface quality was observed for II series of composites. Moreover, for these composites lower (compared to I series) but more uniform strength properties over the cross-section of the bar were noticed. Microstructure and Young-s modulus investigations were made.Keywords: aluminum alloy, extrusion, metal matrix composites, microstructure
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17495801 Constructing Distinct Kinds of Solutions for the Time-Dependent Coefficients Coupled Klein-Gordon-Schrödinger Equation
Authors: Anupma Bansal
Abstract:
We seek exact solutions of the coupled Klein-Gordon-Schrödinger equation with variable coefficients with the aid of Lie classical approach. By using the Lie classical method, we are able to derive symmetries that are used for reducing the coupled system of partial differential equations into ordinary differential equations. From reduced differential equations we have derived some new exact solutions of coupled Klein-Gordon-Schrödinger equations involving some special functions such as Airy wave functions, Bessel functions, Mathieu functions etc.
Keywords: Klein-Gordon-Schödinger Equation, Lie Classical Method, Exact Solutions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 45305800 Adaptive WiFi Fingerprinting for Location Approximation
Authors: Mohd Fikri Azli bin Abdullah, Khairul Anwar bin Kamarul Hatta, Esther Jeganathan
Abstract:
WiFi has become an essential technology that is widely used nowadays. It is famous due to its convenience to be used with mobile devices. This is especially true for Internet users worldwide that use WiFi connections. There are many location based services that are available nowadays which uses Wireless Fidelity (WiFi) signal fingerprinting. A common example that is gaining popularity in this era would be Foursquare. In this work, the WiFi signal would be used to estimate the user or client’s location. Similar to GPS, fingerprinting method needs a floor plan to increase the accuracy of location estimation. Still, the factor of inconsistent WiFi signal makes the estimation defer at different time intervals. Given so, an adaptive method is needed to obtain the most accurate signal at all times. WiFi signals are heavily distorted by external factors such as physical objects, radio frequency interference, electrical interference, and environmental factors to name a few. Due to these factors, this work uses a method of reducing the signal noise and estimation using the Nearest Neighbour based on past activities of the signal to increase the signal accuracy up to more than 80%. The repository yet increases the accuracy by using Artificial Neural Network (ANN) pattern matching. The repository acts as the server cum support of the client side application decision. Numerous previous works has adapted the methods of collecting signal strengths in the repository over the years, but mostly were just static. In this work, proposed solutions on how the adaptive method is done to match the signal received to the data in the repository are highlighted. With the said approach, location estimation can be done more accurately. Adaptive update allows the latest location fingerprint to be stored in the repository. Furthermore, any redundant location fingerprints are removed and only the updated version of the fingerprint is stored in the repository. How the location estimation of the user can be predicted would be highlighted more in the proposed solution section. After some studies on previous works, it is found that the Artificial Neural Network is the most feasible method to deploy in updating the repository and making it adaptive. The Artificial Neural Network functions are to do the pattern matching of the WiFi signal to the existing data available in the repository.
Keywords: Adaptive Repository, Artificial Neural Network, Location Estimation, Nearest Neighbour Euclidean Distance, WiFi RSSI Fingerprinting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34605799 Cryogenic Freezing Process Optimization Based On Desirability Function on the Path of Steepest Ascent
Authors: R. Uporn, P. Luangpaiboon
Abstract:
This paper presents a comparative study of statistical methods for the multi-response surface optimization of a cryogenic freezing process. Taguchi design and analysis and steepest ascent methods based on the desirability function were conducted to ascertain the influential factors of a cryogenic freezing process and their optimal levels. The more preferable levels of the set point, exhaust fan speed, retention time and flow direction are set at -90oC, 20 Hz, 18 minutes and Counter Current, respectively. The overall desirability level is 0.7044.
Keywords: Cryogenic Freezing Process, Taguchi Design and Analysis, Response Surface Method, Steepest Ascent Method and Desirability Function Approach.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18325798 Despeckling of Synthetic Aperture Radar Images Using Inner Product Spaces in Undecimated Wavelet Domain
Authors: Syed Musharaf Ali, Muhammad Younus Javed, Naveed Sarfraz Khattak, Athar Mohsin, UmarFarooq
Abstract:
This paper introduces the effective speckle reduction of synthetic aperture radar (SAR) images using inner product spaces in undecimated wavelet domain. There are two major areas in projection onto span algorithm where improvement can be made. First is the use of undecimated wavelet transformation instead of discrete wavelet transformation. And second area is the use of smoothing filter namely directional smoothing filter which is an additional step. Proposed method does not need any noise estimation and thresholding technique. More over proposed method gives good results on both single polarimetric and fully polarimetric SAR images.Keywords: Directional Smoothing, Inner product, Length ofvector, Undecimated wavelet transformation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16115797 An Alternative Method for Generating Almost Infinite Sequence of Gaussian Variables
Authors: Nyah C. Temaneh, F. A. Phiri, E. Ruhunga
Abstract:
Most of the well known methods for generating Gaussian variables require at least one standard uniform distributed value, for each Gaussian variable generated. The length of the random number generator therefore, limits the number of independent Gaussian distributed variables that can be generated meanwhile the statistical solution of complex systems requires a large number of random numbers for their statistical analysis. We propose an alternative simple method of generating almost infinite number of Gaussian distributed variables using a limited number of standard uniform distributed random numbers.Keywords: Gaussian variable, statistical analysis, simulation ofCommunication Network, Random numbers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14725796 Blind Source Separation for Convoluted Signals Based on Properties of Acoustic Transfer Function in Real Environments
Authors: Takaaki Ishibashi
Abstract:
Frequency domain independent component analysis has a scaling indeterminacy and a permutation problem. The scaling indeterminacy can be solved by use of a decomposed spectrum. For the permutation problem, we have proposed the rules in terms of gain ratio and phase difference derived from the decomposed spectra and the source-s coarse directions. The present paper experimentally clarifies that the gain ratio and the phase difference work effectively in a real environment but their performance depends on frequency bands, a microphone-space and a source-microphone distance. From these facts it is seen that it is difficult to attain a perfect solution for the permutation problem in a real environment only by either the gain ratio or the phase difference. For the perfect solution, this paper gives a solution to the problems in a real environment. The proposed method is simple, the amount of calculation is small. And the method has high correction performance without depending on the frequency bands and distances from source signals to microphones. Furthermore, it can be applied under the real environment. From several experiments in a real room, it clarifies that the proposed method has been verified.Keywords: blind source separation, frequency domain independent component analysys, permutation correction, scale adjustment, target extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14395795 A Study on Improving the Flow Capacity of the Valves
Authors: A. G. Pradeep, Gorantla Giridhar Kumar, Vijay Turaga, Vinod Srinivasa
Abstract:
The major problem in the flow control valve is of lower Flow Capacity (Cv) which will reduce overall efficiency of flow circuit. Designers are continuously working to improve the Cv of the valve, but they need to validate the design ideas they have regarding the improvement of Cv. Traditional method of prototype and testing take a lot of time, that is where CFD comes into picture with very quick and accurate validation along with the visualization which is not possible with traditional testing method. We have developed a method to predict Cv value using CFD analysis by iterating on various Boundary conditions, solver settings and by carrying out grid convergence studies to establish correlation between the CFD model and Test data. The present study investigates 3 different ideas put forward by the designers for improving the flow capacity of the valves like reducing the cage thickness, changing the port position, and using the parabolic plug to guide the flow. Using CFD, we analyzed all design changes using the established methodology that we developed. We were able to evaluate the effect of these design changes on the Valve Cv. We optimized the wetted surface of the valve further by suggesting the design modification to the lower part of the valve to make the flow more streamlined. We could find that changing cage thickness and port position has little impact on the valve Cv. Combination of optimized wetted surface and introduction of parabolic plug improved the Cv of the valve significantly.
Keywords: Flow control valves, flow capacity, CFD simulations, design validation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4385794 Performance Variation of the TEES According to the Changes in Cold-Side Storage Temperature
Authors: Young-Jin Baik, Minsung Kim, Junhyun Cho, Ho-Sang Ra, Young-Soo Lee, Ki-Chang Chang
Abstract:
Surplus electricity can be converted into potential energy via pumped hydroelectric storage for future usage. Similarly, thermo-electric energy storage (TEES) uses heat pumps equipped with thermal storage to convert electrical energy into thermal energy; the stored energy is then converted back into electrical energy when necessary using a heat engine. The greatest advantage of this method is that, unlike pumped hydroelectric storage and compressed air energy storage, TEES is not restricted by geographical constraints. In this study, performance variation of the TEES according to the changes in cold-side storage temperature was investigated by simulation method.
Keywords: Energy Storage System, Heat Pump.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18505793 Expelling Policy Based Buffer Control during Congestion in Differentiated Service Routers
Authors: Kumar Padmanabh, Rajarshi Roy
Abstract:
In this paper a special kind of buffer management policy is studied where the packet are preempted even when sufficient space is available in the buffer for incoming packets. This is done to congestion for future incoming packets to improve QoS for certain type of packets. This type of study has been done in past for ATM type of scenario. We extend the same for heterogeneous traffic where data rate and size of the packets are very versatile in nature. Typical example of this scenario is the buffer management in Differentiated Service Router. There are two aspects that are of interest. First is the packet size: whether all packets have same or different sizes. Second aspect is the value or space priority of the packets, do all packets have the same space priority or different packets have different space priorities. We present two types of policies to achieve QoS goals for packets with different priorities: the push out scheme and the expelling scheme. For this work the scenario of packets of variable length is considered with two space priorities and main goal is to minimize the total weighted packet loss. Simulation and analytical studies show that, expelling policies can outperform the push out policies when it comes to offering variable QoS for packets of two different priorities and expelling policies also help improve the amount of admissible load. Some other comparisons of push out and expelling policies are also presented using simulations.
Keywords: Buffer Management Policy, Diffserv, ATM, Pushout Policy, Expeling Policy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13955792 A Pragmatics Study on Jokes and the Implicature in Broadcast Messages
Authors: Yuli Widiana
Abstract:
The study of implicature which is one of the discussions of pragmatics is such an interesting and challenging topic to discuss. Implicature is such a meaning which is implied in such an utterance which is not the same as its literal meaning. The rapid development of information technology results social networks as media to broadcast messages. The broadcast messages may be in the form of jokes which contain implicature. The research applies the pragmatic equivalent method to analyze the topics of jokes based on the implicatures contained in them. Furthermore, the method is also applied to reveal the purpose of creating implicature in jokes. The findings include the kinds of implicature found in jokes which are classified into conventional implicature and conversational implicature. Then, in detailed analysis, implicature in jokes is divided into implicature related to gender, culture, and social phenomena. Furthermore, implicature in jokes may not only be used to give entertainment but also to soften criticisms or satire so that it does not sound rude and harsh.Keywords: Implicature, broadcast messages, conventional implicature, conversational implicature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32765791 Normalization Discriminant Independent Component Analysis
Authors: Liew Yee Ping, Pang Ying Han, Lau Siong Hoe, Ooi Shih Yin, Housam Khalifa Bashier Babiker
Abstract:
In face recognition, feature extraction techniques attempts to search for appropriate representation of the data. However, when the feature dimension is larger than the samples size, it brings performance degradation. Hence, we propose a method called Normalization Discriminant Independent Component Analysis (NDICA). The input data will be regularized to obtain the most reliable features from the data and processed using Independent Component Analysis (ICA). The proposed method is evaluated on three face databases, Olivetti Research Ltd (ORL), Face Recognition Technology (FERET) and Face Recognition Grand Challenge (FRGC). NDICA showed it effectiveness compared with other unsupervised and supervised techniques.
Keywords: Face recognition, small sample size, regularization, independent component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19545790 Development of User Interface for Multiple Devices Connecting Path Planning System for Bus Network
Authors: Takahiro Takayama, Takao Kawamura, Toshihiko Sasama, Kazunori Sugahara
Abstract:
Recently, web services to access from many type devices are often used. We have developed the shortest path planning system called "Bus-Net" in Tottori prefecture as a web application to sustain the public transport. And it used the same user interface for both devices. To support both devices, the interface cannot use JavaScript and so on. Thus, we developed the method that use individual user interface for each device type to improve its convenience. To be concrete, we defined formats of condition input to the path planning system and result output from it and separate the system into the request processing part and user interface parts that depend on device types. By this method, we have also developed special device for Bus-Net named "Intelligent-Bus-Stop".Keywords: Bus, Path planning, Public transport, User interface
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14835789 Robust Detection of R-Wave Using Wavelet Technique
Authors: Awadhesh Pachauri, Manabendra Bhuyan
Abstract:
Electrocardiogram (ECG) is considered to be the backbone of cardiology. ECG is composed of P, QRS & T waves and information related to cardiac diseases can be extracted from the intervals and amplitudes of these waves. The first step in extracting ECG features starts from the accurate detection of R peaks in the QRS complex. We have developed a robust R wave detector using wavelets. The wavelets used for detection are Daubechies and Symmetric. The method does not require any preprocessing therefore, only needs the ECG correct recordings while implementing the detection. The database has been collected from MIT-BIH arrhythmia database and the signals from Lead-II have been analyzed. MatLab 7.0 has been used to develop the algorithm. The ECG signal under test has been decomposed to the required level using the selected wavelet and the selection of detail coefficient d4 has been done based on energy, frequency and cross-correlation analysis of decomposition structure of ECG signal. The robustness of the method is apparent from the obtained results.Keywords: ECG, P-QRS-T waves, Wavelet Transform, Hard Thresholding, R-wave Detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24745788 Evaluation of Dynamic Behavior of a Rotor-Bearing System in Operating Conditions
Authors: Mohammad Hadi Jalali, Behrooz Shahriari, Mostafa Ghayour, Saeed Ziaei-Rad, Shahram Yousefi
Abstract:
Most flexible rotors can be considered as beam-like structures. In many cases, rotors are modeled as one-dimensional bodies, made basically of beam-like shafts with rigid bodies attached to them. This approach is typical of rotor dynamics, both analytical and numerical, and several rotor dynamic codes, based on the finite element method, follow this trend. In this paper, a finite element model based on Timoshenko beam elements is utilized to analyze the lateral dynamic behavior of a certain rotor-bearing system in operating conditions.
Keywords: Finite element method, Operational deflection shape, Timoshenko beam elements, Unbalance response.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3010