Search results for: Minimum spanning tree
272 Current-Mode Resistorless SIMO Universal Filter and Four-Phase Quadrature Oscillator
Authors: Jie Jin
Abstract:
In this paper, a new CMOS current-mode single input and multi-outputs (SIMO) universal filter and quadrature oscillator with a similar circuit are proposed. The circuits only consist of three Current differencing transconductance amplifiers (CDTA) and two grounded capacitors, which are resistorless, and they are suitable for monolithic integration. The universal filter uses minimum CDTAs and passive elements to realize SIMO type low-pass (LP), high-pass (HP), band-pass (BP) band-stop (BS) and all-pass (AP) filter functions simultaneously without any component matching conditions. The angular frequency (ω0) and the quality factor (Q) of the proposed filter can be electronically controlled and tuned orthogonal. By some modifications of the filter, a new current-mode four-phase quadrature oscillator (QO) can be obtained easily. The condition of oscillation (CO) and frequency of oscillation (FO) of the QO can be controlled electronically and independently through the bias current of the CDTAs, and it is suitable for variable frequency oscillator. Moreover, all the passive and active sensitivities of the circuits are low. SPICE simulation results are included to confirm the theory.
Keywords: Universal Filter, Quadrature Oscillator, Current mode, Current differencing transconductance amplifiers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951271 Numerical Analysis of Wave and Hydrodynamic Models for Energy Balance and Primitive Equations
Authors: Worachat Wannawong, Usa W. Humphries, Prungchan Wongwises, Suphat Vongvisessomjai, Wiriya Lueangaram
Abstract:
A numerical analysis of wave and hydrodynamic models is used to investigate the influence of WAve and Storm Surge (WASS) in the regional and coastal zones. The numerical analyzed system consists of the WAve Model Cycle 4 (WAMC4) and the Princeton Ocean Model (POM) which used to solve the energy balance and primitive equations respectively. The results of both models presented the incorporated surface wave in the regional zone affected the coastal storm surge zone. Specifically, the results indicated that the WASS generally under the approximation is not only the peak surge but also the coastal water level drop which can also cause substantial impact on the coastal environment. The wave–induced surface stress affected the storm surge can significantly improve storm surge prediction. Finally, the calibration of wave module according to the minimum error of the significant wave height (Hs) is not necessarily result in the optimum wave module in the WASS analyzed system for the WASS prediction.Keywords: energy balance equation, numerical analysis, primitiveequation, storm surge, wave.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1939270 Experimental Analysis of the Plate-on-Tube Evaporator on a Domestic Refrigerator’s Performance
Authors: Mert Tosun, Tuğba Tosun
Abstract:
The evaporator is the utmost important component in the refrigeration system, since it enables the refrigerant to draw heat from the desired environment, i.e. the refrigerated space. Studies are being conducted on this component which generally affects the performance of the system, where energy efficient products are important. This study was designed to enhance the effectiveness of the evaporator in the refrigeration cycle of a domestic refrigerator by adjusting the capillary tube length, refrigerant amount, and the evaporator pipe diameter to reduce energy consumption. The experiments were conducted under identical thermal and ambient conditions. Experiment data were analysed using the Design of Experiment (DOE) technique which is a six-sigma method to determine effects of parameters. As a result, it has been determined that the most important parameters affecting the evaporator performance among the selected parameters are found to be the refrigerant amount and pipe diameter. It has been determined that the minimum energy consumption is 6-mm pipe diameter and 16-g refrigerant. It has also been noted that the overall consumption of the experiment sample decreased by 16.6% with respect to the reference system, which has 7-mm pipe diameter and 18-g refrigerant.
Keywords: Heat exchanger, refrigerator, design of experiment, energy consumption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 603269 Optimal Channel Equalization for MIMO Time-Varying Channels
Authors: Ehab F. Badran, Guoxiang Gu
Abstract:
We consider optimal channel equalization for MIMO (multi-input/multi-output) time-varying channels in the sense of MMSE (minimum mean-squared-error), where the observation noise can be non-stationary. We show that all ZF (zero-forcing) receivers can be parameterized in an affine form which eliminates completely the ISI (inter-symbol-interference), and optimal channel equalizers can be designed through minimization of the MSE (mean-squarederror) between the detected signals and the transmitted signals, among all ZF receivers. We demonstrate that the optimal channel equalizer is a modified Kalman filter, and show that under the AWGN (additive white Gaussian noise) assumption, the proposed optimal channel equalizer minimizes the BER (bit error rate) among all possible ZF receivers. Our results are applicable to optimal channel equalization for DWMT (discrete wavelet multitone), multirate transmultiplexers, OFDM (orthogonal frequency division multiplexing), and DS (direct sequence) CDMA (code division multiple access) wireless data communication systems. A design algorithm for optimal channel equalization is developed, and several simulation examples are worked out to illustrate the proposed design algorithm.Keywords: Channel equalization, Kalman filtering, Time-varying systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1834268 Dynamics and Driving Forces of the Alpine Wetlands in the Yarlung Zangbo River Basin of Tibet, China
Authors: Weishou Shen, Dong Liu, Di Ji, Haoyun Shen, Naifeng Lin
Abstract:
Based on the field investigation and long term remote sensing data, the dynamics of the alpine wetland in the river basin and their response to climate change were studied. Results showed the alpine wetlands accounted for 3.73% of total basin in 2010. Lake and river appeared an increasing trend in the past 30 years, with an increase of 34.36 % and 24.57%. However, swamp exhibited a tendency of decreasing with 233.74 km2. Annual average temperature, maximum temperature, minimum temperature and precipitation in the river basin all exhibited an increasing trend, whereas relative humidity exhibited a decreasing trend. Ice and snow melting are main reasons of lake and river area enhancement and swamp area descend. There existed 91.78%-97.86% of reduced swamp converted into lakes on the basis of remote sensing image interpretation. China-s government policy of implementing development in the river basin is the major driving force of artificial wetland growth.Keywords: alpine wetland dynamics, climate change, Yarlung Zangbo River basin
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1654267 Optimum Design of Tall Tube-Type Building: An Approach to Structural Height Premium
Authors: Ali Kheyroddin, Niloufar Mashhadiali, Frazaneh Kheyroddin
Abstract:
In last decades, tubular systems employed for tall buildings were efficient structural systems. However, increasing the height of a building leads to an increase in structural material corresponding to the loads imposed by lateral loads. Based on this approach, new structural systems are emerging to provide strength and stiffness with the minimum premium for height. In this research, selected tube-type structural systems such as framed tubes, braced tubes, diagrids and hexagrid systems were applied as a single tube, tubular structures combined with braced core and outrigger trusses on a set of 48, 72, and 96-story, respectively, to improve integrated structural systems. This paper investigated structural material consumption by model structures focusing on the premium for height. Compared analytical results indicated that as the height of the building increased, combination of the structural systems caused the framed tube, hexagrid and braced tube system to pay fewer premiums to material tonnage while in diagrid system, combining the structural system reduced insignificantly the steel material consumption.
Keywords: Braced tube, diagrid, framed tube, hexagrid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1120266 Energy Efficient Reliable Cooperative Multipath Routing in Wireless Sensor Networks
Authors: Gergely Treplan, Long Tran-Thanh, Janos Levendovszky
Abstract:
In this paper, a reliable cooperative multipath routing algorithm is proposed for data forwarding in wireless sensor networks (WSNs). In this algorithm, data packets are forwarded towards the base station (BS) through a number of paths, using a set of relay nodes. In addition, the Rayleigh fading model is used to calculate the evaluation metric of links. Here, the quality of reliability is guaranteed by selecting optimal relay set with which the probability of correct packet reception at the BS will exceed a predefined threshold. Therefore, the proposed scheme ensures reliable packet transmission to the BS. Furthermore, in the proposed algorithm, energy efficiency is achieved by energy balancing (i.e. minimizing the energy consumption of the bottleneck node of the routing path) at the same time. This work also demonstrates that the proposed algorithm outperforms existing algorithms in extending longevity of the network, with respect to the quality of reliability. Given this, the obtained results make possible reliable path selection with minimum energy consumption in real time.Keywords: wireless sensor networks, reliability, cooperativerouting, Rayleigh fading model, energy balancing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610265 Optimization of Quercus cerris Bark Liquefaction
Authors: Luísa P. Cruz-Lopes, Hugo Costa e Silva, Idalina Domingos, José Ferreira, Luís Teixeira de Lemos, Bruno Esteves
Abstract:
The liquefaction process of cork based tree barks has led to an increase of interest due to its potential innovation in the lumber and wood industries. In this particular study the bark of Quercus cerris (Turkish oak) is used due to its appreciable amount of cork tissue, although of inferior quality when compared to the cork provided by other Quercus trees. This study aims to optimize alkaline catalysis liquefaction conditions, regarding several parameters. To better comprehend the possible chemical characteristics of the bark of Quercus cerris, a complete chemical analysis was performed. The liquefaction process was performed in a double-jacket reactor heated with oil, using glycerol and a mixture of glycerol/ethylene glycol as solvents, potassium hydroxide as a catalyst, and varying the temperature, liquefaction time and granulometry. Due to low liquefaction efficiency resulting from the first experimental procedures a study was made regarding different washing techniques after the filtration process using methanol and methanol/water. The chemical analysis stated that the bark of Quercus cerris is mostly composed by suberin (ca. 30%) and lignin (ca. 24%) as well as insolvent hemicelluloses in hot water (ca. 23%). On the liquefaction stage, the results that led to higher yields were: using a mixture of methanol/ethylene glycol as reagents and a time and temperature of 120 minutes and 200 ºC, respectively. It is concluded that using a granulometry of <80 mesh leads to better results, even if this parameter barely influences the liquefaction efficiency. Regarding the filtration stage, washing the residue with methanol and then distilled water leads to a considerable increase on final liquefaction percentages, which proves that this procedure is effective at liquefying suberin content and lignocellulose fraction.Keywords: Liquefaction, alkaline catalysis, optimization, Quercus cerris bark.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491264 Manipulation of Probiotics Fermentation of Yogurt by Cinnamon and Licorice: Effects on Yogurt Formation and Inhibition of Helicobacter Pylori Growth in vitro
Authors: S. Behrad, M.Y. Yusof, K. L. Goh, A.S. Baba
Abstract:
Probiotic bacteria especially Lactobacillus spp. and Bifidobacterium exert suppressive effect on Helicobacter pylori. Cinnamon and licorice have been traditionally used for the treatment of gastric ulcer. The objectives of this study were to determine the effects of herbs on yogurt fermentation, the level of probiotic bacteria in yogurt during 28 days storage and the effect of herbal yogurt on the growth of H. pylori in vitro. Cinnamon or licorice was mixed with milk and the mixture was fermented with probiotic bacteria to form herbal-yogurt. Changes of pH and total titratable acids were monitored and the viability of probiotic bacteria was evaluated during and after refrigerated storage. The in vitro inhibition of H. pylori growth was determined using agar diffusion and minimum inhibitory concentration (MIC) method. The presence of herbs did not affect the probiotic population during storage. There were no significant differences in pH and TTA between herbal-yogurts and plain-yogurt during fermentation and storage. Water extract of cinnamon-yogurt showed the highest inhibition effect (13.5mm) on H. pylori growth in comparison with licorice-yogurt (11.2mm). The present findings indicate cinnamon and licorice has bioactive components to decrease the growth of H. pylori.
Keywords: Cinnamon, Helicobacter pylori, Herbal-Yogurt, Licorice, Probiotics
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3684263 Shape Optimization of Permanent Magnet Motors Using the Reduced Basis Technique
Authors: A. Jabbari, M. Shakeri, A. Nabavi
Abstract:
In this paper, a tooth shape optimization method for cogging torque reduction in Permanent Magnet (PM) motors is developed by using the Reduced Basis Technique (RBT) coupled by Finite Element Analysis (FEA) and Design of Experiments (DOE) methods. The primary objective of the method is to reduce the enormous number of design variables required to define the tooth shape. RBT is a weighted combination of several basis shapes. The aim of the method is to find the best combination using the weights for each tooth shape as the design variables. A multi-level design process is developed to find suitable basis shapes or trial shapes at each level that can be used in the reduced basis technique. Each level is treated as a separated optimization problem until the required objective – minimum cogging torque – is achieved. The process is started with geometrically simple basis shapes that are defined by their shape co-ordinates. The experimental design of Taguchi method is used to build the approximation model and to perform optimization. This method is demonstrated on the tooth shape optimization of a 8-poles/12-slots PM motor.Keywords: PM motor, cogging torque, tooth shape optimization, RBT, FEA, DOE.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2503262 Reliability Analysis of P-I Diagram Formula for RC Column Subjected to Blast Load
Authors: Masoud Abedini, Azrul A. Mutalib, Shahrizan Baharom, Hong Hao
Abstract:
This study was conducted published to investigate there liability of the equation pressure-impulse (PI) reinforced concrete column inprevious studies. Equation involves three different levels of damage criteria known as D =0. 2, D =0. 5 and D =0. 8.The damage criteria known as a minor when 0-0.2, 0.2-0.5is known as moderate damage, high damage known as 0.5-0.8, and 0.8-1 of the structure is considered a failure. In this study, two types of reliability analyzes conducted. First, using pressure-impulse equation with different parameters. The parameters involved are the concrete strength, depth, width, and height column, the ratio of longitudinal reinforcement and transverse reinforcement ratio. In the first analysis of the reliability of this new equation is derived to improve the previous equations. The second reliability analysis involves three types of columns used to derive the PI curve diagram using the derived equation to compare with the equation derived from other researchers and graph minimum standoff versus weapon yield Federal Emergency Management Agency (FEMA). The results showed that the derived equation is more accurate with FEMA standards than previous researchers.
Keywords: Blast load, RC column, P-I curve, Analytical formulae, Standard FEMA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2912261 Transient Stability Assessment Using Fuzzy SVM and Modified Preventive Control
Authors: B. Dora Arul Selvi, .N. Kamaraj
Abstract:
Transient Stability is an important issue in power systems planning, operation and extension. The objective of transient stability analysis problem is not satisfied with mere transient instability detection or evaluation and it is most important to complement it by defining fast and efficient control measures in order to ensure system security. This paper presents a new Fuzzy Support Vector Machines (FSVM) to investigate the stability status of power systems and a modified generation rescheduling scheme to bring back the identified unstable cases to a more economical and stable operating point. FSVM improves the traditional SVM (Support Vector Machines) by adding fuzzy membership to each training sample to indicate the degree of membership of this sample to different classes. The preventive control based on economic generator rescheduling avoids the instability of the power systems with minimum change in operating cost under disturbed conditions. Numerical results on the New England 39 bus test system show the effectiveness of the proposed method.
Keywords: Fuzzy Support Vector Machine (FSVM), Incremental Cost, Preventive Control, Transient stability
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491260 Network Reconfiguration for Load Balancing in Distribution System with Distributed Generation and Capacitor Placement
Authors: T. Lantharthong, N. Rugthaicharoencheep
Abstract:
This paper presents an efficient algorithm for optimization of radial distribution systems by a network reconfiguration to balance feeder loads and eliminate overload conditions. The system load-balancing index is used to determine the loading conditions of the system and maximum system loading capacity. The index value has to be minimum in the optimal network reconfiguration of load balancing. A method based on Tabu search algorithm, The Tabu search algorithm is employed to search for the optimal network reconfiguration. The basic idea behind the search is a move from a current solution to its neighborhood by effectively utilizing a memory to provide an efficient search for optimality. It presents low computational effort and is able to find good quality configurations. Simulation results for a radial 69-bus system with distributed generations and capacitors placement. The study results show that the optimal on/off patterns of the switches can be identified to give the best network reconfiguration involving balancing of feeder loads while respecting all the constraints.Keywords: Network reconfiguration, Distributed generation Capacitor placement, Load balancing, Optimization technique
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4219259 A New Approach to Face Recognition Using Dual Dimension Reduction
Authors: M. Almas Anjum, M. Younus Javed, A. Basit
Abstract:
In this paper a new approach to face recognition is presented that achieves double dimension reduction, making the system computationally efficient with better recognition results and out perform common DCT technique of face recognition. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results change with change in face image resolution and provide optimal results when arriving at a certain resolution level. In the proposed model of face recognition, initially image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to increased computational speed and feature extraction potential of Discrete Cosine Transform (DCT), it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A tradeoff between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL , Yale and EME color database.Keywords: Biometrics, DCT, Face Recognition, Illumination, Computation, Feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1686258 Adapting Tools for Text Monitoring and for Scenario Analysis Related to the Field of Social Disasters
Authors: Svetlana Cojocaru, Mircea Petic, Inga Titchiev
Abstract:
Humanity faces more and more often with different social disasters, which in turn can generate new accidents and catastrophes. To mitigate their consequences, it is important to obtain early possible signals about the events which are or can occur and to prepare the corresponding scenarios that could be applied. Our research is focused on solving two problems in this domain: identifying signals related that an accident occurred or may occur and mitigation of some consequences of disasters. To solve the first problem, methods of selecting and processing texts from global network Internet are developed. Information in Romanian is of special interest for us. In order to obtain the mentioned tools, we should follow several steps, divided into preparatory stage and processing stage. Throughout the first stage, we manually collected over 724 news articles and classified them into 10 categories of social disasters. It constitutes more than 150 thousand words. Using this information, a controlled vocabulary of more than 300 keywords was elaborated, that will help in the process of classification and identification of the texts related to the field of social disasters. To solve the second problem, the formalism of Petri net has been used. We deal with the problem of inhabitants’ evacuation in useful time. The analysis methods such as reachability or coverability tree and invariants technique to determine dynamic properties of the modeled systems will be used. To perform a case study of properties of extended evacuation system by adding time, the analysis modules of PIPE such as Generalized Stochastic Petri Nets (GSPN) Analysis, Simulation, State Space Analysis, and Invariant Analysis have been used. These modules helped us to obtain the average number of persons situated in the rooms and the other quantitative properties and characteristics related to its dynamics.Keywords: Lexicon of disasters, modelling, Petri nets, text annotation, social disasters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1157257 Determination of Safety Distance Around Gas Pipelines Using Numerical Methods
Authors: Omid Adibi, Nategheh Najafpour, Bijan Farhanieh, Hossein Afshin
Abstract:
Energy transmission pipelines are one of the most vital parts of each country which several strict laws have been conducted to enhance the safety of these lines and their vicinity. One of these laws is the safety distance around high pressure gas pipelines. Safety distance refers to the minimum distance from the pipeline where people and equipment do not confront with serious damages. In the present study, safety distance around high pressure gas transmission pipelines were determined by using numerical methods. For this purpose, gas leakages from cracked pipeline and created jet fires were simulated as continuous ignition, three dimensional, unsteady and turbulent cases. Numerical simulations were based on finite volume method and turbulence of flow was considered using k-ω SST model. Also, the combustion of natural gas and air mixture was applied using the eddy dissipation method. The results show that, due to the high pressure difference between pipeline and environment, flow chocks in the cracked area and velocity of the exhausted gas reaches to sound speed. Also, analysis of the incident radiation results shows that safety distances around 42 inches high pressure natural gas pipeline based on 5 and 15 kW/m2 criteria are 205 and 272 meters, respectively.
Keywords: Gas pipelines, incident radiation, numerical simulation, safety distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1133256 Fuzzy Mathematical Morphology approach in Image Processing
Authors: Yee Yee Htun, Dr. Khaing Khaing Aye
Abstract:
Morphological operators transform the original image into another image through the interaction with the other image of certain shape and size which is known as the structure element. Mathematical morphology provides a systematic approach to analyze the geometric characteristics of signals or images, and has been applied widely too many applications such as edge detection, objection segmentation, noise suppression and so on. Fuzzy Mathematical Morphology aims to extend the binary morphological operators to grey-level images. In order to define the basic morphological operations such as fuzzy erosion, dilation, opening and closing, a general method based upon fuzzy implication and inclusion grade operators is introduced. The fuzzy morphological operations extend the ordinary morphological operations by using fuzzy sets where for fuzzy sets, the union operation is replaced by a maximum operation, and the intersection operation is replaced by a minimum operation. In this work, it consists of two articles. In the first one, fuzzy set theory, fuzzy Mathematical morphology which is based on fuzzy logic and fuzzy set theory; fuzzy Mathematical operations and their properties will be studied in details. As a second part, the application of fuzziness in Mathematical morphology in practical work such as image processing will be discussed with the illustration problems.Keywords: Binary Morphological, Fuzzy sets, Grayscalemorphology, Image processing, Mathematical morphology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3247255 Statistical Distributions of the Lapped Transform Coefficients for Images
Authors: Vijay Kumar Nath, Deepika Hazarika, Anil Mahanta,
Abstract:
Discrete Cosine Transform (DCT) based transform coding is very popular in image, video and speech compression due to its good energy compaction and decorrelating properties. However, at low bit rates, the reconstructed images generally suffer from visually annoying blocking artifacts as a result of coarse quantization. Lapped transform was proposed as an alternative to the DCT with reduced blocking artifacts and increased coding gain. Lapped transforms are popular for their good performance, robustness against oversmoothing and availability of fast implementation algorithms. However, there is no proper study reported in the literature regarding the statistical distributions of block Lapped Orthogonal Transform (LOT) and Lapped Biorthogonal Transform (LBT) coefficients. This study performs two goodness-of-fit tests, the Kolmogorov-Smirnov (KS) test and the 2- test, to determine the distribution that best fits the LOT and LBT coefficients. The experimental results show that the distribution of a majority of the significant AC coefficients can be modeled by the Generalized Gaussian distribution. The knowledge of the statistical distribution of transform coefficients greatly helps in the design of optimal quantizers that may lead to minimum distortion and hence achieve optimal coding efficiency.
Keywords: Lapped orthogonal transform, Lapped biorthogonal transform, Image compression, KS test,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1605254 Optimizing Materials Cost and Mechanical Properties of PVC Electrical Cable-s Insulation by Using Mixture Experimental Design Approach
Authors: Safwan Altarazi, Raghad Hemeimat, Mousa Wakileh, Ra'ad Qsous, Aya Khreisat
Abstract:
With the development of the Polyvinyl chloride (PVC) products in many applications, the challenge of investigating the raw material composition and reducing the cost have both become more and more important. Considerable research has been done investigating the effect of additives on the PVC products. Most of the PVC composites research investigates only the effect of single/few factors, at a time. This isolated consideration of the input factors does not take in consideration the interaction effect of the different factors. This paper implements a mixture experimental design approach to find out a cost-effective PVC composition for the production of electrical-insulation cables considering the ASTM Designation (D) 6096. The results analysis showed that a minimum cost can be achieved through using 20% virgin PVC, 18.75% recycled PVC, 43.75% CaCO3 with participle size 10 microns, 14% DOP plasticizer, and 3.5% CPW plasticizer. For maximum UTS the compound should consist of: 17.5% DOP, 62.5% virgin PVC, and 20.0% CaCO3 of particle size 5 microns. Finally, for the highest ductility the compound should be made of 35% virgin PVC, 20% CaCO3 of particle size 5 microns, and 45.0% DOP plasticizer.Keywords: ASTM 6096, mixture experimental-design approach, PVC electrical cable insulation, recycled PVC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4705253 Effective Defect Prevention Approach in Software Process for Achieving Better Quality Levels
Authors: Suma. V., T. R. Gopalakrishnan Nair
Abstract:
Defect prevention is the most vital but habitually neglected facet of software quality assurance in any project. If functional at all stages of software development, it can condense the time, overheads and wherewithal entailed to engineer a high quality product. The key challenge of an IT industry is to engineer a software product with minimum post deployment defects. This effort is an analysis based on data obtained for five selected projects from leading software companies of varying software production competence. The main aim of this paper is to provide information on various methods and practices supporting defect detection and prevention leading to thriving software generation. The defect prevention technique unearths 99% of defects. Inspection is found to be an essential technique in generating ideal software generation in factories through enhanced methodologies of abetted and unaided inspection schedules. On an average 13 % to 15% of inspection and 25% - 30% of testing out of whole project effort time is required for 99% - 99.75% of defect elimination. A comparison of the end results for the five selected projects between the companies is also brought about throwing light on the possibility of a particular company to position itself with an appropriate complementary ratio of inspection testing.Keywords: Defect Detection and Prevention, Inspections, Software Engineering, Software Process, Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1537252 Impulse Response Shortening for Discrete Multitone Transceivers using Convex Optimization Approach
Authors: Ejaz Khan, Conor Heneghan
Abstract:
In this paper we propose a new criterion for solving the problem of channel shortening in multi-carrier systems. In a discrete multitone receiver, a time-domain equalizer (TEQ) reduces intersymbol interference (ISI) by shortening the effective duration of the channel impulse response. Minimum mean square error (MMSE) method for TEQ does not give satisfactory results. In [1] a new criterion for partially equalizing severe ISI channels to reduce the cyclic prefix overhead of the discrete multitone transceiver (DMT), assuming a fixed transmission bandwidth, is introduced. Due to specific constrained (unit morm constraint on the target impulse response (TIR)) in their method, the freedom to choose optimum vector (TIR) is reduced. Better results can be obtained by avoiding the unit norm constraint on the target impulse response (TIR). In this paper we change the cost function proposed in [1] to the cost function of determining the maximum of a determinant subject to linear matrix inequality (LMI) and quadratic constraint and solve the resulting optimization problem. Usefulness of the proposed method is shown with the help of simulations.Keywords: Equalizer, target impulse response, convex optimization, matrix inequality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1712251 Passive Flow Control in Twin Air-Intakes
Authors: Akshoy R. Paul, Pritanshu Ranjan, Ravi R. Upadhyay, Anuj Jain
Abstract:
Aircraft propulsion systems often use Y-shaped subsonic diffusing ducts as twin air-intakes to supply the ambient air into the engine compressor for thrust generation. Due to space constraint, the diffusers need to be curved, which causes severe flow non-uniformity at the engine face. The present study attempt to control flow in a mild-curved Y-duct diffuser using trapezoidalshaped vortex generators (VG) attached on either both the sidewalls or top and bottom walls of the diffuser at the inflexion plane. A commercial computational fluid dynamics (CFD) code is modified and is used to simulate the effects of SVG in flow of a Y-duct diffuser. A few experiments are conducted for CFD code validation, while the rest are done computationally. The best combination of Yduct diffuser is found with VG-2 arranged in co-rotating sequence and attached to both the sidewalls, which ensures highest static pressure recovery, lowest total pressure loss, minimum flow distortion and less flow separation in Y-duct diffuser. The decrease in VG height while attached to top and bottom walls further improves axial flow uniformity at the diffuser outlet by a great margin as compared to the bare duct.Keywords: Twin air-intake, Vortex generator (VG), Turbulence model, Pressure recovery, Distortion coefficient
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2130250 Economic Evaluation of Bowland Shale Gas Wells Development in the UK
Authors: Elijah Acquah-Andoh
Abstract:
The UK has had its fair share of the shale gas revolutionary waves blowing across the global oil and gas industry at present. Although, its exploitation is widely agreed to have been delayed, shale gas was looked upon favorably by the UK Parliament when they recognized it as genuine energy source and granted licenses to industry to search and extract the resource. This, although a significant progress by industry, there yet remains another test the UK fracking resource must pass in order to render shale gas extraction feasible – it must be economically extractible and sustainably so. Developing unconventional resources is much more expensive and risky, and for shale gas wells, producing in commercial volumes is conditional upon drilling horizontal wells and hydraulic fracturing, techniques which increase CAPEX. Meanwhile, investment in shale gas development projects is sensitive to gas price and technical and geological risks. Using a Two-Factor Model, the economics of the Bowland shale wells were analyzed and the operational conditions under which fracking is profitable in the UK was characterized. We find that there is a great degree of flexibility about Opex spending; hence Opex does not pose much threat to the fracking industry in the UK. However, we discover Bowland shale gas wells fail to add value at gas price of $8/ Mmbtu. A minimum gas price of $12/Mmbtu at Opex of no more than $2/ Mcf and no more than $14.95M Capex are required to create value within the present petroleum tax regime, in the UK fracking industry.Keywords: Capex, economical, investment, profitability, shale gas development, sustainable.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2718249 Atmospheric Oxidation of Carbonyls: Insight to Mechanism, Kinetic and Thermodynamic Parameters
Authors: Olumayede Emmanuel Gbenga, Adeniyi Azeez Adebayo
Abstract:
Carbonyls are the first-generation products from tropospheric degradation reactions of volatile organic compounds (VOCs). This computational study examined the mechanism of removal of carbonyls from the atmosphere via hydroxyl radical. The kinetics of the reactions were computed from the activation energy (using enthalpy (ΔH**) and Gibbs free energy (ΔG**). The minimum energy path (MEP) analysis reveals that in all the molecules, the products have more stable energy than the reactants, which implies that the forward reaction is more thermodynamically favorable. The hydrogen abstraction of the aromatic aldehyde, especially without methyl substituents, is more kinetically favorable compared with the other aldehydes in the order of aromatic (without methyl or meta methyl) > alkene (short chain) > diene > long-chain aldehydes. The activation energy is much lower for the forward reaction than the backward, indicating that the forward reactions are more kinetically stable than their backward reaction. In terms of thermodynamic stability, the aromatic compounds are found to be less favorable in comparison to the aliphatic. The study concludes that the chemistry of the carbonyl bond of the aldehyde changed significantly from the reactants to the products.
Keywords: Atmospheric carbonyls, oxidation, mechanism, kinetic, thermodynamic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 52248 Non-Singular Gravitational Collapse of a Homogeneous Scalar Field in Deformed Phase Space
Authors: Amir Hadi Ziaie
Abstract:
In the present work, we revisit the collapse process of a spherically symmetric homogeneous scalar field (in FRW background) minimally coupled to gravity, when the phase-space deformations are taken into account. Such a deformation is mathematically introduced as a particular type of noncommutativity between the canonical momenta of the scale factor and of the scalar field. In the absence of such deformation, the collapse culminates in a spacetime singularity. However, when the phase-space is deformed, we find that the singularity is removed by a non-singular bounce, beyond which the collapsing cloud re-expands to infinity. More precisely, for negative values of the deformation parameter, we identify the appearance of a negative pressure, which decelerates the collapse to finally avoid the singularity formation. While in the un-deformed case, the horizon curve monotonically decreases to finally cover the singularity, in the deformed case the horizon has a minimum value that this value depends on deformation parameter and initial configuration of the collapse. Such a setting predicts a threshold mass for black hole formation in stellar collapse and manifests the role of non-commutative geometry in physics and especially in stellar collapse and supernova explosion.Keywords: Gravitational collapse, non-commutative geometry, spacetime singularity, black hole physics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1434247 A Speeded up Robust Scale-Invariant Feature Transform Currency Recognition Algorithm
Authors: Daliyah S. Aljutaili, Redna A. Almutlaq, Suha A. Alharbi, Dina M. Ibrahim
Abstract:
All currencies around the world look very different from each other. For instance, the size, color, and pattern of the paper are different. With the development of modern banking services, automatic methods for paper currency recognition become important in many applications like vending machines. One of the currency recognition architecture’s phases is Feature detection and description. There are many algorithms that are used for this phase, but they still have some disadvantages. This paper proposes a feature detection algorithm, which merges the advantages given in the current SIFT and SURF algorithms, which we call, Speeded up Robust Scale-Invariant Feature Transform (SR-SIFT) algorithm. Our proposed SR-SIFT algorithm overcomes the problems of both the SIFT and SURF algorithms. The proposed algorithm aims to speed up the SIFT feature detection algorithm and keep it robust. Simulation results demonstrate that the proposed SR-SIFT algorithm decreases the average response time, especially in small and minimum number of best key points, increases the distribution of the number of best key points on the surface of the currency. Furthermore, the proposed algorithm increases the accuracy of the true best point distribution inside the currency edge than the other two algorithms.
Keywords: Currency recognition, feature detection and description, SIFT algorithm, SURF algorithm, speeded up and robust features.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 864246 Enhanced Multi-Intensity Analysis in Multi-Scenery Classification-Based Macro and Micro Elements
Authors: R. Bremananth
Abstract:
Several computationally challenging issues are encountered while classifying complex natural scenes. In this paper, we address the problems that are encountered in rotation invariance with multi-intensity analysis for multi-scene overlapping. In the present literature, various algorithms proposed techniques for multi-intensity analysis, but there are several restrictions in these algorithms while deploying them in multi-scene overlapping classifications. In order to resolve the problem of multi-scenery overlapping classifications, we present a framework that is based on macro and micro basis functions. This algorithm conquers the minimum classification false alarm while pigeonholing multi-scene overlapping. Furthermore, a quadrangle multi-intensity decay is invoked. Several parameters are utilized to analyze invariance for multi-scenery classifications such as rotation, classification, correlation, contrast, homogeneity, and energy. Benchmark datasets were collected for complex natural scenes and experimented for the framework. The results depict that the framework achieves a significant improvement on gray-level matrix of co-occurrence features for overlapping in diverse degree of orientations while pigeonholing multi-scene overlapping.Keywords: Automatic classification, contrast, homogeneity, invariant analysis, multi-scene analysis, overlapping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1120245 Bipolar Square Wave Pulses for Liquid Food Sterilization using Cascaded H-Bridge Multilevel Inverter
Authors: Hanifah Jambari, Naziha A. Azli, M. Afendi M. Piah
Abstract:
This paper presents the generation of bipolar square wave pulses with characteristics that are suitable for liquid food sterilization using a Cascaded H-bridge Multilevel Inverter (CHMI). Bipolar square waves pulses have been reported as stable for a longer time during the sterilization process with minimum heat emission and increased efficiency. The CHMI allows the system to produce bipolar square wave pulses and yielding high output voltage without using a transformer while fulfilling the pulse requirements for effective liquid food sterilization. This in turn can reduce power consumption and cost of the overall liquid food sterilization system. The simulation results have shown that pulses with peak output voltage of 2.4 kV, pulse width of between 1 2s and 1 ms at frequencies of 50 Hz and 100 Hz can be generated by a 7-level CHMI. Results from the experimental set-up based on a 5-level CHMI has indicated the potential of the proposed circuit in producing bipolar square wave output pulses with peak values that depends on the DC source level supplied to the CHMI modules, pulse width of between 12.5 2s and 1 ms at frequencies of 50 Hz and 100 Hz.Keywords: pulsed electric field, multilevel inverter, bipolarsquare wave, food sterilization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2543244 Power Flow Tracing Based Reactive Power Ancillary Service (AS) in Restructured Power Market
Authors: M. Susithra, R. Gnanadass
Abstract:
Ancillary services are support services which are essential for humanizing and enhancing the reliability and security of the electric power system. Reactive power ancillary service is one of the important ancillary services in a restructured electricity market which determines the cost of supplying ancillary services and finding of how this cost would change with respect to operating decisions. This paper presents a new formation that can be used to minimize the Independent System Operator (ISO)’s total payment for reactive power ancillary service. The modified power flow tracing algorithm estimates the availability of reserve reactive power for ancillary service. In order to find optimum reactive power dispatch, Biogeography based optimization method (BPO) is proposed. Market Reactive Clearing Price (MRCP) is then estimated and it encourages generator companies (GENCOs) to participate in an ancillary service. Finally, optimal weighting factor and real time utilization factor of reactive power give the minimum ISO’s total payment. The effectiveness of proposed design is verified using IEEE 30 bus system.
Keywords: Biogeography based optimization method, Power flow tracing method, Reactive generation capability curve and Reactive power ancillary service.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3235243 Activity Recognition by Smartphone Accelerometer Data Using Ensemble Learning Methods
Authors: Eu Tteum Ha, Kwang Ryel Ryu
Abstract:
As smartphones are equipped with various sensors, there have been many studies focused on using these sensors to create valuable applications. Human activity recognition is one such application motivated by various welfare applications, such as the support for the elderly, measurement of calorie consumption, lifestyle and exercise patterns analyses, and so on. One of the challenges one faces when using smartphone sensors for activity recognition is that the number of sensors should be minimized to save battery power. In this paper, we show that a fairly accurate classifier can be built that can distinguish ten different activities by using only a single sensor data, i.e., the smartphone accelerometer data. The approach that we adopt to deal with this twelve-class problem uses various methods. The features used for classifying these activities include not only the magnitude of acceleration vector at each time point, but also the maximum, the minimum, and the standard deviation of vector magnitude within a time window. The experiments compared the performance of four kinds of basic multi-class classifiers and the performance of four kinds of ensemble learning methods based on three kinds of basic multi-class classifiers. The results show that while the method with the highest accuracy is ECOC based on Random forest.
Keywords: Ensemble learning, activity recognition, smartphone accelerometer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2173