Search results for: Point Isovist
592 Investigation of a Hybrid Process: Multipoint Incremental Forming
Authors: Safa Boudhaouia, Mohamed Amen Gahbiche, Eliane Giraud, Wacef Ben Salem, Philippe Dal Santo
Abstract:
Multi-point forming (MPF) and asymmetric incremental forming (ISF) are two flexible processes for sheet metal manufacturing. To take advantages of these two techniques, a hybrid process has been developed: The Multipoint Incremental Forming (MPIF). This process accumulates at once the advantages of each of these last mentioned forming techniques, which makes it a very interesting and particularly an efficient process for single, small, and medium series production. In this paper, an experimental and a numerical investigation of this technique are presented. To highlight the flexibility of this process and its capacity to manufacture standard and complex shapes, several pieces were produced by using MPIF. The forming experiments are performed on a 3-axis CNC machine. Moreover, a numerical model of the MPIF process has been implemented in ABAQUS and the analysis showed a good agreement with experimental results in terms of deformed shape. Furthermore, the use of an elastomeric interpolator allows avoiding classical local defaults like dimples, which are generally caused by the asymmetric contact and also improves the distribution of residual strain. Future works will apply this approach to other alloys used in aeronautic or automotive applications.Keywords: Incremental forming, numerical simulation, MPIF, multipoint forming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1312591 An Information Theoretic Approach to Rescoring Peptides Produced by De Novo Peptide Sequencing
Authors: John R. Rose, James P. Cleveland, Alvin Fox
Abstract:
Tandem mass spectrometry (MS/MS) is the engine driving high-throughput protein identification. Protein mixtures possibly representing thousands of proteins from multiple species are treated with proteolytic enzymes, cutting the proteins into smaller peptides that are then analyzed generating MS/MS spectra. The task of determining the identity of the peptide from its spectrum is currently the weak point in the process. Current approaches to de novo sequencing are able to compute candidate peptides efficiently. The problem lies in the limitations of current scoring functions. In this paper we introduce the concept of proteome signature. By examining proteins and compiling proteome signatures (amino acid usage) it is possible to characterize likely combinations of amino acids and better distinguish between candidate peptides. Our results strongly support the hypothesis that a scoring function that considers amino acid usage patterns is better able to distinguish between candidate peptides. This in turn leads to higher accuracy in peptide prediction.Keywords: Tandem mass spectrometry, proteomics, scoring, peptide, de novo, mutual information
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1730590 Basic Study of Mammographic Image Magnification System with Eye-Detector and Simple EEG Scanner
Authors: A. Umemuro, M. Sato, M. Narita, S. Hori, S. Sakurai, T. Nakayama, A. Nakazawa, T. Ogura
Abstract:
Mammography requires the detection of very small calcifications, and physicians search for microcalcifications by magnifying the images as they read them. The mouse is necessary to zoom in on the images, but this can be tiring and distracting when many images are read in a single day. Therefore, an image magnification system combining an eye-detector and a simple electroencephalograph (EEG) scanner was devised, and its operability was evaluated. Two experiments were conducted in this study: the measurement of eye-detection error using an eye-detector and the measurement of the time required for image magnification using a simple EEG scanner. Eye-detector validation showed that the mean distance of eye-detection error ranged from 0.64 cm to 2.17 cm, with an overall mean of 1.24 ± 0.81 cm for the observers. The results showed that the eye detection error was small enough for the magnified area of the mammographic image. The average time required for point magnification in the verification of the simple EEG scanner ranged from 5.85 to 16.73 seconds, and individual differences were observed. The reason for this may be that the size of the simple EEG scanner used was not adjustable, so it did not fit well for some subjects. The use of a simple EEG scanner with size adjustment would solve this problem. Therefore, the image magnification system using the eye-detector and the simple EEG scanner is useful.
Keywords: EEG scanner, eye-detector, mammography, observers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 366589 Numerical Analysis of Laminar Flow around Square Cylinders with EHD Phenomenon
Authors: M. Salmanpour, O. Nourani Zonouz
Abstract:
In this research, a numerical simulation of an Electrohydrodynamic (EHD) actuator’s effects on the flow around a square cylinder by using a finite volume method has been investigated. This is one of the newest ways for controlling the fluid flows. Two plate electrodes are flush-mounted on the surface of the cylinder and one wire electrode is placed on the line with zero angle of attack relative to the stagnation point and excited with DC power supply. The discharge produces an electric force and changes the local momentum behaviors in the fluid layers. For this purpose, after selecting proper domain and boundary conditions, the electric field relating to the problem has been analyzed and then the results in the form of electrical body force have been entered in the governing equations of fluid field (Navier-Stokes equations). The effect of ionic wind resulted from the Electrohydrodynamic actuator, on the velocity, pressure and the wake behind cylinder has been considered. According to the results, it is observed that the fluid flow accelerates in the nearest wall of the frontal half of the cylinder and the pressure difference between frontal and hinder cylinder is increased.
Keywords: CFD, corona discharge, electro hydrodynamics, flow around square cylinders.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 878588 Modeling and Simulation of Two-Phase Interleaved Boost Converter Using Open-Source Software Scilab/Xcos
Authors: Yin Yin Phyo, Tun Lin Naing
Abstract:
This paper investigated the simulation of two-phase interleaved boost converter (IBC) with free and open-source software Scilab/Xcos. By using interleaved method, it can reduce current stress on components, components size, input current ripple and output voltage ripple. The required mathematical model is obtained from the equivalent circuit of its different four modes of operation for simulation. The equivalent circuits are considered in continuous conduction mode (CCM). The average values of the system variables are derived from the state-space equation to find the equilibrium point. Scilab is now becoming more and more popular among students, engineers and scientists because it is open-source software and free of charge. It gives a great convenience because it has powerful computation and simulation function. The waveforms of output voltage, input current and inductors current are obtained by using Scilab/Xcos.
Keywords: Two-phase boost converter, continuous conduction mode, free and open-source, interleaved method, dynamic simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 947587 Combined Model Predictive Controller Technique for Enhancing NAO Gait Stabilization
Authors: Brahim Brahmi, Mohammed Hamza Laraki, Mohammad Habibur Rahman, Islam M. Rasedul, M. Assad Uz-Zaman
Abstract:
The humanoid robot, specifically the NAO robot must be able to provide a highly dynamic performance on the soccer field. Maintaining the balance of the humanoid robot during the required motion is considered as one of a challenging problems especially when the robot is subject to external disturbances, as contact with other robots. In this paper, a dynamic controller is proposed in order to ensure a robust walking (stabilization) and to improve the dynamic balance of the robot during its contact with the environment (external disturbances). The generation of the trajectory of the center of mass (CoM) is done by a model predictive controller (MPC) conjoined with zero moment point (ZMP) technique. Taking into account the properties of the rotational dynamics of the whole-body system, a modified previous control mixed with feedback control is employed to manage the angular momentum and the CoM’s acceleration, respectively. This latter is dedicated to provide a robust gait of the robot in the presence of the external disturbances. Simulation results are presented to show the feasibility of the proposed strategy.Keywords: Preview control, walking, stabilization, humanoid robot.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 591586 Application of Biometrics to Obtain High Entropy Cryptographic Keys
Authors: Sanjay Kanade, Danielle Camara, Dijana Petrovska-Delacretaz, Bernadette Dorizzi
Abstract:
In this paper, a two factor scheme is proposed to generate cryptographic keys directly from biometric data, which unlike passwords, are strongly bound to the user. Hash value of the reference iris code is used as a cryptographic key and its length depends only on the hash function, being independent of any other parameter. The entropy of such keys is 94 bits, which is much higher than any other comparable system. The most important and distinct feature of this scheme is that it regenerates the reference iris code by providing a genuine iris sample and the correct user password. Since iris codes obtained from two images of the same eye are not exactly the same, error correcting codes (Hadamard code and Reed-Solomon code) are used to deal with the variability. The scheme proposed here can be used to provide keys for a cryptographic system and/or for user authentication. The performance of this system is evaluated on two publicly available databases for iris biometrics namely CBS and ICE databases. The operating point of the system (values of False Acceptance Rate (FAR) and False Rejection Rate (FRR)) can be set by properly selecting the error correction capacity (ts) of the Reed- Solomon codes, e.g., on the ICE database, at ts = 15, FAR is 0.096% and FRR is 0.76%. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2091585 A Numerical Description of a Fibre Reinforced Concrete Using a Genetic Algorithm
Authors: Henrik L. Funke, Lars Ulke-Winter, Sandra Gelbrich, Lothar Kroll
Abstract:
This work reports about an approach for an automatic adaptation of concrete formulations based on genetic algorithms (GA) to optimize a wide range of different fit-functions. In order to achieve the goal, a method was developed which provides a numerical description of a fibre reinforced concrete (FRC) mixture regarding the production technology and the property spectrum of the concrete. In a first step, the FRC mixture with seven fixed components was characterized by varying amounts of the components. For that purpose, ten concrete mixtures were prepared and tested. The testing procedure comprised flow spread, compressive and bending tensile strength. The analysis and approximation of the determined data was carried out by GAs. The aim was to obtain a closed mathematical expression which best describes the given seven-point cloud of FRC by applying a Gene Expression Programming with Free Coefficients (GEP-FC) strategy. The seven-parametric FRC-mixtures model which is generated according to this method correlated well with the measured data. The developed procedure can be used for concrete mixtures finding closed mathematical expressions, which are based on the measured data.
Keywords: Concrete design, fibre reinforced concrete, genetic algorithms, GEP-FC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 991584 Software Reliability Prediction Model Analysis
Authors: L. Mirtskhulava, M. Khunjgurua, N. Lomineishvili, K. Bakuria
Abstract:
Software reliability prediction gives a great opportunity to measure the software failure rate at any point throughout system test. A software reliability prediction model provides with the technique for improving reliability. Software reliability is very important factor for estimating overall system reliability, which depends on the individual component reliabilities. It differs from hardware reliability in that it reflects the design perfection. Main reason of software reliability problems is high complexity of software. Various approaches can be used to improve the reliability of software. We focus on software reliability model in this article, assuming that there is a time redundancy, the value of which (the number of repeated transmission of basic blocks) can be an optimization parameter. We consider given mathematical model in the assumption that in the system may occur not only irreversible failures, but also a failure that can be taken as self-repairing failures that significantly affect the reliability and accuracy of information transfer. Main task of the given paper is to find a time distribution function (DF) of instructions sequence transmission, which consists of random number of basic blocks. We consider the system software unreliable; the time between adjacent failures has exponential distribution.
Keywords: Exponential distribution, conditional mean time to failure, distribution function, mathematical model, software reliability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1683583 Connected Vertex Cover in 2-Connected Planar Graph with Maximum Degree 4 is NP-complete
Authors: Priyadarsini P. L. K, Hemalatha T.
Abstract:
This paper proves that the problem of finding connected vertex cover in a 2-connected planar graph ( CVC-2 ) with maximum degree 4 is NP-complete. The motivation for proving this result is to give a shorter and simpler proof of NP-Completeness of TRA-MLC (the Top Right Access point Minimum-Length Corridor) problem [1], by finding the reduction from CVC-2. TRA-MLC has many applications in laying optical fibre cables for data communication and electrical wiring in floor plans.The problem of finding connected vertex cover in any planar graph ( CVC ) with maximum degree 4 is NP-complete [2]. We first show that CVC-2 belongs to NP and then we find a polynomial reduction from CVC to CVC-2. Let a graph G0 and an integer K form an instance of CVC, where G0 is a planar graph and K is an upper bound on the size of the connected vertex cover in G0. We construct a 2-connected planar graph, say G, by identifying the blocks and cut vertices of G0, and then finding the planar representation of all the blocks of G0, leading to a plane graph G1. We replace the cut vertices with cycles in such a way that the resultant graph G is a 2-connected planar graph with maximum degree 4. We consider L = K -2t+3 t i=1 di where t is the number of cut vertices in G1 and di is the number of blocks for which ith cut vertex is common. We prove that G will have a connected vertex cover with size less than or equal to L if and only if G0 has a connected vertex cover of size less than or equal to K.Keywords: NP-complete, 2-Connected planar graph, block, cut vertex
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2005582 Measuring the Relationship between Customers- Satisfaction and Cognitions: A Case of Janfusun Fancyworld in Taiwan
Authors: Wan-Yu Liu, Yen-Hsiang Liu, Shing-Yi Huang, Hao-Zhi Wen
Abstract:
The private theme parks are gradually surpassing public-owned scenic areas after many years of development and have become a mainstream choice for domestic tourists. Previous studies show that visitors from different backgrounds differ in consumer behavior and satisfaction factors. An understanding of visitor satisfaction is therefore of extreme importance to operators of privately-owned theme parks. Importance-Performance Analysis (IPA) is used to measure consumer's potential satisfaction with services and has become a widely used management tool for strength and weakness analysis for brands, products, services and point of sales. As IPA has so far not been used to evaluate the visitor satisfaction with privately-owned theme parks, in this study the IPA method is used to analyze visitor satisfaction with Janfusun Fancyworld (one of the most popular private theme parks in Taiwan) and to rank visitor focus and satisfaction on/in theme park facilities and services. Results of the analysis provide private theme park operators with an understanding of user or consumer demands as well as an assessment of the quality of services currently offered.Keywords: Satisfaction, Importance-Performance Analysis, Theme Parks, Janfusun Fancyworld.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2450581 Comparison of GSA, SA and PSO Based Intelligent Controllers for Path Planning of Mobile Robot in Unknown Environment
Authors: P. K. Panigrahi, Saradindu Ghosh, Dayal R. Parhi
Abstract:
Now-a-days autonomous mobile robots have found applications in diverse fields. An autonomous robot system must be able to behave in an intelligent manner to deal with complex and changing environment. This work proposes the performance of path planning and navigation of autonomous mobile robot using Gravitational Search Algorithm (GSA), Simulated Annealing (SA) and Particle Swarm optimization (PSO) based intelligent controllers in an unstructured environment. The approach not only finds a valid collision free path but also optimal one. The main aim of the work is to minimize the length of the path and duration of travel from a starting point to a target while moving in an unknown environment with obstacles without collision. Finally, a comparison is made between the three controllers, it is found that the path length and time duration made by the robot using GSA is better than SA and PSO based controllers for the same work.
Keywords: Autonomous Mobile Robot, Gravitational Search Algorithm, Particle Swarm Optimization, Simulated Annealing Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3120580 The Effect of Confinement Shapes on Over-Reinforced HSC Beams
Authors: Ross Jeffry, Muhammad N. S. Hadi
Abstract:
High strength concrete (HSC) provides high strength but lower ductility than normal strength concrete. This low ductility limits the benefit of using HSC in building safe structures. On the other hand, when designing reinforced concrete beams, designers have to limit the amount of tensile reinforcement to prevent the brittle failure of concrete. Therefore the full potential of the use of steel reinforcement can not be achieved. This paper presents the idea of confining concrete in the compression zone so that the HSC will be in a state of triaxial compression, which leads to improvements in strength and ductility. Five beams made of HSC were cast and tested. The cross section of the beams was 200×300 mm, with a length of 4 m and a clear span of 3.6 m subjected to four-point loading, with emphasis placed on the midspan deflection. The first beam served as a reference beam. The remaining beams had different tensile reinforcement and the confinement shapes were changed to gauge their effectiveness in improving the strength and ductility of the beams. The compressive strength of the concrete was 85 MPa and the tensile strength of the steel was 500 MPa and for the stirrups and helixes was 250 MPa. Results of testing the five beams proved that placing helixes with different diameters as a variable parameter in the compression zone of reinforced concrete beams improve their strength and ductility.Keywords: Confinement, ductility, high strength concrete, reinforced concrete beam.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2241579 A Differential Calculus Based Image Steganography with Crossover
Authors: Srilekha Mukherjee, Subha Ash, Goutam Sanyal
Abstract:
Information security plays a major role in uplifting the standard of secured communications via global media. In this paper, we have suggested a technique of encryption followed by insertion before transmission. Here, we have implemented two different concepts to carry out the above-specified tasks. We have used a two-point crossover technique of the genetic algorithm to facilitate the encryption process. For each of the uniquely identified rows of pixels, different mathematical methodologies are applied for several conditions checking, in order to figure out all the parent pixels on which we perform the crossover operation. This is done by selecting two crossover points within the pixels thereby producing the newly encrypted child pixels, and hence the encrypted cover image. In the next lap, the first and second order derivative operators are evaluated to increase the security and robustness. The last lap further ensures reapplication of the crossover procedure to form the final stego-image. The complexity of this system as a whole is huge, thereby dissuading the third party interferences. Also, the embedding capacity is very high. Therefore, a larger amount of secret image information can be hidden. The imperceptible vision of the obtained stego-image clearly proves the proficiency of this approach.Keywords: Steganography, Crossover, Differential Calculus, Peak Signal to Noise Ratio, Cross-correlation Coefficient.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1399578 First-Principle Investigation of the Electronic Band Structure and Dielectric Response Function of ZnIn2Se4 and ZnIn2Te4
Authors: Nnamdi N. Omehe, Chibuzo Emeruwa
Abstract:
ZnIn2Se4 and ZnIn2Te4 are vacancy defect materials whose properties have been investigated using Density Functional Theory (DFT) framework. The pseudopotential method in conjunction with the LDA+U technique and the Projector Augmented Wave (PAW) was used to calculate the electronic band structure, total density of state, and the partial density of state; while the norm-conserving pseudopotential was used to calculate the dielectric response function with scissors shift. Both ZnIn2Se4 and ZnIn2Te4 were predicted to be semiconductors with energy band gap of 1.66 eV and 1.33 eV respectively, and they both have direct energy band gap at the gamma point of high symmetry. The topmost valence subband for ZnIn2Se4 and ZnIn2Te4 has an energy width of 5.7 eV and 6.0 eV respectively. The calculations of partial density of state (PDOS) show that for ZnIn2Se4, the top of the valence band is dominated by Se-4p orbital, while the bottom of the conduction band is composed of In-5p, In-5s, and Zn-4s states. PDOS for ZnIn2Te4, shows that the top of the valence band is mostly of Te-5p states, while its conduction band bottom is composed mainly of Zn-4s, Te-5p, Te-5s, and In-5s states. Dielectric response function calculation yielded (0) of 11.9 and 36 for ZnIn2Se4 and ZnIn2Te4 respectively.
Keywords: Optoelectronic, Dielectric Response Function, LDA+U, band structure calculation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 112577 Daemon- Based Distributed Deadlock Detection and Resolution
Authors: Z. RahimAlipour, A. T. Haghighat
Abstract:
detecting the deadlock is one of the important problems in distributed systems and different solutions have been proposed for it. Among the many deadlock detection algorithms, Edge-chasing has been the most widely used. In Edge-chasing algorithm, a special message called probe is made and sent along dependency edges. When the initiator of a probe receives the probe back the existence of a deadlock is revealed. But these algorithms are not problem-free. One of the problems associated with them is that they cannot detect some deadlocks and they even identify false deadlocks. A key point not mentioned in the literature is that when the process is waiting to obtain the required resources and its execution has been blocked, how it can actually respond to probe messages in the system. Also the question of 'which process should be victimized in order to achieve a better performance when multiple cycles exist within one single process in the system' has received little attention. In this paper, one of the basic concepts of the operating system - daemon - will be used to solve the problems mentioned. The proposed Algorithm becomes engaged in sending probe messages to the mandatory daemons and collects enough information to effectively identify and resolve multi-cycle deadlocks in distributed systems.Keywords: Distributed system, distributed deadlock detectionand resolution, daemon, false deadlock.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1937576 Effect of Co3O4 Nanoparticles Addition on (Bi,Pb)-2223 Superconductor
Authors: A. N. Jannah, R. Abd-Shukor, H. Abdullah
Abstract:
The effect of nano Co3O4 addition on the superconducting properties of (Bi, Pb)-2223 system was studied. The samples were prepared by the acetate coprecipitation method. The Co3O4 with different sizes (10-30 nm and 30-50 nm) from x=0.00 to 0.05 was added to Bi1.6Pb0.4Sr2Ca2Cu3Oy(Co3O4)x. Phase analysis by XRD method, microstructural examination by SEM and dc electrical resistivity by four point probe method were done to characterize the samples. The X-ray diffraction patterns of all the samples indicated the majority Bi-2223 phase along with minor Bi-2212 and Bi-2201 phases. The volume fraction was estimated from the intensities of Bi- 2223, Bi-2212 and Bi-2201 phase. The sample with x=0.01 wt% of the added Co3O4 (10-30 nm size) showed the highest volume fraction of Bi-2223 phase (72%) and the highest superconducting transition temperature, Tc (~102 K). The non-added sample showed the highest Tc(~103 K) compared to added samples with nano Co3O4 (30-50 nm size) added samples. Both the onset critical temperature Tc(onset) and zero electrical resistivity temperature Tc(R=0) were in the range of 103-115 ±1K and 91-103 ±1K respectively for samples with added Co3O4 (10-30 nm and 30-50 nm).Keywords: Bi(Pb)-Sr-Ca-Cu-O superconductor, coprecipitation, nano Co3O4, transition temperature TC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2042575 Multiple Job Shop-Scheduling using Hybrid Heuristic Algorithm
Authors: R.A.Mahdavinejad
Abstract:
In this paper, multi-processors job shop scheduling problems are solved by a heuristic algorithm based on the hybrid of priority dispatching rules according to an ant colony optimization algorithm. The objective function is to minimize the makespan, i.e. total completion time, in which a simultanous presence of various kinds of ferons is allowed. By using the suitable hybrid of priority dispatching rules, the process of finding the best solution will be improved. Ant colony optimization algorithm, not only promote the ability of this proposed algorithm, but also decreases the total working time because of decreasing in setup times and modifying the working production line. Thus, the similar work has the same production lines. Other advantage of this algorithm is that the similar machines (not the same) can be considered. So, these machines are able to process a job with different processing and setup times. According to this capability and from this algorithm evaluation point of view, a number of test problems are solved and the associated results are analyzed. The results show a significant decrease in throughput time. It also shows that, this algorithm is able to recognize the bottleneck machine and to schedule jobs in an efficient way.
Keywords: Job shops scheduling, Priority dispatching rules, Makespan, Hybrid heuristic algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1670574 Coordination between SC and SVC for Voltage Stability Improvement
Authors: Ali Reza Rajabi, Shahab Rashnoei, Mojtaba Hakimzadeh, Amir Habibi
Abstract:
At any point of time, a power system operating condition should be stable, meeting various operational criteria and it should also be secure in the event of any credible contingency. Present day power systems are being operated closer to their stability limits due to economic and environmental constraints. Maintaining a stable and secure operation of a power system is therefore a very important and challenging issue. Voltage instability has been given much attention by power system researchers and planners in recent years, and is being regarded as one of the major sources of power system insecurity. Voltage instability phenomena are the ones in which the receiving end voltage decreases well below its normal value and does not come back even after setting restoring mechanisms such as VAR compensators, or continues to oscillate for lack of damping against the disturbances. Reactive power limit of power system is one of the major causes of voltage instability. This paper investigates the effects of coordinated series capacitors (SC) with static VAR compensators (SVC) on steady-state voltage stability of a power system. Also, the influence of the presence of series capacitor on static VAR compensator controller parameters and ratings required to stabilize load voltages at certain values are highlighted.
Keywords: Static VAR Compensator (SVC), Series Capacitor (SC), voltage stability, reactive power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1967573 Multi Switched Split Vector Quantization of Narrowband Speech Signals
Authors: M. Satya Sai Ram, P. Siddaiah, M. Madhavi Latha
Abstract:
Vector quantization is a powerful tool for speech coding applications. This paper deals with LPC Coding of speech signals which uses a new technique called Multi Switched Split Vector Quantization (MSSVQ), which is a hybrid of Multi, switched, split vector quantization techniques. The spectral distortion performance, computational complexity, and memory requirements of MSSVQ are compared to split vector quantization (SVQ), multi stage vector quantization(MSVQ) and switched split vector quantization (SSVQ) techniques. It has been proved from results that MSSVQ has better spectral distortion performance, lower computational complexity and lower memory requirements when compared to all the above mentioned product code vector quantization techniques. Computational complexity is measured in floating point operations (flops), and memory requirements is measured in (floats).Keywords: Linear predictive Coding, Multi stage vectorquantization, Switched Split vector quantization, Split vectorquantization, Line Spectral Frequencies (LSF).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1675572 Theoretical Analysis of Self-Starting Busemann Intake Family
Authors: N. Moradian, E. Timofeev, R. Tahir
Abstract:
In this work, startability of the Busemann intake family with weak/strong conical shock, as most efficient intakes, via overboard mass spillage method is theoretically analyzed. Masterix and Candifix codes are used to numerically simulate few models of this type of intake and verify the theoretical results. Portions of the intake corresponding to various flow capture angles are considered to have mass spillage in the starting process of this intake. This approach allows for overboard mass spillage via a V-shaped slot with the tip of V coinciding with the focal point of the Busemann flow. The theoretical results, achieved using two different theories, of self-started Busemann takes with weak/strong conical shock show that significant improve in intake startability using overboard spillage technique. The starting phenomena of Busemann intakes with weak conical shock and seven different capture angles are numerically simulated at freestream Mach number of 3 to find the minimum area ratios of self-started intakes. The numerical results confirm the theoretical ones achieved by authors.Keywords: Busemann intake, conical shock, overboard spillage, startability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1139571 Investigating the Behavior of Underground Structures in the Event of an Earthquake
Authors: Davoud Beheshtizadeh, Farzin Malekpour
Abstract:
The progress of technology and producing new machinery have made a big change in excavation operations and construction of underground structures. The limitations of space and some other economic, politic and military considerations gained the attention of most developed and developing countries towards the construction of these structures for mine, military, and development objectives. Underground highways, tunnels, subways, oil reservoir resources, fuels, nuclear wastes burying reservoir and underground stores are increasingly developing and being used in these countries. The existence and habitability of the cities depend on these underground installations or in other words these vital arteries. Stopping the flow of water, gas leakage and explosion, collapsing of sewage paths, etc., resulting from the earthquake are among the factors that can severely harm the environment and increase the casualty. Lack of sewage network and complete stoppage of the flow of water in Bam (Iran) is a good example of this kind. In this paper, we investigate the effect of wave orientation on structures and deformation of them and the effect of faulting on underground structures, and then, we study resistance of reinforced concrete against earthquake, simulate two different samples, analyze the result and point out the importance of paying attention to underground installations.
Keywords: Earthquake, underground structures, underground installations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 968570 Novel Rao-Blackwellized Particle Filter for Mobile Robot SLAM Using Monocular Vision
Authors: Maohai Li, Bingrong Hong, Zesu Cai, Ronghua Luo
Abstract:
This paper presents the novel Rao-Blackwellised particle filter (RBPF) for mobile robot simultaneous localization and mapping (SLAM) using monocular vision. The particle filter is combined with unscented Kalman filter (UKF) to extending the path posterior by sampling new poses that integrate the current observation which drastically reduces the uncertainty about the robot pose. The landmark position estimation and update is also implemented through UKF. Furthermore, the number of resampling steps is determined adaptively, which seriously reduces the particle depletion problem, and introducing the evolution strategies (ES) for avoiding particle impoverishment. The 3D natural point landmarks are structured with matching Scale Invariant Feature Transform (SIFT) feature pairs. The matching for multi-dimension SIFT features is implemented with a KD-Tree in the time cost of O(log2 N). Experiment results on real robot in our indoor environment show the advantages of our methods over previous approaches.Keywords: Mobile robot, simultaneous localization and mapping, Rao-Blackwellised particle filter, evolution strategies, scale invariant feature transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2147569 A Structural Support Vector Machine Approach for Biometric Recognition
Authors: Vishal Awasthi, Atul Kumar Agnihotri
Abstract:
Face is a non-intrusive strong biometrics for identification of original and dummy facial by different artificial means. Face recognition is extremely important in the contexts of computer vision, psychology, surveillance, pattern recognition, neural network, content based video processing. The availability of a widespread face database is crucial to test the performance of these face recognition algorithms. The openly available face databases include face images with a wide range of poses, illumination, gestures and face occlusions but there is no dummy face database accessible in public domain. This paper presents a face detection algorithm based on the image segmentation in terms of distance from a fixed point and template matching methods. This proposed work is having the most appropriate number of nodal points resulting in most appropriate outcomes in terms of face recognition and detection. The time taken to identify and extract distinctive facial features is improved in the range of 90 to 110 sec. with the increment of efficiency by 3%.Keywords: Face recognition, Principal Component Analysis, PCA, Linear Discriminant Analysis, LDA, Improved Support Vector Machine, iSVM, elastic bunch mapping technique.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 498568 Hydrogen Rich Fuel Gas Production from 2- Propanol Using Pt/Al2O3 and Ni/Al2O3 Catalysts in Supercritical Water
Authors: Yağmur Karakuş, Fatih Aynacı, Ekin Kıpçak, Mesut Akgün
Abstract:
Hydrogen is an important chemical in many industries and it is expected to become one of the major fuels for energy generation in the future. Unfortunately, hydrogen does not exist in its elemental form in nature and therefore has to be produced from hydrocarbons, hydrogen-containing compounds or water. Above its critical point (374.8oC and 22.1MPa), water has lower density and viscosity, and a higher heat capacity than those of ambient water. Mass transfer in supercritical water (SCW) is enhanced due to its increased diffusivity and transport ability. The reduced dielectric constant makes supercritical water a better solvent for organic compounds and gases. Hence, due to the aforementioned desirable properties, there is a growing interest toward studies regarding the gasification of organic matter containing biomass or model biomass solutions in supercritical water. In this study, hydrogen and biofuel production by the catalytic gasification of 2-Propanol in supercritical conditions of water was investigated. Pt/Al2O3and Ni/Al2O3were the catalysts used in the gasification reactions. All of the experiments were performed under a constant pressure of 25MPa. The effects of five reaction temperatures (400, 450, 500, 550 and 600°C) and five reaction times (10, 15, 20, 25 and 30 s) on the gasification yield and flammable component content were investigated.Keywords: 2-Propanol, Gasification, Ni/Al2O3, Pt/Al2O3, Supercritical water.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2055567 Review and Classification of the Indicators and Trends Used in Bridge Performance Modeling
Authors: S. Rezaei, Z. Mirzaei, M. Khalighi, J. Bahrami
Abstract:
Bridges, as an essential part of road infrastructures, are affected by various deterioration mechanisms over time due to the changes in their performance. As changes in performance can have many negative impacts on society, it is essential to be able to evaluate and measure the performance of bridges throughout their life. This evaluation includes the development or the choice of the appropriate performance indicators, which, in turn, are measured based on the selection of appropriate models for the existing deterioration mechanism. The purpose of this article is a statistical study of indicators and deterioration mechanisms of bridges in order to discover further research capacities in bridges performance assessment. For this purpose, some of the most common indicators of bridge performance, including reliability, risk, vulnerability, robustness, and resilience, were selected. The researches performed on each index based on the desired deterioration mechanisms and hazards were comprehensively reviewed. In addition, the formulation of the indicators and their relationship with each other were studied. The research conducted on the mentioned indicators were classified from the point of view of deterministic or probabilistic method, the level of study (element level, object level, etc.), and the type of hazard and the deterioration mechanism of interest. For each of the indicators, a number of challenges and recommendations were presented according to the review of previous studies.
Keywords: Bridge, deterioration mechanism, lifecycle, performance indicator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 460566 Bipolar PWM and LCL Filter Configuration to Reduce Leakage Currents in Transformerless PV System Connected to Utility Grid
Authors: Shanmuka Naga Raju
Abstract:
This paper presents PV system without considering transformer connected to electric grid. This is considered more economic compared to present PV system. The problem that occurs when transformer is not considered appears with a leakage current near capacitor connected to ground. Bipolar Pulse Width Modulation (BPWM) technique along with filter L-C-L configuration in the circuit is modeled to shrink the leakage current in the circuit. The DC/AC inverter is modeled using H-bridge Insulated Gate Bipolar Transistor (IGBT) module which is controlled using proposed Bipolar PWM control technique. To extract maximum power, Maximum Power Point Technique (MPPT) controller is used in this model. Voltage and current regulators are used to determine the reference voltage for the inverter from active and reactive current where reactive current is set to zero. The PLL is modeled to synchronize the measurements. The model is designed with MATLAB Simulation blocks and compared with the methods available in literature survey to show its effectiveness.Keywords: Photovoltaic, PV, pulse width modulation, PWM, perturb and observe, phase locked loop.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1024565 A Comprehensive Evaluation of IGBTs Performance under Zero Current Switching
Authors: Ly. Benbahouche
Abstract:
Currently, several soft switching topologies have been studied to achieve high power switching efficiency, reduced cost, improved reliability and reduced parasites. It is well known that improvement in power electronics systems always depend on advanced in power devices. The IGBT has been successfully used in a variety of switching applications such as motor drives and appliance control because of its superior characteristics.
The aim of this paper is focuses on simulation and explication of the internal dynamics of IGBTs behaviour under the most popular soft switching schemas that is Zero Current Switching (ZCS) environments.
The main purpose of this paper is to point out some mechanisms relating to current tail during the turn-off and examination of the response at turn-off with variation of temperature, inductance L, snubber capacitors Cs, and bus voltage in order to achieve an improved understanding of internal carrier dynamics. It is shown that the snubber capacitor, the inductance and even the temperature controls the magnitude and extent of the tail current, hence the turn-off time (switching speed of the device).
Moreover, it has also been demonstrated that the ZCS switching can be utilized efficiently to improve and reduce the power losses as well as the turn-off time. Furthermore, the turn-off loss in ZCS was found to depend on the time of switching of the device.
Keywords: PT-IGBT, ZCS, turn-off losses, dV/dt.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2587564 A Novel Method Based on Monte Carlo for Simulation of Variable Resolution X-ray CT Scanner: Measurement of System Presampling MTF
Authors: H. Arabi, A.R. Kamali Asl
Abstract:
The purpose of this work is measurement of the system presampling MTF of a variable resolution x-ray (VRX) CT scanner. In this paper, we used the parameters of an actual VRX CT scanner for simulation and study of effect of different focal spot sizes on system presampling MTF by Monte Carlo method (GATE simulation software). Focal spot size of 0.6 mm limited the spatial resolution of the system to 5.5 cy/mm at incident angles of below 17º for cell#1. By focal spot size of 0.3 mm the spatial resolution increased up to 11 cy/mm and the limiting effect of focal spot size appeared at incident angles of below 9º. The focal spot size of 0.3 mm could improve the spatial resolution to some extent but because of magnification non-uniformity, there is a 10 cy/mm difference between spatial resolution of cell#1 and cell#256. The focal spot size of 0.1 mm acted as an ideal point source for this system. The spatial resolution increased to more than 35 cy/mm and at all incident angles the spatial resolution was a function of incident angle. By the way focal spot size of 0.1 mm minimized the effect of magnification nonuniformity.Keywords: Focal spot, Spatial resolution, Monte Carlosimulation, Variable resolution x-ray (VRX) CT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1529563 Axiomatic Systems as an Alternative to Teach Physics
Authors: Liliana M. Marinelli, Cristina T. Varanese
Abstract:
In the last few years, students from higher education have difficulties in grasping mathematical concepts which support physical matters, especially those in the first years of this education. Classical Physics teaching turns to be complex when students are not able to make use of mathematical tools which lead to the conceptual structure of Physics. When derivation and integration rules are not used or developed in parallel with other disciplines, the physical meaning that we attempt to convey turns to be complicated. Due to this fact, it could be of great use to see the Classical Mechanics from an axiomatic approach, where the correspondence rules give physical meaning, if we expect students to understand concepts clearly and accurately. Using the Minkowski point of view adapted to a two-dimensional space and time where vectors, matrices, and straight lines (worked from an affine space) give mathematical and physical rigorosity even when it is more abstract. An interesting option would be to develop the disciplinary contents from an axiomatic version which embraces the Classical Mechanics as a particular case of Relativistic Mechanics. The observation about the increase in the difficulties stated by students in the first years of education allows this idea to grow as a possible option to improve performance and understanding of the concepts of this subject.
Keywords: Axiom, classical physics, physical concepts, relativity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1284