Search results for: lifting point adjustment
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1942

Search results for: lifting point adjustment

592 Biomechanics Analysis of Bicross Start

Authors: M. Kalichová, S. Hřebíčková, R. Labounková, P. Hedbávný, G. Bago

Abstract:

The article deals with a biomechanics analysis of the classic bicross start with a backward movement of the bike. This is a case study analyzing this type of start in two bicross riders representing the Czech Republic. Based on the 3D kinematic analysis and with a special emphasis on the ankle movement we have divided the start into five phases – phase n. 1 – reaction time, phase n. 2 – preparation movements time, phase n. 3 – first pedal stroke time, phase n. 4 – dead point pedal passage time, phase n. 5 – second pedal stroke time. Further we have demonstrated the significance of kinematic characteristics in various stages of the bicross start including their values and the extent of change. These primarily include the vector of the instantaneous velocity of the head, wrists, elbows, shoulders, hip and knee joints. The significant angle characteristics have been noted in elbow, shoulder, hip and knee joints. The results of this work indicate the types of movement prevailing in the respective phases and as such are expected to serve as a basis for further analyses of this movement structure performed, however, on a large research sample.

Keywords: Bicross, start, kinematic analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2580
591 Application of Company Financial Crisis Early Warning Model- Use of “Financial Reference Database“

Authors: Chiung-ying Lee, Chia-hua Chang

Abstract:

In July 1, 2007, Taiwan Stock Exchange (TWSE) on market observation post system (MOPS) adds a new "Financial reference database" for investors to do investment reference. This database as a warning to public offering companies listed on the public financial information and it original within eight targets. In this paper, this database provided by the indicators for the application of company financial crisis early warning model verify that the database provided by the indicator forecast for the financial crisis, whether or not companies have a high accuracy rate as opposed to domestic and foreign scholars have positive results. There is use of Logistic Regression Model application of the financial early warning model, in which no joined back-conditions is the first model, joined it in is the second model, has been taken occurred in the financial crisis of companies to research samples and then business took place before the financial crisis point with T-1 and T-2 sample data to do positive analysis. The results show that this database provided the debt ratio and net per share for the best forecast variables.

Keywords: Financial reference database, Financial early warning model, Logistic Regression.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1431
590 Packet Losses Interpretation in Mobile Internet

Authors: Hossam el-ddin Mostafa, Pavel Čičak

Abstract:

The mobile users with Laptops need to have an efficient access to i.e. their home personal data or to the Internet from any place in the world, regardless of their location or point of attachment, especially while roaming outside the home subnet. An efficient interpretation of packet losses problem that is encountered from this roaming is to the centric of all aspects in this work, to be over-highlighted. The main previous works, such as BER-systems, Amigos, and ns-2 implementation that are considered to be in conjunction with that problem under study are reviewed and discussed. Their drawbacks and limitations, of stopping only at monitoring, and not to provide an actual solution for eliminating or even restricting these losses, are mentioned. Besides that, the framework around which we built a Triple-R sequence as a costeffective solution to eliminate the packet losses and bridge the gap between subnets, an area that until now has been largely neglected, is presented. The results show that, in addition to the high bit error rate of wireless mobile networks, mainly the low efficiency of mobile-IP registration procedure is a direct cause of these packet losses. Furthermore, the output of packet losses interpretation resulted an illustrated triangle of the registration process. This triangle should be further researched and analyzed in our future work.

Keywords: Amigos, BER-systems, ns-2 implementation, packetlosses, registration process, roaming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1474
589 Enhancing Human Mobility Exoskeleton Comfort Using Admittance Controller

Authors: Alexandre Rabaseda, Emelie Seguin, Marc Doumit

Abstract:

Human mobility exoskeletons have been in development for several years and are becoming increasingly efficient. Unfortunately, user comfort was not always a priority design criterion throughout their development. To further improve this technology, exoskeletons should operate and deliver assistance without causing discomfort to the user. For this, improvements are necessary from an ergonomic point of view. The device’s control method is important when endeavoring to enhance user comfort. Exoskeleton or rehabilitation device controllers use methods of control called interaction controls (admittance and impedance controls). This paper proposes an extended version of an admittance controller to enhance user comfort. The control method used consists of adding an inner loop that is controlled by a proportional-integral-derivative (PID) controller. This allows the interaction force to be kept as close as possible to the desired force trajectory. The force-tracking admittance controller modifies the actuation force of the system in order to follow both the desired motion trajectory and the desired relative force between the user and the exoskeleton.

Keywords: Mobility assistive device, exoskeleton, force-tracking admittance controller, user comfort.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 455
588 Model based Soft-Sensor for Industrial Crystallization: On-line Mass of Crystals and Solubility Measurement

Authors: Cédric Damour, Michel Benne, Brigitte Grondin-Perez, Jean-Pierre Chabriat

Abstract:

Monitoring and control of cane sugar crystallization processes depend on the stability of the supersaturation (σ ) state. The most widely used information to represent σ is the electrical conductivity κ of the solutions. Nevertheless, previous studies point out the shortcomings of this approach: κ may be regarded as inappropriate to guarantee an accurate estimation of σ in impure solutions. To improve the process control efficiency, additional information is necessary. The mass of crystals in the solution ( c m ) and the solubility (mass ratio of sugar to water / s w m m ) are relevant to complete information. Indeed, c m inherently contains information about the mass balance and / s w m m contains information about the supersaturation state of the solution. The main problem is that c m and / s w m m are not available on-line. In this paper, a model based soft-sensor is presented for a final crystallization stage (C sugar). Simulation results obtained on industrial data show the reliability of this approach, c m and the crystal content ( cc ) being estimated with a sufficient accuracy for achieving on-line monitoring in industry

Keywords: Soft-sensor, on-line monitoring, cane sugarcrystallization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2139
587 Investigation of a Hybrid Process: Multipoint Incremental Forming

Authors: Safa Boudhaouia, Mohamed Amen Gahbiche, Eliane Giraud, Wacef Ben Salem, Philippe Dal Santo

Abstract:

Multi-point forming (MPF) and asymmetric incremental forming (ISF) are two flexible processes for sheet metal manufacturing. To take advantages of these two techniques, a hybrid process has been developed: The Multipoint Incremental Forming (MPIF). This process accumulates at once the advantages of each of these last mentioned forming techniques, which makes it a very interesting and particularly an efficient process for single, small, and medium series production. In this paper, an experimental and a numerical investigation of this technique are presented. To highlight the flexibility of this process and its capacity to manufacture standard and complex shapes, several pieces were produced by using MPIF. The forming experiments are performed on a 3-axis CNC machine. Moreover, a numerical model of the MPIF process has been implemented in ABAQUS and the analysis showed a good agreement with experimental results in terms of deformed shape. Furthermore, the use of an elastomeric interpolator allows avoiding classical local defaults like dimples, which are generally caused by the asymmetric contact and also improves the distribution of residual strain. Future works will apply this approach to other alloys used in aeronautic or automotive applications.

Keywords: Incremental forming, numerical simulation, MPIF, multipoint forming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1314
586 An Information Theoretic Approach to Rescoring Peptides Produced by De Novo Peptide Sequencing

Authors: John R. Rose, James P. Cleveland, Alvin Fox

Abstract:

Tandem mass spectrometry (MS/MS) is the engine driving high-throughput protein identification. Protein mixtures possibly representing thousands of proteins from multiple species are treated with proteolytic enzymes, cutting the proteins into smaller peptides that are then analyzed generating MS/MS spectra. The task of determining the identity of the peptide from its spectrum is currently the weak point in the process. Current approaches to de novo sequencing are able to compute candidate peptides efficiently. The problem lies in the limitations of current scoring functions. In this paper we introduce the concept of proteome signature. By examining proteins and compiling proteome signatures (amino acid usage) it is possible to characterize likely combinations of amino acids and better distinguish between candidate peptides. Our results strongly support the hypothesis that a scoring function that considers amino acid usage patterns is better able to distinguish between candidate peptides. This in turn leads to higher accuracy in peptide prediction.

Keywords: Tandem mass spectrometry, proteomics, scoring, peptide, de novo, mutual information

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1733
585 Numerical Analysis of Laminar Flow around Square Cylinders with EHD Phenomenon

Authors: M. Salmanpour, O. Nourani Zonouz

Abstract:

In this research, a numerical simulation of an Electrohydrodynamic (EHD) actuator’s effects on the flow around a square cylinder by using a finite volume method has been investigated. This is one of the newest ways for controlling the fluid flows. Two plate electrodes are flush-mounted on the surface of the cylinder and one wire electrode is placed on the line with zero angle of attack relative to the stagnation point and excited with DC power supply. The discharge produces an electric force and changes the local momentum behaviors in the fluid layers. For this purpose, after selecting proper domain and boundary conditions, the electric field relating to the problem has been analyzed and then the results in the form of electrical body force have been entered in the governing equations of fluid field (Navier-Stokes equations). The effect of ionic wind resulted from the Electrohydrodynamic actuator, on the velocity, pressure and the wake behind cylinder has been considered. According to the results, it is observed that the fluid flow accelerates in the nearest wall of the frontal half of the cylinder and the pressure difference between frontal and hinder cylinder is increased.

Keywords: CFD, corona discharge, electro hydrodynamics, flow around square cylinders.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 879
584 Modeling and Simulation of Two-Phase Interleaved Boost Converter Using Open-Source Software Scilab/Xcos

Authors: Yin Yin Phyo, Tun Lin Naing

Abstract:

This paper investigated the simulation of two-phase interleaved boost converter (IBC) with free and open-source software Scilab/Xcos. By using interleaved method, it can reduce current stress on components, components size, input current ripple and output voltage ripple. The required mathematical model is obtained from the equivalent circuit of its different four modes of operation for simulation. The equivalent circuits are considered in continuous conduction mode (CCM). The average values of the system variables are derived from the state-space equation to find the equilibrium point. Scilab is now becoming more and more popular among students, engineers and scientists because it is open-source software and free of charge. It gives a great convenience because it has powerful computation and simulation function. The waveforms of output voltage, input current and inductors current are obtained by using Scilab/Xcos.

Keywords: Two-phase boost converter, continuous conduction mode, free and open-source, interleaved method, dynamic simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 949
583 Combined Model Predictive Controller Technique for Enhancing NAO Gait Stabilization

Authors: Brahim Brahmi, Mohammed Hamza Laraki, Mohammad Habibur Rahman, Islam M. Rasedul, M. Assad Uz-Zaman

Abstract:

The humanoid robot, specifically the NAO robot must be able to provide a highly dynamic performance on the soccer field. Maintaining the balance of the humanoid robot during the required motion is considered as one of a challenging problems especially when the robot is subject to external disturbances, as contact with other robots. In this paper, a dynamic controller is proposed in order to ensure a robust walking (stabilization) and to improve the dynamic balance of the robot during its contact with the environment (external disturbances). The generation of the trajectory of the center of mass (CoM) is done by a model predictive controller (MPC) conjoined with zero moment point (ZMP) technique. Taking into account the properties of the rotational dynamics of the whole-body system, a modified previous control mixed with feedback control is employed to manage the angular momentum and the CoM’s acceleration, respectively. This latter is dedicated to provide a robust gait of the robot in the presence of the external disturbances. Simulation results are presented to show the feasibility of the proposed strategy.

Keywords: Preview control, walking, stabilization, humanoid robot.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 594
582 Application of Biometrics to Obtain High Entropy Cryptographic Keys

Authors: Sanjay Kanade, Danielle Camara, Dijana Petrovska-Delacretaz, Bernadette Dorizzi

Abstract:

In this paper, a two factor scheme is proposed to generate cryptographic keys directly from biometric data, which unlike passwords, are strongly bound to the user. Hash value of the reference iris code is used as a cryptographic key and its length depends only on the hash function, being independent of any other parameter. The entropy of such keys is 94 bits, which is much higher than any other comparable system. The most important and distinct feature of this scheme is that it regenerates the reference iris code by providing a genuine iris sample and the correct user password. Since iris codes obtained from two images of the same eye are not exactly the same, error correcting codes (Hadamard code and Reed-Solomon code) are used to deal with the variability. The scheme proposed here can be used to provide keys for a cryptographic system and/or for user authentication. The performance of this system is evaluated on two publicly available databases for iris biometrics namely CBS and ICE databases. The operating point of the system (values of False Acceptance Rate (FAR) and False Rejection Rate (FRR)) can be set by properly selecting the error correction capacity (ts) of the Reed- Solomon codes, e.g., on the ICE database, at ts = 15, FAR is 0.096% and FRR is 0.76%.

Keywords:

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2092
581 A Numerical Description of a Fibre Reinforced Concrete Using a Genetic Algorithm

Authors: Henrik L. Funke, Lars Ulke-Winter, Sandra Gelbrich, Lothar Kroll

Abstract:

This work reports about an approach for an automatic adaptation of concrete formulations based on genetic algorithms (GA) to optimize a wide range of different fit-functions. In order to achieve the goal, a method was developed which provides a numerical description of a fibre reinforced concrete (FRC) mixture regarding the production technology and the property spectrum of the concrete. In a first step, the FRC mixture with seven fixed components was characterized by varying amounts of the components. For that purpose, ten concrete mixtures were prepared and tested. The testing procedure comprised flow spread, compressive and bending tensile strength. The analysis and approximation of the determined data was carried out by GAs. The aim was to obtain a closed mathematical expression which best describes the given seven-point cloud of FRC by applying a Gene Expression Programming with Free Coefficients (GEP-FC) strategy. The seven-parametric FRC-mixtures model which is generated according to this method correlated well with the measured data. The developed procedure can be used for concrete mixtures finding closed mathematical expressions, which are based on the measured data.

Keywords: Concrete design, fibre reinforced concrete, genetic algorithms, GEP-FC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 992
580 Software Reliability Prediction Model Analysis

Authors: L. Mirtskhulava, M. Khunjgurua, N. Lomineishvili, K. Bakuria

Abstract:

Software reliability prediction gives a great opportunity to measure the software failure rate at any point throughout system test. A software reliability prediction model provides with the technique for improving reliability. Software reliability is very important factor for estimating overall system reliability, which depends on the individual component reliabilities. It differs from hardware reliability in that it reflects the design perfection. Main reason of software reliability problems is high complexity of software. Various approaches can be used to improve the reliability of software. We focus on software reliability model in this article, assuming that there is a time redundancy, the value of which (the number of repeated transmission of basic blocks) can be an optimization parameter. We consider given mathematical model in the assumption that in the system may occur not only irreversible failures, but also a failure that can be taken as self-repairing failures that significantly affect the reliability and accuracy of information transfer. Main task of the given paper is to find a time distribution function (DF) of instructions sequence transmission, which consists of random number of basic blocks. We consider the system software unreliable; the time between adjacent failures has exponential distribution.

Keywords: Exponential distribution, conditional mean time to failure, distribution function, mathematical model, software reliability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1684
579 Connected Vertex Cover in 2-Connected Planar Graph with Maximum Degree 4 is NP-complete

Authors: Priyadarsini P. L. K, Hemalatha T.

Abstract:

This paper proves that the problem of finding connected vertex cover in a 2-connected planar graph ( CVC-2 ) with maximum degree 4 is NP-complete. The motivation for proving this result is to give a shorter and simpler proof of NP-Completeness of TRA-MLC (the Top Right Access point Minimum-Length Corridor) problem [1], by finding the reduction from CVC-2. TRA-MLC has many applications in laying optical fibre cables for data communication and electrical wiring in floor plans.The problem of finding connected vertex cover in any planar graph ( CVC ) with maximum degree 4 is NP-complete [2]. We first show that CVC-2 belongs to NP and then we find a polynomial reduction from CVC to CVC-2. Let a graph G0 and an integer K form an instance of CVC, where G0 is a planar graph and K is an upper bound on the size of the connected vertex cover in G0. We construct a 2-connected planar graph, say G, by identifying the blocks and cut vertices of G0, and then finding the planar representation of all the blocks of G0, leading to a plane graph G1. We replace the cut vertices with cycles in such a way that the resultant graph G is a 2-connected planar graph with maximum degree 4. We consider L = K -2t+3 t i=1 di where t is the number of cut vertices in G1 and di is the number of blocks for which ith cut vertex is common. We prove that G will have a connected vertex cover with size less than or equal to L if and only if G0 has a connected vertex cover of size less than or equal to K.

Keywords: NP-complete, 2-Connected planar graph, block, cut vertex

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2006
578 Measuring the Relationship between Customers- Satisfaction and Cognitions: A Case of Janfusun Fancyworld in Taiwan

Authors: Wan-Yu Liu, Yen-Hsiang Liu, Shing-Yi Huang, Hao-Zhi Wen

Abstract:

The private theme parks are gradually surpassing public-owned scenic areas after many years of development and have become a mainstream choice for domestic tourists. Previous studies show that visitors from different backgrounds differ in consumer behavior and satisfaction factors. An understanding of visitor satisfaction is therefore of extreme importance to operators of privately-owned theme parks. Importance-Performance Analysis (IPA) is used to measure consumer's potential satisfaction with services and has become a widely used management tool for strength and weakness analysis for brands, products, services and point of sales. As IPA has so far not been used to evaluate the visitor satisfaction with privately-owned theme parks, in this study the IPA method is used to analyze visitor satisfaction with Janfusun Fancyworld (one of the most popular private theme parks in Taiwan) and to rank visitor focus and satisfaction on/in theme park facilities and services. Results of the analysis provide private theme park operators with an understanding of user or consumer demands as well as an assessment of the quality of services currently offered.

Keywords: Satisfaction, Importance-Performance Analysis, Theme Parks, Janfusun Fancyworld.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2451
577 Comparison of GSA, SA and PSO Based Intelligent Controllers for Path Planning of Mobile Robot in Unknown Environment

Authors: P. K. Panigrahi, Saradindu Ghosh, Dayal R. Parhi

Abstract:

Now-a-days autonomous mobile robots have found applications in diverse fields. An autonomous robot system must be able to behave in an intelligent manner to deal with complex and changing environment. This work proposes the performance of path planning and navigation of autonomous mobile robot using Gravitational Search Algorithm (GSA), Simulated Annealing (SA) and Particle Swarm optimization (PSO) based intelligent controllers in an unstructured environment. The approach not only finds a valid collision free path but also optimal one. The main aim of the work is to minimize the length of the path and duration of travel from a starting point to a target while moving in an unknown environment with obstacles without collision. Finally, a comparison is made between the three controllers, it is found that the path length and time duration made by the robot using GSA is better than SA and PSO based controllers for the same work.

Keywords: Autonomous Mobile Robot, Gravitational Search Algorithm, Particle Swarm Optimization, Simulated Annealing Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3121
576 The Effect of Confinement Shapes on Over-Reinforced HSC Beams

Authors: Ross Jeffry, Muhammad N. S. Hadi

Abstract:

High strength concrete (HSC) provides high strength but lower ductility than normal strength concrete. This low ductility limits the benefit of using HSC in building safe structures. On the other hand, when designing reinforced concrete beams, designers have to limit the amount of tensile reinforcement to prevent the brittle failure of concrete. Therefore the full potential of the use of steel reinforcement can not be achieved. This paper presents the idea of confining concrete in the compression zone so that the HSC will be in a state of triaxial compression, which leads to improvements in strength and ductility. Five beams made of HSC were cast and tested. The cross section of the beams was 200×300 mm, with a length of 4 m and a clear span of 3.6 m subjected to four-point loading, with emphasis placed on the midspan deflection. The first beam served as a reference beam. The remaining beams had different tensile reinforcement and the confinement shapes were changed to gauge their effectiveness in improving the strength and ductility of the beams. The compressive strength of the concrete was 85 MPa and the tensile strength of the steel was 500 MPa and for the stirrups and helixes was 250 MPa. Results of testing the five beams proved that placing helixes with different diameters as a variable parameter in the compression zone of reinforced concrete beams improve their strength and ductility.

Keywords: Confinement, ductility, high strength concrete, reinforced concrete beam.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2243
575 A Differential Calculus Based Image Steganography with Crossover

Authors: Srilekha Mukherjee, Subha Ash, Goutam Sanyal

Abstract:

Information security plays a major role in uplifting the standard of secured communications via global media. In this paper, we have suggested a technique of encryption followed by insertion before transmission. Here, we have implemented two different concepts to carry out the above-specified tasks. We have used a two-point crossover technique of the genetic algorithm to facilitate the encryption process. For each of the uniquely identified rows of pixels, different mathematical methodologies are applied for several conditions checking, in order to figure out all the parent pixels on which we perform the crossover operation. This is done by selecting two crossover points within the pixels thereby producing the newly encrypted child pixels, and hence the encrypted cover image. In the next lap, the first and second order derivative operators are evaluated to increase the security and robustness. The last lap further ensures reapplication of the crossover procedure to form the final stego-image. The complexity of this system as a whole is huge, thereby dissuading the third party interferences. Also, the embedding capacity is very high. Therefore, a larger amount of secret image information can be hidden. The imperceptible vision of the obtained stego-image clearly proves the proficiency of this approach.

Keywords: Steganography, Crossover, Differential Calculus, Peak Signal to Noise Ratio, Cross-correlation Coefficient.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1399
574 First-Principle Investigation of the Electronic Band Structure and Dielectric Response Function of ZnIn2Se4 and ZnIn2Te4

Authors: Nnamdi N. Omehe, Chibuzo Emeruwa

Abstract:

ZnIn2Se4 and ZnIn2Te4 are vacancy defect materials whose properties have been investigated using Density Functional Theory (DFT) framework. The pseudopotential method in conjunction with the LDA+U technique and the Projector Augmented Wave (PAW) was used to calculate the electronic band structure, total density of state, and the partial density of state; while the norm-conserving pseudopotential was used to calculate the dielectric response function with scissors shift. Both ZnIn2Se4 and ZnIn2Te4 were predicted to be semiconductors with energy band gap of 1.66 eV and 1.33 eV respectively, and they both have direct energy band gap at the gamma point of high symmetry. The topmost valence subband for ZnIn2Se4 and ZnIn2Te4 has an energy width of 5.7 eV and 6.0 eV respectively. The calculations of partial density of state (PDOS) show that for ZnIn2Se4, the top of the valence band is dominated by Se-4p orbital, while the bottom of the conduction band is composed of In-5p, In-5s, and Zn-4s states. PDOS for ZnIn2Te4, shows that the top of the valence band is mostly of Te-5p states, while its conduction band bottom is composed mainly of Zn-4s, Te-5p, Te-5s, and In-5s states. Dielectric response function calculation yielded (0) of 11.9 and 36 for ZnIn2Se4 and ZnIn2Te4 respectively.

Keywords: Optoelectronic, Dielectric Response Function, LDA+U, band structure calculation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 116
573 Daemon- Based Distributed Deadlock Detection and Resolution

Authors: Z. RahimAlipour, A. T. Haghighat

Abstract:

detecting the deadlock is one of the important problems in distributed systems and different solutions have been proposed for it. Among the many deadlock detection algorithms, Edge-chasing has been the most widely used. In Edge-chasing algorithm, a special message called probe is made and sent along dependency edges. When the initiator of a probe receives the probe back the existence of a deadlock is revealed. But these algorithms are not problem-free. One of the problems associated with them is that they cannot detect some deadlocks and they even identify false deadlocks. A key point not mentioned in the literature is that when the process is waiting to obtain the required resources and its execution has been blocked, how it can actually respond to probe messages in the system. Also the question of 'which process should be victimized in order to achieve a better performance when multiple cycles exist within one single process in the system' has received little attention. In this paper, one of the basic concepts of the operating system - daemon - will be used to solve the problems mentioned. The proposed Algorithm becomes engaged in sending probe messages to the mandatory daemons and collects enough information to effectively identify and resolve multi-cycle deadlocks in distributed systems.

Keywords: Distributed system, distributed deadlock detectionand resolution, daemon, false deadlock.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1938
572 Effect of Co3O4 Nanoparticles Addition on (Bi,Pb)-2223 Superconductor

Authors: A. N. Jannah, R. Abd-Shukor, H. Abdullah

Abstract:

The effect of nano Co3O4 addition on the superconducting properties of (Bi, Pb)-2223 system was studied. The samples were prepared by the acetate coprecipitation method. The Co3O4 with different sizes (10-30 nm and 30-50 nm) from x=0.00 to 0.05 was added to Bi1.6Pb0.4Sr2Ca2Cu3Oy(Co3O4)x. Phase analysis by XRD method, microstructural examination by SEM and dc electrical resistivity by four point probe method were done to characterize the samples. The X-ray diffraction patterns of all the samples indicated the majority Bi-2223 phase along with minor Bi-2212 and Bi-2201 phases. The volume fraction was estimated from the intensities of Bi- 2223, Bi-2212 and Bi-2201 phase. The sample with x=0.01 wt% of the added Co3O4 (10-30 nm size) showed the highest volume fraction of Bi-2223 phase (72%) and the highest superconducting transition temperature, Tc (~102 K). The non-added sample showed the highest Tc(~103 K) compared to added samples with nano Co3O4 (30-50 nm size) added samples. Both the onset critical temperature Tc(onset) and zero electrical resistivity temperature Tc(R=0) were in the range of 103-115 ±1K and 91-103 ±1K respectively for samples with added Co3O4 (10-30 nm and 30-50 nm).

Keywords: Bi(Pb)-Sr-Ca-Cu-O superconductor, coprecipitation, nano Co3O4, transition temperature TC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2043
571 Multiple Job Shop-Scheduling using Hybrid Heuristic Algorithm

Authors: R.A.Mahdavinejad

Abstract:

In this paper, multi-processors job shop scheduling problems are solved by a heuristic algorithm based on the hybrid of priority dispatching rules according to an ant colony optimization algorithm. The objective function is to minimize the makespan, i.e. total completion time, in which a simultanous presence of various kinds of ferons is allowed. By using the suitable hybrid of priority dispatching rules, the process of finding the best solution will be improved. Ant colony optimization algorithm, not only promote the ability of this proposed algorithm, but also decreases the total working time because of decreasing in setup times and modifying the working production line. Thus, the similar work has the same production lines. Other advantage of this algorithm is that the similar machines (not the same) can be considered. So, these machines are able to process a job with different processing and setup times. According to this capability and from this algorithm evaluation point of view, a number of test problems are solved and the associated results are analyzed. The results show a significant decrease in throughput time. It also shows that, this algorithm is able to recognize the bottleneck machine and to schedule jobs in an efficient way.

Keywords: Job shops scheduling, Priority dispatching rules, Makespan, Hybrid heuristic algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672
570 Coordination between SC and SVC for Voltage Stability Improvement

Authors: Ali Reza Rajabi, Shahab Rashnoei, Mojtaba Hakimzadeh, Amir Habibi

Abstract:

At any point of time, a power system operating condition should be stable, meeting various operational criteria and it should also be secure in the event of any credible contingency. Present day power systems are being operated closer to their stability limits due to economic and environmental constraints. Maintaining a stable and secure operation of a power system is therefore a very important and challenging issue. Voltage instability has been given much attention by power system researchers and planners in recent years, and is being regarded as one of the major sources of power system insecurity. Voltage instability phenomena are the ones in which the receiving end voltage decreases well below its normal value and does not come back even after setting restoring mechanisms such as VAR compensators, or continues to oscillate for lack of damping against the disturbances. Reactive power limit of power system is one of the major causes of voltage instability. This paper investigates the effects of coordinated series capacitors (SC) with static VAR compensators (SVC) on steady-state voltage stability of a power system. Also, the influence of the presence of series capacitor on static VAR compensator controller parameters and ratings required to stabilize load voltages at certain values are highlighted.

Keywords: Static VAR Compensator (SVC), Series Capacitor (SC), voltage stability, reactive power.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1970
569 Multi Switched Split Vector Quantization of Narrowband Speech Signals

Authors: M. Satya Sai Ram, P. Siddaiah, M. Madhavi Latha

Abstract:

Vector quantization is a powerful tool for speech coding applications. This paper deals with LPC Coding of speech signals which uses a new technique called Multi Switched Split Vector Quantization (MSSVQ), which is a hybrid of Multi, switched, split vector quantization techniques. The spectral distortion performance, computational complexity, and memory requirements of MSSVQ are compared to split vector quantization (SVQ), multi stage vector quantization(MSVQ) and switched split vector quantization (SSVQ) techniques. It has been proved from results that MSSVQ has better spectral distortion performance, lower computational complexity and lower memory requirements when compared to all the above mentioned product code vector quantization techniques. Computational complexity is measured in floating point operations (flops), and memory requirements is measured in (floats).

Keywords: Linear predictive Coding, Multi stage vectorquantization, Switched Split vector quantization, Split vectorquantization, Line Spectral Frequencies (LSF).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1677
568 Theoretical Analysis of Self-Starting Busemann Intake Family

Authors: N. Moradian, E. Timofeev, R. Tahir

Abstract:

In this work, startability of the Busemann intake family with weak/strong conical shock, as most efficient intakes, via overboard mass spillage method is theoretically analyzed. Masterix and Candifix codes are used to numerically simulate few models of this type of intake and verify the theoretical results. Portions of the intake corresponding to various flow capture angles are considered to have mass spillage in the starting process of this intake. This approach allows for overboard mass spillage via a V-shaped slot with the tip of V coinciding with the focal point of the Busemann flow. The theoretical results, achieved using two different theories, of self-started Busemann takes with weak/strong conical shock show that significant improve in intake startability using overboard spillage technique. The starting phenomena of Busemann intakes with weak conical shock and seven different capture angles are numerically simulated at freestream Mach number of 3 to find the minimum area ratios of self-started intakes. The numerical results confirm the theoretical ones achieved by authors.

Keywords: Busemann intake, conical shock, overboard spillage, startability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1141
567 Investigating the Behavior of Underground Structures in the Event of an Earthquake

Authors: Davoud Beheshtizadeh, Farzin Malekpour

Abstract:

The progress of technology and producing new machinery have made a big change in excavation operations and construction of underground structures. The limitations of space and some other economic, politic and military considerations gained the attention of most developed and developing countries towards the construction of these structures for mine, military, and development objectives. Underground highways, tunnels, subways, oil reservoir resources, fuels, nuclear wastes burying reservoir and underground stores are increasingly developing and being used in these countries. The existence and habitability of the cities depend on these underground installations or in other words these vital arteries. Stopping the flow of water, gas leakage and explosion, collapsing of sewage paths, etc., resulting from the earthquake are among the factors that can severely harm the environment and increase the casualty. Lack of sewage network and complete stoppage of the flow of water in Bam (Iran) is a good example of this kind. In this paper, we investigate the effect of wave orientation on structures and deformation of them and the effect of faulting on underground structures, and then, we study resistance of reinforced concrete against earthquake, simulate two different samples, analyze the result and point out the importance of paying attention to underground installations.

Keywords: Earthquake, underground structures, underground installations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 971
566 Novel Rao-Blackwellized Particle Filter for Mobile Robot SLAM Using Monocular Vision

Authors: Maohai Li, Bingrong Hong, Zesu Cai, Ronghua Luo

Abstract:

This paper presents the novel Rao-Blackwellised particle filter (RBPF) for mobile robot simultaneous localization and mapping (SLAM) using monocular vision. The particle filter is combined with unscented Kalman filter (UKF) to extending the path posterior by sampling new poses that integrate the current observation which drastically reduces the uncertainty about the robot pose. The landmark position estimation and update is also implemented through UKF. Furthermore, the number of resampling steps is determined adaptively, which seriously reduces the particle depletion problem, and introducing the evolution strategies (ES) for avoiding particle impoverishment. The 3D natural point landmarks are structured with matching Scale Invariant Feature Transform (SIFT) feature pairs. The matching for multi-dimension SIFT features is implemented with a KD-Tree in the time cost of O(log2 N). Experiment results on real robot in our indoor environment show the advantages of our methods over previous approaches.

Keywords: Mobile robot, simultaneous localization and mapping, Rao-Blackwellised particle filter, evolution strategies, scale invariant feature transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2148
565 A Structural Support Vector Machine Approach for Biometric Recognition

Authors: Vishal Awasthi, Atul Kumar Agnihotri

Abstract:

Face is a non-intrusive strong biometrics for identification of original and dummy facial by different artificial means. Face recognition is extremely important in the contexts of computer vision, psychology, surveillance, pattern recognition, neural network, content based video processing. The availability of a widespread face database is crucial to test the performance of these face recognition algorithms. The openly available face databases include face images with a wide range of poses, illumination, gestures and face occlusions but there is no dummy face database accessible in public domain. This paper presents a face detection algorithm based on the image segmentation in terms of distance from a fixed point and template matching methods. This proposed work is having the most appropriate number of nodal points resulting in most appropriate outcomes in terms of face recognition and detection. The time taken to identify and extract distinctive facial features is improved in the range of 90 to 110 sec. with the increment of efficiency by 3%.

Keywords: Face recognition, Principal Component Analysis, PCA, Linear Discriminant Analysis, LDA, Improved Support Vector Machine, iSVM, elastic bunch mapping technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 499
564 Hydrogen Rich Fuel Gas Production from 2- Propanol Using Pt/Al2O3 and Ni/Al2O3 Catalysts in Supercritical Water

Authors: Yağmur Karakuş, Fatih Aynacı, Ekin Kıpçak, Mesut Akgün

Abstract:

Hydrogen is an important chemical in many industries and it is expected to become one of the major fuels for energy generation in the future. Unfortunately, hydrogen does not exist in its elemental form in nature and therefore has to be produced from hydrocarbons, hydrogen-containing compounds or water. Above its critical point (374.8oC and 22.1MPa), water has lower density and viscosity, and a higher heat capacity than those of ambient water. Mass transfer in supercritical water (SCW) is enhanced due to its increased diffusivity and transport ability. The reduced dielectric constant makes supercritical water a better solvent for organic compounds and gases. Hence, due to the aforementioned desirable properties, there is a growing interest toward studies regarding the gasification of organic matter containing biomass or model biomass solutions in supercritical water. In this study, hydrogen and biofuel production by the catalytic gasification of 2-Propanol in supercritical conditions of water was investigated. Pt/Al2O3and Ni/Al2O3were the catalysts used in the gasification reactions. All of the experiments were performed under a constant pressure of 25MPa. The effects of five reaction temperatures (400, 450, 500, 550 and 600°C) and five reaction times (10, 15, 20, 25 and 30 s) on the gasification yield and flammable component content were investigated.

Keywords: 2-Propanol, Gasification, Ni/Al2O3, Pt/Al2O3, Supercritical water.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2056
563 Review and Classification of the Indicators and Trends Used in Bridge Performance Modeling

Authors: S. Rezaei, Z. Mirzaei, M. Khalighi, J. Bahrami

Abstract:

Bridges, as an essential part of road infrastructures, are affected by various deterioration mechanisms over time due to the changes in their performance. As changes in performance can have many negative impacts on society, it is essential to be able to evaluate and measure the performance of bridges throughout their life. This evaluation includes the development or the choice of the appropriate performance indicators, which, in turn, are measured based on the selection of appropriate models for the existing deterioration mechanism. The purpose of this article is a statistical study of indicators and deterioration mechanisms of bridges in order to discover further research capacities in bridges performance assessment. For this purpose, some of the most common indicators of bridge performance, including reliability, risk, vulnerability, robustness, and resilience, were selected. The researches performed on each index based on the desired deterioration mechanisms and hazards were comprehensively reviewed. In addition, the formulation of the indicators and their relationship with each other were studied. The research conducted on the mentioned indicators were classified from the point of view of deterministic or probabilistic method, the level of study (element level, object level, etc.), and the type of hazard and the deterioration mechanism of interest. For each of the indicators, a number of challenges and recommendations were presented according to the review of previous studies.

Keywords: Bridge, deterioration mechanism, lifecycle, performance indicator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 461