Search results for: Fault Coverage
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 531

Search results for: Fault Coverage

141 Robust Power System Stabilizer Design Using Particle Swarm Optimization Technique

Authors: Sidhartha Panda, N. P. Padhy

Abstract:

Power system stabilizers (PSS) are now routinely used in the industry to damp out power system oscillations. In this paper, particle swarm optimization (PSO) technique is applied to design a robust power system stabilizer (PSS). The design problem of the proposed controller is formulated as an optimization problem and PSO is employed to search for optimal controller parameters. By minimizing the time-domain based objective function, in which the deviation in the oscillatory rotor speed of the generator is involved; stability performance of the system is improved. The non-linear simulation results are presented under wide range of operating conditions; disturbances at different locations as well as for various fault clearing sequences to show the effectiveness and robustness of the proposed controller and their ability to provide efficient damping of low frequency oscillations. Further, all the simulations results are compared with a conventionally designed power system stabilizer to show the superiority of the proposed design approach.

Keywords: Particle swarm optimization, power system stabilizer, low frequency oscillations, power system stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2353
140 Generating Class-Based Test Cases for Interface Classes of Object-Oriented Black Box Frameworks

Authors: Jehad Al Dallal, Paul Sorenson

Abstract:

An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define the Framework Interface Classes (FICs) and their possible specifications, which helps in building reusable test cases for the implementations of these classes. This paper introduces a novel technique called all paths-state to generate state-based test cases to test the FICs at class level. The technique is experimentally evaluated. The empirical evaluation shows that all paths-state technique produces test cases with a high degree of coverage for the specifications of the implemented FICs comparing to test cases generated using round-trip path and all-transition techniques.

Keywords: Hooks, object-oriented framework, frameworkinterface classes (FICs), specification-based testing, test casegeneration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1321
139 An Application of Extreme Value Theory as a Risk Measurement Approach in Frontier Markets

Authors: Dany Ng Cheong Vee, Preethee Nunkoo Gonpot, Noor-Ul-Hacq Sookia

Abstract:

In this paper, we consider the application of Extreme Value Theory as a risk measurement tool. The Value at Risk, for a set of indices, from six Stock Exchanges of Frontier markets is calculated using the Peaks over Threshold method and the performance of the model index-wise is evaluated using coverage tests and loss functions. Our results show that “fattailedness” alone of the data is not enough to justify the use of EVT as a VaR approach. The structure of the returns dynamics is also a determining factor. This approach works fine in markets which have had extremes occurring in the past thus making the model capable of coping with extremes coming up (Colombo, Tunisia and Zagreb Stock Exchanges). On the other hand, we find that indices with lower past than present volatility fail to adequately deal with future extremes (Mauritius and Kazakhstan). We also conclude that using EVT alone produces quite static VaR figures not reflecting the actual dynamics of the data.

Keywords: Extreme Value theory, Financial Crisis 2008, Frontier Markets, Value at Risk.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2379
138 Li-Fi Technology: Data Transmission through Visible Light

Authors: Shahzad Hassan, Kamran Saeed

Abstract:

People are always in search of Wi-Fi hotspots because Internet is a major demand nowadays. But like all other technologies, there is still room for improvement in the Wi-Fi technology with regards to the speed and quality of connectivity. In order to address these aspects, Harald Haas, a professor at the University of Edinburgh, proposed what we know as the Li-Fi (Light Fidelity). Li-Fi is a new technology in the field of wireless communication to provide connectivity within a network environment. It is a two-way mode of wireless communication using light. Basically, the data is transmitted through Light Emitting Diodes which can vary the intensity of light very fast, even faster than the blink of an eye. From the research and experiments conducted so far, it can be said that Li-Fi can increase the speed and reliability of the transfer of data. This paper pays particular attention on the assessment of the performance of this technology. In other words, it is a 5G technology which uses LED as the medium of data transfer. For coverage within the buildings, Wi-Fi is good but Li-Fi can be considered favorable in situations where large amounts of data are to be transferred in areas with electromagnetic interferences. It brings a lot of data related qualities such as efficiency, security as well as large throughputs to the table of wireless communication. All in all, it can be said that Li-Fi is going to be a future phenomenon where the presence of light will mean access to the Internet as well as speedy data transfer.

Keywords: Communication, LED, Li-Fi, Wi-Fi.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2156
137 Discussing Embedded versus Central Machine Learning in Wireless Sensor Networks

Authors: Anne-Lena Kampen, Øivind Kure

Abstract:

Machine learning (ML) can be implemented in Wireless Sensor Networks (WSNs) as a central solution or distributed solution where the ML is embedded in the nodes. Embedding improves privacy and may reduce prediction delay. In addition, the number of transmissions is reduced. However, quality factors such as prediction accuracy, fault detection efficiency and coordinated control of the overall system suffer. Here, we discuss and highlight the trade-offs that should be considered when choosing between embedding and centralized ML, especially for multihop networks. In addition, we present estimations that demonstrate the energy trade-offs between embedded and centralized ML. Although the total network energy consumption is lower with central prediction, it makes the network more prone for partitioning due to the high forwarding load on the one-hop nodes. Moreover, the continuous improvements in the number of operations per joule for embedded devices will move the energy balance toward embedded prediction.

Keywords: Central ML, embedded machine learning, energy consumption, local ML, Wireless Sensor Networks, WSN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 817
136 Neural Network Based Icing Identification and Fault Tolerant Control of a 340 Aircraft

Authors: F. Caliskan

Abstract:

This paper presents a Neural Network (NN) identification of icing parameters in an A340 aircraft and a reconfiguration technique to keep the A/C performance close to the performance prior to icing. Five aircraft parameters are assumed to be considerably affected by icing. The off-line training for identifying the clear and iced dynamics is based on the Levenberg-Marquard Backpropagation algorithm. The icing parameters are located in the system matrix. The physical locations of the icing are assumed at the right and left wings. The reconfiguration is based on the technique known as the control mixer approach or pseudo inverse technique. This technique generates the new control input vector such that the A/C dynamics is not much affected by icing. In the simulations, the longitudinal and lateral dynamics of an Airbus A340 aircraft model are considered, and the stability derivatives affected by icing are identified. The simulation results show the successful NN identification of the icing parameters and the reconfigured flight dynamics having the similar performance before the icing. In other words, the destabilizing icing affect is compensated.

Keywords: Aircraft Icing, Stability Derivatives, Neural NetworkIdentification, Reconfiguration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1699
135 Real Time Remote Monitoring and Fault Detection in Wind Turbine

Authors: Saad Chakkor, Mostafa Baghouri, Abderrahmane Hajraoui

Abstract:

In new energy development, wind power has boomed. It is due to the proliferation of wind parks and their operation in supplying the national electric grid with low cost and clean resources. Hence, there is an increased need to establish a proactive maintenance for wind turbine machines based on remote control and monitoring. That is necessary with a real-time wireless connection in offshore or inaccessible locations while the wired method has many flaws. The objective of this strategy is to prolong wind turbine lifetime and to increase productivity. The hardware of a remote control and monitoring system for wind turbine parks is designed. It takes advantage of GPRS or Wi-Max wireless module to collect data measurements from different wind machine sensors through IP based multi-hop communication. Computer simulations with Proteus ISIS and OPNET software tools have been conducted to evaluate the performance of the studied system. Study findings show that the designed device is suitable for application in a wind park.

Keywords: Embedded System, Monitoring, Wind Turbine, Faults Diagnosis, TCP/IP Protocol, Real Time, Web.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3972
134 On Mobile Checkpointing using Index and Time Together

Authors: Awadhesh Kumar Singh

Abstract:

Checkpointing is one of the commonly used techniques to provide fault-tolerance in distributed systems so that the system can operate even if one or more components have failed. However, mobile computing systems are constrained by low bandwidth, mobility, lack of stable storage, frequent disconnections and limited battery life. Hence, checkpointing protocols having lesser number of synchronization messages and fewer checkpoints are preferred in mobile environment. There are two different approaches, although not orthogonal, to checkpoint mobile computing systems namely, time-based and index-based. Our protocol is a fusion of these two approaches, though not first of its kind. In the present exposition, an index-based checkpointing protocol has been developed, which uses time to indirectly coordinate the creation of consistent global checkpoints for mobile computing systems. The proposed algorithm is non-blocking, adaptive, and does not use any control message. Compared to other contemporary checkpointing algorithms, it is computationally more efficient because it takes lesser number of checkpoints and does not need to compute dependency relationships. A brief account of important and relevant works in both the fields, time-based and index-based, has also been included in the presentation.

Keywords: Checkpointing, forced checkpoint, mobile computing, recovery, time-coordinated.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1482
133 Reliability Modeling and Data Analysis of Vacuum Circuit Breaker Subject to Random Shocks

Authors: Rafik Medjoudj, Rabah Medjoudj, D. Aissani

Abstract:

The electrical substation components are often subject to degradation due to over-voltage or over-current, caused by a short circuit or a lightning. A particular interest is given to the circuit breaker, regarding the importance of its function and its dangerous failure. This component degrades gradually due to the use, and it is also subject to the shock process resulted from the stress of isolating the fault when a short circuit occurs in the system. In this paper, based on failure mechanisms developments, the wear out of the circuit breaker contacts is modeled. The aim of this work is to evaluate its reliability and consequently its residual lifetime. The shock process is based on two random variables such as: the arrival of shocks and their magnitudes. The arrival of shocks was modeled using homogeneous Poisson process (HPP). By simulation, the dates of short-circuit arrivals were generated accompanied with their magnitudes. The same principle of simulation is applied to the amount of cumulative wear out contacts. The objective reached is to find the formulation of the wear function depending on the number of solicitations of the circuit breaker.

Keywords: reliability, short-circuit, models of shocks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930
132 Identifying Relationships between Technology-based Services and ICTs: A Patent Analysis Approach

Authors: Chulhyun Kim, Seungkyum Kim, Moon-soo Kim

Abstract:

A variety of new technology-based services have emerged with the development of Information and Communication Technologies (ICTs). Since technology-based services have technology-driven characteristics, the identification of relationships between technology-based services and ICTs would give meaningful implications. Thus, this paper proposes an approach for identifying the relationships between technology-based services and ICTs by analyzing patent documents. First, business model (BM) patents are classified into relevant service categories. Second, patent citation analysis is conducted to investigate the technological linkage and impacts between technology-based services and ICTs at macro level. Third, as a micro level analysis, patent co-classification analysis is employed to identify the technological linkage and coverage. The proposed approach could guide and help managers and designers of technology-based services to discover the opportunity of the development of new technology-based services in emerging service sectors.

Keywords: Technology-based Services, Information and Communication Technology (ICT), Business Model (BM) Patent, Patent Analysis, Technological Relationship

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1977
131 Evaluation Process for the Hardware Safety Integrity Level

Authors: Sung Kyu Kim, Yong Soo Kim

Abstract:

Safety instrumented systems (SISs) are becoming increasingly complex and the proportion of programmable electronic parts is growing. The IEC 61508 global standard was established to ensure the functional safety of SISs, but it was expressed in highly macroscopic terms. This study introduces an evaluation process for hardware safety integrity levels through failure modes, effects, and diagnostic analysis (FMEDA).FMEDA is widely used to evaluate safety levels, and it provides the information on failure rates and failure mode distributions necessary to calculate a diagnostic coverage factor for a given component. In our evaluation process, the components of the SIS subsystem are first defined in terms of failure modes and effects. Then, the failure rate and failure mechanism distribution are assigned to each component. The safety mode and detectability of each failure mode are determined for each component. Finally, the hardware safety integrity level is evaluated based on the calculated results.

Keywords: Safety instrumented system; Safety integrity level; Failure modes, effects, and diagnostic analysis; IEC 61508.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2523
130 Monte Carlo Analysis and Fuzzy Sets for Uncertainty Propagation in SIS Performance Assessment

Authors: Fares Innal, Yves Dutuit, Mourad Chebila

Abstract:

The object of this work is the probabilistic performance evaluation of safety instrumented systems (SIS), i.e. the average probability of dangerous failure on demand (PFDavg) and the average frequency of failure (PFH), taking into account the uncertainties related to the different parameters that come into play: failure rate (λ), common cause failure proportion (β), diagnostic coverage (DC)... This leads to an accurate and safe assessment of the safety integrity level (SIL) inherent to the safety function performed by such systems. This aim is in keeping with the requirement of the IEC 61508 standard with respect to handling uncertainty. To do this, we propose an approach that combines (1) Monte Carlo simulation and (2) fuzzy sets. Indeed, the first method is appropriate where representative statistical data are available (using pdf of the relating parameters), while the latter applies in the case characterized by vague and subjective information (using membership function). The proposed approach is fully supported with a suitable computer code.

Keywords: Fuzzy sets, Monte Carlo simulation, Safety instrumented system, Safety integrity level.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2774
129 Optical Repeater Assisted Visible Light Device-to-Device Communications

Authors: Samrat Vikramaditya Tiwari, Atul Sewaiwar, Yeon-Ho Chung

Abstract:

Device-to-device (D2D) communication is considered a promising technique to provide wireless peer-to-peer communication services. Due to increasing demand on mobile services, available spectrum for radio frequency (RF) based communications becomes scarce. Recently, visible light communications (VLC) has evolved as a high speed wireless data transmission technology for indoor environments with abundant available bandwidth. In this paper, a novel VLC based D2D communication that provides wireless peer-to-peer communication is proposed. Potential low operating power devices for an efficient D2D communication over increasing distance of separation between devices is analyzed. Optical repeaters (OR) are also proposed to enhance the performance in an environment where direct D2D communications yield degraded performance. Simulation results show that VLC plays an important role in providing efficient D2D communication up to a distance of 1 m between devices. It is also found that the OR significantly improves the coverage distance up to 3.5 m.

Keywords: Visible light communication, light emitting diode, device-to-device, optical repeater.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2100
128 Investigations into Effect of Neural Network Predictive Control of UPFC for Improving Transient Stability Performance of Multimachine Power System

Authors: Sheela Tiwari, R. Naresh, R. Jha

Abstract:

The paper presents an investigation in to the effect of neural network predictive control of UPFC on the transient stability performance of a multimachine power system. The proposed controller consists of a neural network model of the test system. This model is used to predict the future control inputs using the damped Gauss-Newton method which employs ‘backtracking’ as the line search method for step selection. The benchmark 2 area, 4 machine system that mimics the behavior of large power systems is taken as the test system for the study and is subjected to three phase short circuit faults at different locations over a wide range of operating conditions. The simulation results clearly establish the robustness of the proposed controller to the fault location, an increase in the critical clearing time for the circuit breakers, and an improved damping of the power oscillations as compared to the conventional PI controller.

Keywords: Identification, Neural networks, Predictive control, Transient stability, UPFC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2072
127 Alignment of MG-63 Osteoblasts on Fibronectin-Coated Phosphorous Doping Lattices in Silicon

Authors: Andreas Körtge, Susanne Stählke, Regina Lange, Mario Birkholz, Mirko Fraschke, Katrin Schulz, Barbara Nebe, Patrick Elter

Abstract:

A major challenge in biomaterials research is the regulation of protein adsorption which is a key factor for controlling the subsequent cell adhesion at implant surfaces. The aim of the present study was to control the adsorption of fibronectin (FN) and the attachment of MG-63 osteoblasts with an electronic nanostructure. Shallow doping line lattices with a period of 260 nm were produced for this purpose by implantation of phosphorous in silicon wafers. Protein coverage was determined after incubating the substrate with FN by means of an immunostaining procedure and the measurement of the fluorescence intensity with a TECAN analyzer. We observed an increased amount of adsorbed FN on the nanostructure compared to control substrates. MG-63 osteoblasts were cultivated for 24h on FN-incubated substrates and their morphology was assessed by SEM. Preferred orientation and elongation of the cells in direction of the doping lattice lines was observed on FN-coated nanostructures.

Keywords: Cell adhesion, electronic nanostructures, doping lattice, fibronectin, MG-63 osteoblasts, protein adsorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2026
126 Climate Change Finger Prints in Mountainous Upper Euphrates Basin

Authors: Abdullah Gokhan Yilmaz, Monzur Alam Imteaz, Shirley Gato-Trinidad, Iqbal Hossain

Abstract:

Climate change leading to global warming affects the earth through many different ways such as weather (temperature, precipitation, humidity and the other parameters of weather), snow coverage and ice melting, sea level rise, hydrological cycles, quality of water, agriculture, forests, ecosystems and health. One of the most affected areas by climate change is hydrology and water resources. Regions where majority of runoff consists of snow melt are more sensitive to climate change. The first step of climate change studies is to establish trends of significant climate variables including precipitation, temperature and flow data to detect any potential climate change impacts already happened. Two popular non-parametric trend analysis methods, Mann-Kendal and Spearman-s Rho were applied to Upper Euphrates Basin (Turkey) to detect trends of precipitation, temperatures (maximum, minimum and average) and streamflow.

Keywords: Climate change, precipitation, snow hydrology, trend analysis and Upper Euphrates Basin

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 55585
125 Levels and Trends of Under-Five Mortality in South Africa from 1998 to 2012

Authors: T. Motsima, K. Zuma, E Rapoo

Abstract:

Childhood mortality is a key sign of the coverage of child survival interventions, social and economic progressions. Although the level of under-five mortality has been declining, it is still unacceptably high. The primary aim of this paper is to establish and analyse the levels and trends of under-five mortality for the periods 1998, 2003 and 2012 in South Africa. Methods: The data used for analysis came from the 1998 SADHS, the 2003 SADHS and the 2012 SABSSM which collected information on the survival status of children. The Kaplan-Meier estimate of the survival function method was used to determine the probabilities of failure (death) from birth up to 59 months. Results and Conclusion: The overall U5MR declined by 28.2% from 53.1 in 1998 to 38.1 in 2012. The U5MR of male children declined from 59.2 in 1998 to 46.2 in 2003 and dropped further to 41.4 in 2012. The U5MR of children of mothers aged 40 years and older increased from 64.0 in 1998 to 89.0 in 2003 and rose further to 129.9 in 2012. The U5MR of children of mothers with education level of 12 years or more increased from 32.2 in 1998 to 35.2 in 2003 and declined substantially to 17.5 in 2012.

Keywords: Demographic and health survey, Kaplan-Meier, levels and trends, under-five mortality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 690
124 Memory Leak Detection in Distributed System

Authors: Roohi Shabrin S., Devi Prasad B., Prabu D., Pallavi R. S., Revathi P.

Abstract:

Due to memory leaks, often-valuable system memory gets wasted and denied for other processes thereby affecting the computational performance. If an application-s memory usage exceeds virtual memory size, it can leads to system crash. Current memory leak detection techniques for clusters are reactive and display the memory leak information after the execution of the process (they detect memory leak only after it occur). This paper presents a Dynamic Memory Monitoring Agent (DMMA) technique. DMMA framework is a dynamic memory leak detection, that detects the memory leak while application is in execution phase, when memory leak in any process in the cluster is identified by DMMA it gives information to the end users to enable them to take corrective actions and also DMMA submit the affected process to healthy node in the system. Thus provides reliable service to the user. DMMA maintains information about memory consumption of executing processes and based on this information and critical states, DMMA can improve reliability and efficaciousness of cluster computing.

Keywords: Dynamic Memory Monitoring Agent (DMMA), Cluster Computing, Memory Leak, Fault Tolerant Framework, Dynamic Memory Leak Detection (DMLD).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2274
123 Slovenian Text-to-Speech Synthesis for Speech User Interfaces

Authors: Jerneja Žganec Gros, Aleš Mihelič, Nikola Pavešić, Mario Žganec, Stanislav Gruden

Abstract:

The paper presents the design concept of a unitselection text-to-speech synthesis system for the Slovenian language. Due to its modular and upgradable architecture, the system can be used in a variety of speech user interface applications, ranging from server carrier-grade voice portal applications, desktop user interfaces to specialized embedded devices. Since memory and processing power requirements are important factors for a possible implementation in embedded devices, lexica and speech corpora need to be reduced. We describe a simple and efficient implementation of a greedy subset selection algorithm that extracts a compact subset of high coverage text sentences. The experiment on a reference text corpus showed that the subset selection algorithm produced a compact sentence subset with a small redundancy. The adequacy of the spoken output was evaluated by several subjective tests as they are recommended by the International Telecommunication Union ITU.

Keywords: text-to-speech synthesis, prosody modeling, speech user interface.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1450
122 Technique for Grounding System Design in Distribution Substation

Authors: N. Rugthaicharoencheep, A. Charlangsut, B. Ainsuk, A. Phayomhom

Abstract:

This paper presents the significant factor and give some suggestion that should know before design. The main objective of this paper is guide the first step for someone who attends to design of grounding system before study in details later. The overview of grounding system can protect damage from fault such as can save a human life and power system equipment. The unsafe conditions have three cases. Case 1) maximum touch voltage exceeds the safety criteria. In this case, the conductor compression ratio of the ground gird should be first adjusted to have optimal spacing of ground grid conductors. If it still over limit, earth resistivity should be consider afterward. Case 2) maximum step voltage exceeds the safety criteria. In this case, increasing the number of ground grid conductors around the boundary can solve this problem. Case 3) both of maximum touch and step voltage exceed the safety criteria. In this case, follow the solutions explained in case 1 and case 2. Another suggestion, vary depth of ground grid until maximum step and touch voltage do not exceed the safety criteria.

Keywords: Grounding System, Touch Voltage, Step Voltage, Safety Criteria.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3520
121 An Agent-Based Modelling Simulation Approach to Calculate Processing Delay of GEO Satellite Payload

Authors: V. Vicente E. Mujica, Gustavo Gonzalez

Abstract:

The global coverage of broadband multimedia and internet-based services in terrestrial-satellite networks demand particular interests for satellite providers in order to enhance services with low latencies and high signal quality to diverse users. In particular, the delay of on-board processing is an inherent source of latency in a satellite communication that sometimes is discarded for the end-to-end delay of the satellite link. The frame work for this paper includes modelling of an on-orbit satellite payload using an agent model that can reproduce the properties of processing delays. In essence, a comparison of different spatial interpolation methods is carried out to evaluate physical data obtained by an GEO satellite in order to define a discretization function for determining that delay. Furthermore, the performance of the proposed agent and the development of a delay discretization function are together validated by simulating an hybrid satellite and terrestrial network. Simulation results show high accuracy according to the characteristics of initial data points of processing delay for Ku bands.

Keywords: Terrestrial-satellite networks, latency, on-orbit satellite payload, simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 880
120 Scalable Cloud-Based LEO Satellite Constellation Simulator

Authors: Karim Sobh, Khaled El-Ayat, Fady Morcos, Amr El-Kadi

Abstract:

Distributed applications deployed on LEO satellites and ground stations require substantial communication between different members in a constellation to overcome the earth coverage barriers imposed by GEOs. Applications running on LEO constellations suffer the earth line-of-sight blockage effect. They need adequate lab testing before launching to space. We propose a scalable cloud-based network simulation framework to simulate problems created by the earth line-of-sight blockage. The framework utilized cloud IaaS virtual machines to simulate LEO satellites and ground stations distributed software. A factorial ANOVA statistical analysis is conducted to measure simulator overhead on overall communication performance. The results showed a very low simulator communication overhead. Consequently, the simulation framework is proposed as a candidate for testing LEO constellations with distributed software in the lab before space launch.

Keywords: LEO, Cloud Computing, Constellation, Satellite, Network Simulation, Netfilter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2572
119 Comparative Evaluation of Adaptive and Conventional Distance Relay for Parallel Transmission Line with Mutual Coupling

Authors: S.G. Srivani, Chandrasekhar Reddy Atla, K.P.Vittal

Abstract:

This paper presents the development of adaptive distance relay for protection of parallel transmission line with mutual coupling. The proposed adaptive relay, automatically adjusts its operation based on the acquisition of the data from distance relay of adjacent line and status of adjacent line from line circuit breaker IED (Intelligent Electronic Device). The zero sequence current of the adjacent parallel transmission line is used to compute zero sequence current ratio and the mutual coupling effect is fully compensated. The relay adapts to changing circumstances, like failure in communication from other relays and non - availability of adjacent transmission line. The performance of the proposed adaptive relay is tested using steady state and dynamic test procedures. The fault transients are obtained by simulating a realistic parallel transmission line system with mutual coupling effect in PSCAD. The evaluation test results show the efficacy of adaptive distance relay over the conventional distance relay.

Keywords: Adaptive relaying, distance measurement, mutualcoupling, quadrilateral trip characteristic, zones of protection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3141
118 A Consistency Protocol Multi-Layer for Replicas Management in Large Scale Systems

Authors: Ghalem Belalem, Yahya Slimani

Abstract:

Large scale systems such as computational Grid is a distributed computing infrastructure that can provide globally available network resources. The evolution of information processing systems in Data Grid is characterized by a strong decentralization of data in several fields whose objective is to ensure the availability and the reliability of the data in the reason to provide a fault tolerance and scalability, which cannot be possible only with the use of the techniques of replication. Unfortunately the use of these techniques has a height cost, because it is necessary to maintain consistency between the distributed data. Nevertheless, to agree to live with certain imperfections can improve the performance of the system by improving competition. In this paper, we propose a multi-layer protocol combining the pessimistic and optimistic approaches conceived for the data consistency maintenance in large scale systems. Our approach is based on a hierarchical representation model with tree layers, whose objective is with double vocation, because it initially makes it possible to reduce response times compared to completely pessimistic approach and it the second time to improve the quality of service compared to an optimistic approach.

Keywords: Data Grid, replication, consistency, optimistic approach, pessimistic approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1568
117 Double Manifold Sliding Mode Observer for Sensorless Control of Multiphase Induction Machine under Fault Condition

Authors: Mohammad Jafarifar

Abstract:

Multiphase Induction Machine (IM) is normally controlled using rotor field oriented vector control. Under phase(s) loss, the machine currents can be optimally controlled to satisfy certain optimization criteria. In this paper we discuss the performance of double manifold sliding mode observer (DM-SMO) in Sensorless control of multiphase induction machine under unsymmetrical condition (one phase loss). This observer is developed using the IM model in the stationary reference frame. DM-SMO is constructed by adding extra feedback term to conventional single mode sliding mode observer (SM-SMO) which proposed in many literature. This leads to a fully convergent observer that also yields an accurate estimate of the speed and stator currents. It will be shown by the simulation results that the estimated speed and currents by the method are very well and error between real and estimated quantities is negligible. Also parameter sensitivity analysis shows that this method is rather robust against parameter variation.

Keywords: Multiphase induction machine, field oriented control, sliding mode, unsymmetrical condition, manifold.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1813
116 Performance Improvement of Information System of a Banking System Based on Integrated Resilience Engineering Design

Authors: S. H. Iranmanesh, L. Aliabadi, A. Mollajan

Abstract:

Integrated resilience engineering (IRE) is capable of returning banking systems to the normal state in extensive economic circumstances. In this study, information system of a large bank (with several branches) is assessed and optimized under severe economic conditions. Data envelopment analysis (DEA) models are employed to achieve the objective of this study. Nine IRE factors are considered to be the outputs, and a dummy variable is defined as the input of the DEA models. A standard questionnaire is designed and distributed among executive managers to be considered as the decision-making units (DMUs). Reliability and validity of the questionnaire is examined based on Cronbach's alpha and t-test. The most appropriate DEA model is determined based on average efficiency and normality test. It is shown that the proposed integrated design provides higher efficiency than the conventional RE design. Results of sensitivity and perturbation analysis indicate that self-organization, fault tolerance, and reporting culture respectively compose about 50 percent of total weight.

Keywords: Banking system, data envelopment analysis, DEA, integrated resilience engineering, IRE, performance evaluation, perturbation analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 832
115 Design and Performance Evaluation of Hybrid Corrugated-GFRP Infill Panels

Authors: WooYoung Jung, HoYoung Son

Abstract:

This study presented to reduce earthquake damage and emergency rehabilitation of critical structures such as schools, hightech factories, and hospitals due to strong ground motions associated with climate changes. Regarding recent trend, a strong earthquake causes serious damage to critical structures and then the critical structure might be influenced by sequence aftershocks (or tsunami) due to fault plane adjustments. Therefore, in order to improve seismic performance of critical structures, retrofitted or strengthening study of the structures under aftershocks sequence after emergency rehabilitation of the structures subjected to strong earthquakes is widely carried out. Consequently, this study used composite material for emergency rehabilitation of the structure rather than concrete and steel materials because of high strength and stiffness, lightweight, rapid manufacturing, and dynamic performance. Also, this study was to develop or improve the seismic performance or seismic retrofit of critical structures subjected to strong ground motions and earthquake aftershocks, by utilizing GFRP-Corrugated Infill Panels (GCIP).

Keywords: Composite material, GFRP, Infill Panel, Aftershock, Seismic Retrofitting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2276
114 Automated Process Quality Monitoring with Prediction of Fault Condition Using Measurement Data

Authors: Hyun-Woo Cho

Abstract:

Detection of incipient abnormal events is important to improve safety and reliability of machine operations and reduce losses caused by failures. Improper set-ups or aligning of parts often leads to severe problems in many machines. The construction of prediction models for predicting faulty conditions is quite essential in making decisions on when to perform machine maintenance. This paper presents a multivariate calibration monitoring approach based on the statistical analysis of machine measurement data. The calibration model is used to predict two faulty conditions from historical reference data. This approach utilizes genetic algorithms (GA) based variable selection, and we evaluate the predictive performance of several prediction methods using real data. The results shows that the calibration model based on supervised probabilistic principal component analysis (SPPCA) yielded best performance in this work. By adopting a proper variable selection scheme in calibration models, the prediction performance can be improved by excluding non-informative variables from their model building steps.

Keywords: Prediction, operation monitoring, on-line data, nonlinear statistical methods, empirical model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1654
113 Simultaneous Tuning of Static Var Compensator and Power System Stabilizer Employing Real- Coded Genetic Algorithm

Authors: S. Panda, N. P. Patidar, R. Singh

Abstract:

Power system stability enhancement by simultaneous tuning of a Power System Stabilizer (PSS) and a Static Var Compensator (SVC)-based controller is thoroughly investigated in this paper. The coordination among the proposed damping stabilizers and the SVC internal voltage regulators has also been taken into consideration. The design problem is formulated as an optimization problem with a time-domain simulation-based objective function and Real-Coded Genetic Algorithm (RCGA) is employed to search for optimal controller parameters. The proposed stabilizers are tested on a weakly connected power system with different disturbances and loading conditions. The nonlinear simulation results are presented to show the effectiveness and robustness of the proposed control schemes over a wide range of loading conditions and disturbances. Further, the proposed design approach is found to be robust and improves stability effectively even under small disturbance and unbalanced fault conditions.

Keywords: Real-Coded Genetic Algorithm (RCGA), Static Var Compensator (SVC), Power System Stabilizer (PSS), Low Frequency Oscillations, Power System Stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2248
112 Application of Genetic Algorithm for FACTS-based Controller Design

Authors: Sidhartha Panda, N. P. Padhy, R.N.Patel

Abstract:

In this paper, genetic algorithm (GA) opmization technique is applied to design Flexible AC Transmission System (FACTS)-based damping controllers. Two types of controller structures, namely a proportional-integral (PI) and a lead-lag (LL) are considered. The design problem of the proposed controllers is formulated as an optimization problem and GA is employed to search for optimal controller parameters. By minimizing the time-domain based objective function, in which the deviation in the oscillatory rotor speed of the generator is involved; stability performance of the system is improved. The proposed controllers are tested on a weakly connected power system subjected to different disturbances. The non-linear simulation results are presented to show the effectiveness of the proposed controller and their ability to provide efficient damping of low frequency oscillations. It is also observed that the proposed SSSC-based controllers improve greatly the voltage profile of the system under severe disturbances. Further, the dynamic performances of both the PI and LL structured FACTS-controller are analyzed at different loading conditions and under various disturbance condition as well as under unbalanced fault conditions..

Keywords: Genetic algorithm, proportional-integral controller, lead-lag controller, power system stability, FACTS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2534