Search results for: reliability allocation
2139 Studying the Beginnings of Strategic Behavior
Authors: Taher Abofol, Yaakov Kareev, Judith Avrahami, Peter M. Todd
Abstract:
Are children sensitive to their relative strength in competitions against others? Performance on tasks that require cooperation or coordination (e.g. the Ultimatum Game) indicates that early precursors of adult-like notions of fairness and reciprocity, as well as altruistic behavior, are evident at an early age. However, not much is known regarding developmental changes in interactive decision-making, especially in competitive interactions. Thus, it is important to study the developmental aspects of strategic behavior in these situations. The present research focused on cognitive-developmental changes in a competitive interaction. Specifically, it aimed at revealing how children engage in strategic interactions that involve the allocation of limited resources over a number of fields of competition, by manipulating relative strength. Relative strength refers to situations in which player strength changes midway through the game: the stronger player becomes the weaker one, while the weaker player becomes the stronger one. An experiment was conducted to find out if the behavior of children of different age groups differs in the following three aspects: 1. Perception of relative strength. 2. Ability to learn while gaining experience. 3. Ability to adapt to change in relative strength. The task was composed of a resource allocation game. After the players allocated their resources (privately and simultaneously), a competition field was randomly chosen for each player. The player who allocated more resources to the field chosen was declared the winner of that round. The resources available to the two competitors were unequal (or equal, for control). The theoretical solution for this game is that the weaker player should give up on a certain number of fields, depending on the stronger opponent’s relative strength, in order to be able to compete with the opponent on equal footing in the remaining fields. Participants were of three age groups, first-graders (N = 36, mean age = 6), fourth-graders (N = 36, mean age = 10), and eleventh-graders (N = 72, mean age = 16). The games took place between players of the same age and lasted for 16 rounds. There were two experimental conditions – a control condition, in which players were of equal strength, and an experimental condition, in which players differed in strength. In the experimental condition, players' strength was changed midway through the session. Results indicated that players in all age groups were sensitive to their relative strength, and played in line with the theoretical solution: the weaker players gave up on more fields than the stronger ones. This understanding, as well as the consequent difference in allocation between weak and strong players, was more pronounced among older participants. Experience led only to minimal behavioral change. Finally, the children from the two older groups, particularly the eleventh graders adapted quickly to the midway switch in relative strength. In contrast, the first-graders hardly changed their behavior with the change in their relative strength, indicating a limited ability to adapt. These findings highlight young children’s ability to consider their relative strength in strategic interactions and its boundaries.Keywords: children, competition, decision making, developmental changes, strategic behavior
Procedia PDF Downloads 3122138 Neuro-Fuzzy Approach to Improve Reliability in Auxiliary Power Supply System for Nuclear Power Plant
Authors: John K. Avor, Choong-Koo Chang
Abstract:
The transfer of electrical loads at power generation stations from Standby Auxiliary Transformer (SAT) to Unit Auxiliary Transformer (UAT) and vice versa is through a fast bus transfer scheme. Fast bus transfer is a time-critical application where the transfer process depends on various parameters, thus transfer schemes apply advance algorithms to ensure power supply reliability and continuity. In a nuclear power generation station, supply continuity is essential, especially for critical class 1E electrical loads. Bus transfers must, therefore, be executed accurately within 4 to 10 cycles in order to achieve safety system requirements. However, the main problem is that there are instances where transfer schemes scrambled due to inaccurate interpretation of key parameters; and consequently, have failed to transfer several critical loads from UAT to the SAT during main generator trip event. Although several techniques have been adopted to develop robust transfer schemes, a combination of Artificial Neural Network and Fuzzy Systems (Neuro-Fuzzy) has not been extensively used. In this paper, we apply the concept of Neuro-Fuzzy to determine plant operating mode and dynamic prediction of the appropriate bus transfer algorithm to be selected based on the first cycle of voltage information. The performance of Sequential Fast Transfer and Residual Bus Transfer schemes was evaluated through simulation and integration of the Neuro-Fuzzy system. The objective for adopting Neuro-Fuzzy approach in the bus transfer scheme is to utilize the signal validation capabilities of artificial neural network, specifically the back-propagation algorithm which is very accurate in learning completely new systems. This research presents a combined effect of artificial neural network and fuzzy systems to accurately interpret key bus transfer parameters such as magnitude of the residual voltage, decay time, and the associated phase angle of the residual voltage in order to determine the possibility of high speed bus transfer for a particular bus and the corresponding transfer algorithm. This demonstrates potential for general applicability to improve reliability of the auxiliary power distribution system. The performance of the scheme is implemented on APR1400 nuclear power plant auxiliary system.Keywords: auxiliary power system, bus transfer scheme, fuzzy logic, neural networks, reliability
Procedia PDF Downloads 1712137 Analysis of Fault Tolerance on Grid Computing in Real Time Approach
Authors: Parampal Kaur, Deepak Aggarwal
Abstract:
In the computational Grid, fault tolerance is an imperative issue to be considered during job scheduling. Due to the widespread use of resources, systems are highly prone to errors and failures. Hence, fault tolerance plays a key role in the grid to avoid the problem of unreliability. Scheduling the task to the appropriate resource is a vital requirement in computational Grid. The fittest resource scheduling algorithm searches for the appropriate resource based on the job requirements, in contrary to the general scheduling algorithms where jobs are scheduled to the resources with best performance factor. The proposed method is to improve the fault tolerance of the fittest resource scheduling algorithm by scheduling the job in coordination with job replication when the resource has low reliability. Based on the reliability index of the resource, the resource is identified as critical. The tasks are scheduled based on the criticality of the resources. Results show that the execution time of the tasks is comparatively reduced with the proposed algorithm using real-time approach rather than a simulator.Keywords: computational grid, fault tolerance, task replication, job scheduling
Procedia PDF Downloads 4362136 A Study on the Accelerated Life Cycle Test Method of the Motor for Home Appliances by Using Acceleration Factor
Authors: Youn-Sung Kim, Mi-Sung Kim, Jae-Kun Lee
Abstract:
This paper deals with the accelerated life cycle test method of the motor for home appliances that demand high reliability. Life Cycle of parts in home appliances also should be 10 years because life cycle of the home appliances such as washing machine, refrigerator, TV is at least 10 years. In case of washing machine, the life cycle test method of motor is advanced for 3000 cycle test (1cycle = 2hours). However, 3000 cycle test incurs loss for the time and cost. Objectives of this study are to reduce the life cycle test time and the number of test samples, which could be realized by using acceleration factor for the test time and reduction factor for the number of sample.Keywords: accelerated life cycle test, motor reliability test, motor for washing machine, BLDC motor
Procedia PDF Downloads 6342135 The Design and Implementation of an Enhanced 2D Mesh Switch
Authors: Manel Langar, Riad Bourguiba, Jaouhar Mouine
Abstract:
In this paper, we propose the design and implementation of an enhanced wormhole virtual channel on chip router. It is a heart of a mesh NoC using the XY deterministic routing algorithm. It is characterized by its simple virtual channel allocation strategy which allows reducing area and complexity of connections without affecting the performance. We implemented our router on a Tezzaron process to validate its performances. This router is a basic element that will be used later to design a 3D mesh NoC.Keywords: NoC, mesh, router, 3D NoC
Procedia PDF Downloads 5682134 Derivation of a Risk-Based Level of Service Index for Surface Street Network Using Reliability Analysis
Authors: Chang-Jen Lan
Abstract:
Current Level of Service (LOS) index adopted in Highway Capacity Manual (HCM) for signalized intersections on surface streets is based on the intersection average delay. The delay thresholds for defining LOS grades are subjective and is unrelated to critical traffic condition. For example, an intersection delay of 80 sec per vehicle for failing LOS grade F does not necessarily correspond to the intersection capacity. Also, a specific measure of average delay may result from delay minimization, delay equality, or other meaningful optimization criteria. To that end, a reliability version of the intersection critical degree of saturation (v/c) as the LOS index is introduced. Traditionally, the level of saturation at a signalized intersection is defined as the ratio of critical volume sum (per lane) to the average saturation flow (per lane) during all available effective green time within a cycle. The critical sum is the sum of the maximal conflicting movement-pair volumes in northbound-southbound and eastbound/westbound right of ways. In this study, both movement volume and saturation flow are assumed log-normal distributions. Because, when the conditions of central limit theorem obtain, multiplication of the independent, positive random variables tends to result in a log-normal distributed outcome in the limit, the critical degree of saturation is expected to be a log-normal distribution as well. Derivation of the risk index predictive limits is complex due to the maximum and absolute value operators, as well as the ratio of random variables. A fairly accurate functional form for the predictive limit at a user-specified significant level is yielded. The predictive limit is then compared with the designated LOS thresholds for the intersection critical degree of saturation (denoted as XKeywords: reliability analysis, level of service, intersection critical degree of saturation, risk based index
Procedia PDF Downloads 1312133 Reliability of 2D Motion Analysis System for Sagittal Plane Lower Limb Kinematics during Running
Authors: Seyed Hamed Mousavi, Juha M. Hijmans, Reza Rajabi, Ron Diercks, Johannes Zwerver, Henk van der Worp
Abstract:
Introduction: Running is one of the most popular sports activity among people. Improper sagittal plane ankle, knee and hip kinematics are considered to be associated with the increase of injury risk in runners. Motion assessing smart-phone applications are increasingly used to measure kinematics both in the field and laboratory setting, as they are cheaper, more portable, accessible, and easier to use relative to 3D motion analysis system. The aims of this study are 1) to compare the results of 3D gait analysis system and CE; 2) to evaluate the test-retest and intra-rater reliability of coach’s eye (CE) app for the sagittal plane hip, knee, and ankle angles in the touchdown and toe-off while running. Method: Twenty subjects participated in this study. Sixteen reflective markers and cluster markers were attached to the subject’s body. Subjects were asked to run at a self-selected speed on a treadmill. Twenty-five seconds of running were collected for analyzing kinematics of interest. To measure sagittal plane hip, knee and ankle joint angles at touchdown (TD) and toe off (TO), the mean of first ten acceptable consecutive strides was calculated for each angle. A smartphone (Samsung Note5, android) was placed on the right side of the subject so that whole body was simultaneously filmed with 3D gait system during running. All subjects repeated the task with the same running speed after a short interval of 5 minutes in between. The CE app, installed on the smartphone, was used to measure the sagittal plane hip, knee and ankle joint angles at touchdown and toe off the stance phase. Results: Intraclass correlation coefficient (ICC) was used to assess test-retest and intra-rater reliability. To analyze the agreement between 3D and 2D outcomes, the Bland and Altman plot was used. The values of ICC were for Ankle at TD (TRR=0.8,IRR=0.94), ankle at TO (TRR=0.9,IRR=0.97), knee at TD (TRR=0.78,IRR=0.98), knee at TO (TRR=0.9,IRR=0.96), hip at TD (TRR=0.75,IRR=0.97), hip at TO (TRR=0.87,IRR=0.98). The Bland and Altman plots displaying a mean difference (MD) and ±2 standard deviation of MD (2SDMD) of 3D and 2D outcomes were for Ankle at TD (MD=3.71,+2SDMD=8.19, -2SDMD=-0.77), ankle at TO (MD=-1.27, +2SDMD=6.22, -2SDMD=-8.76), knee at TD (MD=1.48, +2SDMD=8.21, -2SDMD=-5.25), knee at TO (MD=-6.63, +2SDMD=3.94, -2SDMD=-17.19), hip at TD (MD=1.51, +2SDMD=9.05, -2SDMD=-6.03), hip at TO (MD=-0.18, +2SDMD=12.22, -2SDMD=-12.59). Discussion: The ability that the measurements are accurately reproduced is valuable in the performance and clinical assessment of outcomes of joint angles. The results of this study showed that the intra-rater and test-retest reliability of CE app for all kinematics measured are excellent (ICC ≥ 0.75). The Bland and Altman plots display that there are high differences of values for ankle at TD and knee at TO. Measuring ankle at TD by 2D gait analysis depends on the plane of movement. Since ankle at TD mostly occurs in the none-sagittal plane, the measurements can be different as foot progression angle at TD increases during running. The difference in values of the knee at TD can depend on how 3D and the rater detect the TO during the stance phase of running.Keywords: reliability, running, sagittal plane, two dimensional
Procedia PDF Downloads 2012132 Progress in Accuracy, Reliability and Safety in Firedamp Detection
Authors: José Luis Lorenzo Bayona, Ljiljana Medic-Pejic, Isabel Amez Arenillas, Blanca Castells Somoza
Abstract:
The communication presents the study results carried out by the Official Laboratory J. M. Madariaga (LOM) of the Polytechnic University of Madrid to analyze the reliability of methane detection systems used in underground mining. Poor firedamp control in work can cause from production stoppages to fatal accidents and since there is currently a great variety of equipment with different functional characteristics, a study is needed to indicate which measurement principles have the highest degree of confidence. For the development of the project, a series of fixed, transportable and portable methane detectors with different measurement principles have been selected to subject them to laboratory tests following the methods described in the applicable regulations. The test equipment has been the one usually used in the certification and calibration of these devices, subject to the LOM quality system, and the tests have been carried out on detectors accessible in the market. The conclusions establish the main advantages and disadvantages of the equipment according to the measurement principle used; catalytic combustion, interferometry and infrared absorption.Keywords: ATEX standards, gas detector, methane meter, mining safety
Procedia PDF Downloads 1372131 The Happiness Pulse: A Measure of Individual Wellbeing at a City Scale, Development and Validation
Authors: Rosemary Hiscock, Clive Sabel, David Manley, Sam Wren-Lewis
Abstract:
As part of the Happy City Index Project, Happy City have developed a survey instrument to measure experienced wellbeing: how people are feeling and functioning in their everyday lives. The survey instrument, called the Happiness Pulse, was developed in partnership with the New Economics Foundation (NEF) with the dual aim of collecting citywide wellbeing data and engaging individuals and communities in the measurement and promotion of their own wellbeing. The survey domains and items were selected through a review of the academic literature and a stakeholder engagement process, including local policymakers, community organisations and individuals. The Happiness Pulse was included in the Bristol pilot of the Happy City Index (n=722). The experienced wellbeing items were subjected to factor analysis. A reduced number of items to be included in a revised scale for future data collection were again entered into a factor analysis. These revised factors were tested for reliability and validity. Among items to be included in a revised scale for future data collection three factors emerged: Be, Do and Connect. The Be factor had good reliability, convergent and criterion validity. The Do factor had good discriminant validity. The Connect factor had adequate reliability and good discriminant and criterion validity. Some age, gender and socioeconomic differentiation was found. The properties of a new scale to measure experienced wellbeing, intended for use by municipal authorities, are described. Happiness Pulse data can be combined with local data on wellbeing conditions to determine what matters for peoples wellbeing across a city and why.Keywords: city wellbeing , community wellbeing, engaging individuals and communities, measuring wellbeing and happiness
Procedia PDF Downloads 2612130 Simultaneous Optimization of Design and Maintenance through a Hybrid Process Using Genetic Algorithms
Authors: O. Adjoul, A. Feugier, K. Benfriha, A. Aoussat
Abstract:
In general, issues related to design and maintenance are considered in an independent manner. However, the decisions made in these two sets influence each other. The design for maintenance is considered an opportunity to optimize the life cycle cost of a product, particularly in the nuclear or aeronautical field, where maintenance expenses represent more than 60% of life cycle costs. The design of large-scale systems starts with product architecture, a choice of components in terms of cost, reliability, weight and other attributes, corresponding to the specifications. On the other hand, the design must take into account maintenance by improving, in particular, real-time monitoring of equipment through the integration of new technologies such as connected sensors and intelligent actuators. We noticed that different approaches used in the Design For Maintenance (DFM) methods are limited to the simultaneous characterization of the reliability and maintainability of a multi-component system. This article proposes a method of DFM that assists designers to propose dynamic maintenance for multi-component industrial systems. The term "dynamic" refers to the ability to integrate available monitoring data to adapt the maintenance decision in real time. The goal is to maximize the availability of the system at a given life cycle cost. This paper presents an approach for simultaneous optimization of the design and maintenance of multi-component systems. Here the design is characterized by four decision variables for each component (reliability level, maintainability level, redundancy level, and level of monitoring data). The maintenance is characterized by two decision variables (the dates of the maintenance stops and the maintenance operations to be performed on the system during these stops). The DFM model helps the designers choose technical solutions for the large-scale industrial products. Large-scale refers to the complex multi-component industrial systems and long life-cycle, such as trains, aircraft, etc. The method is based on a two-level hybrid algorithm for simultaneous optimization of design and maintenance, using genetic algorithms. The first level is to select a design solution for a given system that considers the life cycle cost and the reliability. The second level consists of determining a dynamic and optimal maintenance plan to be deployed for a design solution. This level is based on the Maintenance Free Operating Period (MFOP) concept, which takes into account the decision criteria such as, total reliability, maintenance cost and maintenance time. Depending on the life cycle duration, the desired availability, and the desired business model (sales or rental), this tool provides visibility of overall costs and optimal product architecture.Keywords: availability, design for maintenance (DFM), dynamic maintenance, life cycle cost (LCC), maintenance free operating period (MFOP), simultaneous optimization
Procedia PDF Downloads 1182129 Contention Window Adjustment in IEEE 802.11-based Industrial Wireless Networks
Authors: Mohsen Maadani, Seyed Ahmad Motamedi
Abstract:
The use of wireless technology in industrial networks has gained vast attraction in recent years. In this paper, we have thoroughly analyzed the effect of contention window (CW) size on the performance of IEEE 802.11-based industrial wireless networks (IWN), from delay and reliability perspective. Results show that the default values of CWmin, CWmax, and retry limit (RL) are far from the optimum performance due to the industrial application characteristics, including short packet and noisy environment. An adaptive CW algorithm (payload-dependent) has been proposed to minimize the average delay. Finally a simple, but effective CW and RL setting has been proposed for industrial applications which outperforms the minimum-average-delay solution from maximum delay and jitter perspective, at the cost of a little higher average delay. Simulation results show an improvement of up to 20%, 25%, and 30% in average delay, maximum delay and jitter respectively.Keywords: average delay, contention window, distributed coordination function (DCF), jitter, industrial wireless network (IWN), maximum delay, reliability, retry limit
Procedia PDF Downloads 4152128 Psychological Testing in Industrial/Organizational Psychology: Validity and Reliability of Psychological Assessments in the Workplace
Authors: Melissa C. Monney
Abstract:
Psychological testing has been of interest to researchers for many years as useful tools in assessing and diagnosing various disorders as well as to assist in understanding human behavior. However, for over 20 years now, researchers and laypersons alike have been interested in using them for other purposes, such as determining factors in employee selection, promotion, and even termination. In recent years, psychological assessments have been useful in facilitating workplace decision processing, regarding employee circulation within organizations. This literature review explores four of the most commonly used psychological tests in workplace environments, namely cognitive ability, emotional intelligence, integrity, and personality tests, as organizations have used these tests to assess different factors of human behavior as predictive measures of future employee behaviors. The findings suggest that while there is much controversy and debate regarding the validity and reliability of these tests in workplace settings as they were not originally designed for these purposes, the use of such assessments in the workplace has been useful in decreasing costs and employee turnover as well as increase job satisfaction by ensuring the right employees are selected for their roles.Keywords: cognitive ability, personality testing, predictive validity, workplace behavior
Procedia PDF Downloads 2422127 The Attitude and Willingness to Use Telecare for Arthritis Patients
Authors: Jui-Chen Huang
Abstract:
Nowadays, the population is aging, the number of people who need to be taken care of is increased, but the manpower and funding are insufficient. Therefore, this study aims to explore the attitudes and willingness of arthritis patients to adopt telecare and to take a large medical institution in the central area of Taiwan as a sample hospital. A structured questionnaire (using the Likert five-point scale) was used to collect chronic patients over 20 years old as sample data, and a total of 500 valid questionnaires were effectively collected. The SPSS 18.0 statistical software was used for reliability analysis and independent sample t-test to explore the differences in attitudes and willingness to use telecare for arthritis patients and non-arthritic patients. The Cronbach's alpha value of this study questionnaire was above 0.94, showing good reliability. Arthritis patients and non-arthritic patients had statistically significant differences in attitudes toward telecare, while the willingness to use did not reach statistically significant differences. In addition, the average attitude and intention of arthritis patients for telecare are 3.38 and 3.41, respectively, indicating that arthritis patients have a certain degree of attitude and willingness to adopt telecare, which is worthy of follow-up research and practical industry push.Keywords: telecare, arthritis patients, attitudes, intention
Procedia PDF Downloads 1422126 Design and Assessment of Traffic Management Strategies for Improved Mobility on Major Arterial Roads in Lahore City
Authors: N. Ali, S. Nakayama, H. Yamaguchi, M. Nadeem
Abstract:
Traffic congestion is a matter of prime concern in developing countries. This can be primarily attributed due to poor design practices and biased allocation of resources based on political will neglecting the technical feasibilities in infrastructure design. During the last decade, Lahore has expanded at an unprecedented rate as compared to surrounding cities due to more funding and resource allocation by the previous governments. As a result of this, people from surrounding cities and areas moved to the Lahore city for better opportunities and quality of life. This migration inflow inherited the city with an increased population yielding the inefficiency of the existing infrastructure to accommodate enhanced traffic demand. This leads to traffic congestion on major arterial roads of the city. In this simulation study, a major arterial road was selected to evaluate the performance of the five intersections by changing the geometry of the intersections or signal control type. Simulations were done in two software; Highway Capacity Software (HCS) and Synchro Studio and Sim Traffic Software. Some of the traffic management strategies that were employed include actuated-signal control, semi-actuated signal control, fixed-time signal control, and roundabout. The most feasible solution for each intersection in the above-mentioned traffic management techniques was selected with the least delay time (seconds) and improved Level of Service (LOS). The results showed that Jinnah Hospital Intersection and Akbar Chowk Intersection improved 92.97% and 92.67% in delay time reduction, respectively. These results can be used by traffic planners and policy makers for decision making for the expansion of these intersections keeping in mind the traffic demand in future years.Keywords: traffic congestion, traffic simulation, traffic management, congestion problems
Procedia PDF Downloads 4702125 Stochastic Analysis of Linux Operating System through Copula Distribution
Authors: Vijay Vir Singh
Abstract:
This work is focused studying the Linux operating system connected in a LAN (local area network). The STAR topology (to be called subsystem-1) and BUS topology (to be called subsystem-2) are taken into account, which are placed at two different locations and connected to a server through a hub. In the both topologies BUS topology and STAR topology, we have assumed n clients. The system has two types of failures i.e. partial failure and complete failure. Further, the partial failure has been categorized as minor and major partial failure. It is assumed that the minor partial failure degrades the sub-systems and the major partial failure make the subsystem break down mode. The system may completely fail due to failure of server hacking and blocking etc. The system is studied using supplementary variable technique and Laplace transform by using different types of failure and two types of repair. The various measures of reliability for example, availability of system, reliability of system, MTTF, profit function for different parametric values have been discussed.Keywords: star topology, bus topology, blocking, hacking, Linux operating system, Gumbel-Hougaard family copula, supplementary variable
Procedia PDF Downloads 3702124 Strategies to Mitigate Disasters at the Hajj Religious Festival Using GIS and Agent Based Modelling
Authors: Muteb Alotaibi, Graham Clarke, Nick Malleson
Abstract:
The Hajj religious festival at Mina in Saudi Arabia has always presented the opportunity for injuries or deaths. For example, in 1990, a stampede killed 1426 pilgrims, whilst in 1997, 343 people were killed and 1500 injured due to a fire fuelled by high winds sweeping through the tent city in Mina.Many more minor incidents have occurred since then. It is predicted that 5 million pilgrims will soon perform the ritual at Mina (which is, in effect, a temporary city built each year in the desert), which might lead in the future to severe congestion and accidents unless the research is conducted on actions that contribute positively to improving the management of the crowd and facilitating the flow of pilgrims safely and securely. To help prevent further disasters, it is important to first plan better, more accessible locations for emergency services across Mina to ensure a good service for pilgrims. In this paper, we first use a Location Allocation Model (LAM) within a network GIS to examine the optimal locations for key services in the temporary city of Mina. This has been undertaken in relation to the location and movement of the pilgrims during the six day religious festival. The results of various what-if scenarios have been compared against the current location of services. A major argument is that planners should be flexible and locate facilities at different locations throughout the day and night. The use of location-allocation models in this type of comparative static mode has rarely been operationalised in the literature. Second, we model pilgrim movements and behaviours along with the most crowded parts of the network. This has been modelled using an agent-based model. This model allows planners to understand the key bottlenecks in the network and at what usage levels the paths become critically congested. Thus the paper has important implications and recommendations for future disaster planning strategies. This will enable planners to see at what stage in the movements of pilgrims problems occur in terms of potential crushes and trampling incidents. The main application of this research was only customised for pedestrians as the concentration only for pedestrians who move to Jamarat via foot. Further, the network in the middle of Mina was only dedicated for pedestrians for safety, so no Buses, trains and private cars were allowed in this area to prevent the congestion within this network. Initially, this research focus on Mina city as ‘temporary city’ and also about service provision in temporary cities, which is not highlighted in literature so far. Further, it is the first study which use the dynamic demand to optimise the services in the case of day and night time. Moreover, it is the first study which link the location allocation model for optimising services with ABM to find out whether or not the service location is located in the proper location in which it’s not affecting on crowd movement in mainstream flow where some pilgrims need to have health services.Keywords: ABM, crowd management, hajj, temporary city
Procedia PDF Downloads 1222123 An Investigation on Organisation Cyber Resilience
Authors: Arniyati Ahmad, Christopher Johnson, Timothy Storer
Abstract:
Cyber exercises used to assess the preparedness of a community against cyber crises, technology failures and critical information infrastructure (CII) incidents. The cyber exercises also called cyber crisis exercise or cyber drill, involved partnerships or collaboration of public and private agencies from several sectors. This study investigates organisation cyber resilience (OCR) of participation sectors in cyber exercise called X Maya in Malaysia. This study used a principal based cyber resilience survey called C-Suite Executive checklist developed by World Economic Forum in 2012. To ensure suitability of the survey to investigate the OCR, the reliability test was conducted on C-Suite Executive checklist items. The research further investigates the differences of OCR in ten Critical National Infrastructure Information (CNII) sectors participated in the cyber exercise. The One Way ANOVA test result showed a statistically significant difference of OCR among ten CNII sectors participated in the cyber exercise.Keywords: critical information infrastructure, cyber resilience, organisation cyber resilience, reliability test
Procedia PDF Downloads 3572122 Additive Weibull Model Using Warranty Claim and Finite Element Analysis Fatigue Analysis
Authors: Kanchan Mondal, Dasharath Koulage, Dattatray Manerikar, Asmita Ghate
Abstract:
This paper presents an additive reliability model using warranty data and Finite Element Analysis (FEA) data. Warranty data for any product gives insight to its underlying issues. This is often used by Reliability Engineers to build prediction model to forecast failure rate of parts. But there is one major limitation in using warranty data for prediction. Warranty periods constitute only a small fraction of total lifetime of a product, most of the time it covers only the infant mortality and useful life zone of a bathtub curve. Predicting with warranty data alone in these cases is not generally provide results with desired accuracy. Failure rate of a mechanical part is driven by random issues initially and wear-out or usage related issues at later stages of the lifetime. For better predictability of failure rate, one need to explore the failure rate behavior at wear out zone of a bathtub curve. Due to cost and time constraints, it is not always possible to test samples till failure, but FEA-Fatigue analysis can provide the failure rate behavior of a part much beyond warranty period in a quicker time and at lesser cost. In this work, the authors proposed an Additive Weibull Model, which make use of both warranty and FEA fatigue analysis data for predicting failure rates. It involves modeling of two data sets of a part, one with existing warranty claims and other with fatigue life data. Hazard rate base Weibull estimation has been used for the modeling the warranty data whereas S-N curved based Weibull parameter estimation is used for FEA data. Two separate Weibull models’ parameters are estimated and combined to form the proposed Additive Weibull Model for prediction.Keywords: bathtub curve, fatigue, FEA, reliability, warranty, Weibull
Procedia PDF Downloads 732121 Simplifying the Migration of Architectures in Embedded Applications Introducing a Pattern Language to Support the Workforce
Authors: Farha Lakhani, Michael J. Pont
Abstract:
There are two main architectures used to develop software for modern embedded systems: these can be labelled as “event-triggered” (ET) and “time-triggered” (TT). The research presented in this paper is concerned with the issues involved in migration between these two architectures. Although TT architectures are widely used in safety-critical applications they are less familiar to developers of mainstream embedded systems. The research presented in this paper began from the premise that–for a broad class of systems that have been implemented using an ET architecture–migration to a TT architecture would improve reliability. It may be tempting to assume that conversion between ET and TT designs will simply involve converting all event-handling software routines into periodic activities. However, the required changes to the software architecture are, in many cases rather more profound. The main contribution of the work presented in this paper is to identify ways in which the significant effort involved in migrating between existing ET architectures and “equivalent” (and effective) TT architectures could be reduced. The research described in this paper has taken an innovative step in this regard by introducing the use of ‘Design patterns’ for this purpose for the first time.Keywords: embedded applications, software architectures, reliability, pattern
Procedia PDF Downloads 3292120 On Reliability of a Credit Default Swap Contract during the EMU Debt Crisis
Authors: Petra Buzkova, Milos Kopa
Abstract:
Reliability of the credit default swap market had been questioned repeatedly during the EMU debt crisis. The article examines whether this development influenced sovereign EMU CDS prices in general. We regress the CDS market price on a model risk neutral CDS price obtained from an adopted reduced form valuation model in the 2009-2013 period. We look for a break point in the single-equation and multi-equation econometric models in order to show the changes in relations between CDS market and model prices. Our results differ according to the risk profile of a country. We find that in the case of riskier countries, the relationship between the market and model price changed when market participants started to question the ability of CDS contracts to protect their buyers. Specifically, it weakened after the change. In the case of less risky countries, the change happened earlier and the effect of a weakened relationship is not observed.Keywords: chow stability test, credit default swap, debt crisis, reduced form valuation model, seemingly unrelated regression
Procedia PDF Downloads 2622119 Structure Design of Vacuum Vessel with Large Openings for Spacecraft Thermal Vacuum Test
Authors: Han Xiao, Ruan Qi, Zhang Lei, Qi Yan
Abstract:
Space environment simulator is a facility used to conduct thermal test for spacecraft, and vacuum vessel is the main body of it. According to the requirements for thermal tests of the spacecraft and its solar array panels, the primary vessel and the side vessels are designed to be a combinative structure connected with aperture, which ratio reaches 0.7. Since the vacuum vessel suffers 0.1MPa external pressure during the process of thermal test, in order to ensure the simulator’s reliability and safety, it’s necessary to calculate the vacuum vessel’s intensity and stability. Based on the impact of large openings to vacuum vessel structure, this paper explored the reinforce design and analytical way of vacuum vessel with large openings, using a large space environment simulator’s vacuum vessel design as an example. Tests showed that the reinforce structure is effective to fulfill the requirements of external pressure and the gravity. This ensured the reliability of the space environment simulator, providing a guarantee for developing the spacecraft.Keywords: vacuum vessel, large opening, space environment simulator, structure design
Procedia PDF Downloads 5352118 Prediction of Structural Response of Reinforced Concrete Buildings Using Artificial Intelligence
Authors: Juan Bojórquez, Henry E. Reyes, Edén Bojórquez, Alfredo Reyes-Salazar
Abstract:
This paper addressed the use of Artificial Intelligence to obtain the structural reliability of reinforced concrete buildings. For this purpose, artificial neuronal networks (ANN) are developed to predict seismic demand hazard curves. In order to have enough input-output data to train the ANN, a set of reinforced concrete buildings (low, mid, and high rise) are designed, then a probabilistic seismic hazard analysis is made to obtain the seismic demand hazard curves. The results are then used as input-output data to train the ANN in a feedforward backpropagation model. The predicted values of the seismic demand hazard curves found by the ANN are then compared. Finally, it is concluded that the computer time analysis is significantly lower and the predictions obtained from the ANN were accurate in comparison to the values obtained from the conventional methods.Keywords: structural reliability, seismic design, machine learning, artificial neural network, probabilistic seismic hazard analysis, seismic demand hazard curves
Procedia PDF Downloads 1962117 Validating the Theme Park Service Quality Scale: A Case Study of Zhuhai Chimelong Ocean Kingdom
Authors: Kat Jingjing Luo
Abstract:
The development of theme parks in China has been through a rapid growth in the past decades. Increasing competition within service quality has forced theme park managers concerned the relationship between service quality and visitors’ satisfaction. Even though those existing service quality measurements such as SERVQUAL and THEMEQUAL have been applied in related researches, none of them is exclusive for Chinese theme park service quality. This study aims to investigate the service quality of the most popular theme park in China currently and develop a unique, reliable and valid scale. The reliability and validity analysis results from a survey of over 200 tourists in Chimelong ocean kingdom in Zhuhai city, south of China, indicate that the dimension of waiting time is a discover factor in the measurement of Chinese theme park service quality excluding in the THEMEQUAL instrument (i.e., tangibles, reliability, responsiveness and access, assurance, empathy and courtesy). The newly developed scale gives a better understand service quality in Chinese theme park industry, and the managerial implications in regard to the research, how to improve theme park service quality are discussed.Keywords: theme park, scale development, China, service quality
Procedia PDF Downloads 2792116 Subsea Control Module (SCM) - A Vital Factor for Well Integrity and Production Performance in Deep Water Oil and Gas Fields
Authors: Okoro Ikechukwu Ralph, Fuat Kara
Abstract:
The discoveries of hydrocarbon reserves has clearly drifted offshore, and in deeper waters - areas where the industry still has limited knowledge; and that were hitherto, regarded as being out of reach. This shift presents significant and increased challenges in technology requirements needed to guarantee safety of personnel, environment and equipment; ensure high reliability of installed equipment; and provide high level of confidence in security of investment and company reputation. Nowhere are these challenges more apparent than on subsea well integrity and production performance. The past two decades has witnessed enormous rise in deep and ultra-deep water offshore field developments for the recovery of hydrocarbons. Subsea installed equipment at the seabed has been the technology of choice for these developments. This paper discusses the role of Subsea Control module (SCM) as a vital factor for deep-water well integrity and production performance. A case study for Deep-water well integrity and production performance is analysed.Keywords: offshore reliability, production performance, subsea control module, well integrity
Procedia PDF Downloads 5112115 Comparative Isotherms Studies on Adsorptive Removal of Methyl Orange from Wastewater by Watermelon Rinds and Neem-Tree Leaves
Authors: Sadiq Sani, Muhammad B. Ibrahim
Abstract:
Watermelon rinds powder (WRP) and neem-tree leaves powder (NLP) were used as adsorbents for equilibrium adsorption isotherms studies for detoxification of methyl orange dye (MO) from simulated wastewater. The applicability of the process to various isotherm models was tested. All isotherms from the experimental data showed excellent linear reliability (R2: 0.9487-0.9992) but adsorptions onto WRP were more reliable (R2: 0.9724-0.9992) than onto NLP (R2: 0.9487-0.9989) except for Temkin’s Isotherm where reliability was better onto NLP (R2: 0.9937) than onto WRP (R2: 0.9935). Dubinin-Radushkevich’s monolayer adsorption capacities for both WRP and NLP (qD: 20.72 mg/g, 23.09 mg/g) were better than Langmuir’s (qm: 18.62 mg/g, 21.23 mg/g) with both capacities higher for adsorption onto NLP (qD: 23.09 mg/g; qm: 21.23 mg/g) than onto WRP (qD: 20.72 mg/g; qm: 18.62 mg/g). While values for Langmuir’s separation factor (RL) for both adsorbents suggested unfavourable adsorption processes (RL: -0.0461, -0.0250), Freundlich constant (nF) indicated favourable process onto both WRP (nF: 3.78) and NLP (nF: 5.47). Adsorption onto NLP had higher Dubinin-Radushkevich’s mean free energy of adsorption (E: 0.13 kJ/mol) than WRP (E: 0.08 kJ/mol) and Temkin’s heat of adsorption (bT) was better onto NLP (bT: -0.54 kJ/mol) than onto WRP (bT: -0.95 kJ/mol) all of which suggested physical adsorption.Keywords: adsorption isotherms, methyl orange, neem leaves, watermelon rinds
Procedia PDF Downloads 2732114 Validity and Reliability of Communication Activities of Daily Living- Second Edition and Assessment of Language-related Functional Activities: Comparative Evidence from Arab Aphasics
Authors: Sadeq Al Yaari, Ayman Al Yaari, Adham Al Yaari, Montaha Al Yaari, Aayah Al Yaari, Sajedah Al Yaari
Abstract:
Background: Validation of communication activities of daily living-second edition (CADL-2) and assessment of language-related functional activities (ALFA) tests is a critical investment decision, and activities related to language impairments often are underestimated. Literature indicates that age factors, and gender differences may affect the performance of the aphasics. Thus, understanding these influential factors is highly important to neuropsycholinguists and speech language pathologists (SLPs). Purpose: The goal of this study is twofold: (1) to in/validate CADL-2 and ALFA tests, and (2) to investigate whether or not the two assessment tests are reliable. Design: A comparative study is made between the results obtained from the analyses of the Arabic versions of CADL-2 and ALFA tests. Participants: The communication activities of daily-living and language-related functional activities were assessed from the obtained results of 100 adult aphasics (50 males, 50 females; ages 16 to 65). Procedures: Firstly, the two translated and standardized Arabic versions of CADL-2 and ALFA tests were introduced to the Arab aphasics under investigation. Armed with the new two versions of the tests, one of the researchers assessed the language-related functional communication and activities. Outcomes drawn from the obtained analysis of the comparative studies were then qualitatively and statistically analyzed. Main outcomes and Results: Regarding the validity of CADL-2 and ALFA, it is found that …. Is more valid in both pre-and posttests. Concerning the reliability of the two tests, it is found that ….is more reliable in both pre-and-posttests which undoubtedly means that …..is more trustable. Nor must we forget to indicate here that the relationship between age and gender was very weak due to that no remarkable gender differences between the two in both CADL-2 and ALFA pre-and-posttests. Conclusions & Implications: CADL-2 and ALFA tests were found to be valid and reliable tests. In contrast to previous studies, age and gender were not significantly associated with the results of validity and reliability of the two assessment tests. In clearer terms, age and gender patterns do not affect the validation of these two tests. Future studies might focus on complex questions including the use of CADL-2 and ALFA functionally; how gender and puberty influence the results in case the sample is large; the effects of each type of aphasia on the final outcomes, and measurements’ results of imaging techniques.Keywords: CADL-2, ALFA, comparison, language test, arab aphasics, validity, reliability, neuropsycholinguistics, comparison
Procedia PDF Downloads 362113 A Pilot Study on the Development and Validation of an Instrument to Evaluate Inpatient Beliefs, Expectations and Attitudes toward Reflexology (IBEAR)-16
Authors: Samuel Attias, Elad Schiff, Zahi Arnon, Eran Ben-Arye, Yael Keshet, Ibrahim Matter, Boker Lital Keinan
Abstract:
Background: Despite the extensive use of manual therapies, reflexology in particular, no validated tools have been developed to evaluate patients' beliefs, attitudes and expectations regarding reflexology. Such tools however are essential to improve the results of the reflexology treatment, by better adjusting it to the patients' attitudes and expectations. The tool also enables assessing correlations with clinical results of interventional studies using reflexology. Methods: The IBEAR (Inpatient Beliefs, Expectations and Attitudes toward Reflexology) tool contains 25 questions (8 demographic and 17 specifically addressing reflexology), and was constructed in several stages: brainstorming by a multidisciplinary team of experts; evaluation of each of the proposed questions by the experts' team; and assessment of the experts' degree of agreement per each question, based on a Likert 1-7 scale (1 – don't agree at all; 7 – agree completely). Cronbach's Alpha was computed to evaluate the questionnaire's reliability while the Factor analysis test was used for further validation (228 patients). The questionnaire was tested and re-tested (48h) on a group of 199 patients to assure clarity and reliability, using the Pearson coefficient and the Kappa test. It was modified based on these results into its final form. Results: After its construction, the IBEAR questionnaire passed the expert group's preliminary consensus, evaluation of the questions' clarity (from 5.1 to 7.0), inner validation (from 5.5 to 7) and structural validation (from 5.5 to 6.75). Factor analysis pointed to two content worlds in a division into 4 questions discussing attitudes and expectations versus 5 questions on belief and attitudes. Of the 221 questionnaires collected, a Cronbach's Alpha coefficient was calculated on nine questions relating to beliefs, expectations, and attitudes regarding reflexology. This measure stood at 0.716 (satisfactory reliability). At the Test-Retest stage, 199 research participants filled in the questionnaire a second time. The Pearson coefficient for all questions ranged between 0.73 and 0.94 (good to excellent reliability). As for dichotomic answers, Kappa scores ranged between 0.66 and 1.0 (mediocre to high). One of the questions was removed from the IBEAR following questionnaire validation. Conclusions: The present study provides evidence that the proposed IBEAR-16 questionnaire is a valid and reliable tool for the characterization of potential reflexology patients and may be effectively used in settings which include the evaluation of inpatients' beliefs, expectations, and attitudes toward reflexology.Keywords: reflexology, attitude, expectation, belief, CAM, inpatient
Procedia PDF Downloads 2282112 Radio Regulation Development and Radio Spectrum Analysis of Earth Station in Motion Service
Authors: Fei Peng, Jun Yuan, Chen Fan, Fan Jiang, Qian Sun, Yudi Liu
Abstract:
Although Earth Station in Motion (ESIM) services are widely used and there is a huge market demand around the world, International Telecommunication Union (ITU) does not have unified conclusion for the use of ESIM yet. ESIM are Mobile Satellite Services (MSS) due to its mobile-based attributes, while multiple administrations want to use ESIM in Fixed Satellite Service (FSS). However, Radio Regulations (RR) have strict distinction between MSS and FSS. In this case, ITU has been very controversial because this kind of application will violate the RR Article and the conflict will bring risks to the global deployment. Thus, this paper illustrates the development of rules, regulations, standards concerning ESIM and the radio spectrum usage of ESIM in different regions around the world. Firstly, the basic rules, standard and definition of ITU’s Radiocommunication Sector (ITU-R) is introduced. Secondly, the World Radiocommunication Conference (WRC) agenda item on radio spectrum allocation for ESIM, e.g. in C/Ku/Ka band, is introduced and multi-view on the radio spectrum allocation is elaborated, especially on 19.7-20.2 GHz & 29.5-30.0 GHz. Then, some ITU-R Recommendations and Reports are analyzed on the specific technique to enable these ESIM to communicate with Geostationary Earth Orbit Satellite (GSO) space stations in the FSS without causing interference at levels in excess of that caused by conventional FSS earth stations. Meanwhile, the opposite opinion on not allocating EISM service in FSS frequency band is also elaborated. Finally, based on the ESIM’s future application, the ITU-R standards development trend is forecasted. In conclusion, using radio spectrum resource in an equitable, rational and efficient manner is the basic guideline of ITU. Although it is not a good approach to obstruct the revise of RR when there is a large demand for radio spectrum resource in satellite industry, still the propulsion and global demand of the whole industry may face difficulties on the unclear application in modify rules of RR.Keywords: earth station in motion, ITU standards, radio regulations, radio spectrum, satellite communication
Procedia PDF Downloads 2882111 Quality of Service of Transportation Networks: A Hybrid Measurement of Travel Time and Reliability
Authors: Chin-Chia Jane
Abstract:
In a transportation network, travel time refers to the transmission time from source node to destination node, whereas reliability refers to the probability of a successful connection from source node to destination node. With an increasing emphasis on quality of service (QoS), both performance indexes are significant in the design and analysis of transportation systems. In this work, we extend the well-known flow network model for transportation networks so that travel time and reliability are integrated into the QoS measurement simultaneously. In the extended model, in addition to the general arc capacities, each intermediate node has a time weight which is the travel time for per unit of commodity going through the node. Meanwhile, arcs and nodes are treated as binary random variables that switch between operation and failure with associated probabilities. For pre-specified travel time limitation and demand requirement, the QoS of a transportation network is the probability that source can successfully transport the demand requirement to destination while the total transmission time is under the travel time limitation. This work is pioneering, since existing literatures that evaluate travel time reliability via a single optimization path, the proposed QoS focuses the performance of the whole network system. To compute the QoS of transportation networks, we first transfer the extended network model into an equivalent min-cost max-flow network model. In the transferred network, each arc has a new travel time weight which takes value 0. Each intermediate node is replaced by two nodes u and v, and an arc directed from u to v. The newly generated nodes u and v are perfect nodes. The new direct arc has three weights: travel time, capacity, and operation probability. Then the universal set of state vectors is recursively decomposed into disjoint subsets of reliable, unreliable, and stochastic vectors until no stochastic vector is left. The decomposition is made possible by applying existing efficient min-cost max-flow algorithm. Because the reliable subsets are disjoint, QoS can be obtained directly by summing the probabilities of these reliable subsets. Computational experiments are conducted on a benchmark network which has 11 nodes and 21 arcs. Five travel time limitations and five demand requirements are set to compute the QoS value. To make a comparison, we test the exhaustive complete enumeration method. Computational results reveal the proposed algorithm is much more efficient than the complete enumeration method. In this work, a transportation network is analyzed by an extended flow network model where each arc has a fixed capacity, each intermediate node has a time weight, and both arcs and nodes are independent binary random variables. The quality of service of the transportation network is an integration of customer demands, travel time, and the probability of connection. We present a decomposition algorithm to compute the QoS efficiently. Computational experiments conducted on a prototype network show that the proposed algorithm is superior to existing complete enumeration methods.Keywords: quality of service, reliability, transportation network, travel time
Procedia PDF Downloads 2212110 Improvement of the Reliability and the Availability of a Production System
Authors: Lakhoua Najeh
Abstract:
Aims of the work: The aim of this paper is to improve the reliability and the availability of a Packer production line of cigarettes based on two methods: The SADT method (Structured Analysis Design Technique) and the FMECA approach (Failure Mode Effects and Critically Analysis). The first method enables us to describe the functionality of the Packer production line of cigarettes and the second method enables us to establish an FMECA analysis. Methods: The methodology adopted in order to contribute to the improvement of the reliability and the availability of a Packer production line of cigarettes has been proposed in this paper, and it is based on the use of Structured Analysis Design Technique (SADT) and Failure mode, effects, and criticality analysis (FMECA) methods. This methodology consists of using a diagnosis of the existing of all of the equipment of a production line of a factory in order to determine the most critical machine. In fact, we use, on the one hand, a functional analysis based on the SADT method of the production line and on the other hand, a diagnosis and classification of mechanical and electrical failures of the line production by their criticality analysis based on the FMECA approach. Results: Based on the methodology adopted in this paper, the results are the creation and the launch of a preventive maintenance plan. They contain the different elements of a Packer production line of cigarettes; the list of the intervention preventive activities and their period of realization. Conclusion: The diagnosis of the existing state helped us to found that the machine of cigarettes used in the Packer production line of cigarettes is the most critical machine in the factory. Then this enables us in the one hand, to describe the functionality of the production line of cigarettes by SADT method and on the other hand, to study the FMECA machine in order to improve the availability and the performance of this machine.Keywords: production system, diagnosis, SADT method, FMECA method
Procedia PDF Downloads 142