Search results for: Regression method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8626

Search results for: Regression method

7036 Turbine Compressor Vibration Analysis and Rotor Movement Evaluation by Shaft Center Line Method (The Case History Related to Main Turbine Compressor of an Olefin Plant in Iran Oil Industries)

Authors: Omid A. Zargar

Abstract:

Vibration monitoring methods of most critical equipment like main turbine and compressors always plays important role in preventive maintenance and management consideration in big industrial plants. There are a number of traditional methods like monitoring the overall vibration data from Bently Nevada panel and the time wave form (TWF) or fast Fourier transform (FFT) monitoring. Besides, Shaft centerline monitoring method developed too much in recent years. There are a number of arguments both in favor of and against this method between people who work in preventive maintenance and condition monitoring systems (vibration analysts). In this paper basic principal of Turbine compressor vibration analysis and rotor movement evaluation by shaft centerline method discussed in details through a case history. This case history is related to main turbine compressor of an olefin plant in Iran oil industry. In addition, some common mistakes that may occur by vibration analyst during the process discussed in details. It is worthy to know that, these mistakes may one of the reasons that sometimes this method seems to be not effective. Furthermore, recent patent and innovation in shaft position and movement evaluation are discussed in this paper.

Keywords: Shaft centerline position, attitude angle, journal bearing, sleeve bearing, tilting pad, steam turbine, main compressor, multistage compressor, condition monitoring, non-contact probe

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7098
7035 Human Action Recognition Using Variational Bayesian HMM with Dirichlet Process Mixture of Gaussian Wishart Emission Model

Authors: Wanhyun Cho, Soonja Kang, Sangkyoon Kim, Soonyoung Park

Abstract:

In this paper, we present the human action recognition method using the variational Bayesian HMM with the Dirichlet process mixture (DPM) of the Gaussian-Wishart emission model (GWEM). First, we define the Bayesian HMM based on the Dirichlet process, which allows an infinite number of Gaussian-Wishart components to support continuous emission observations. Second, we have considered an efficient variational Bayesian inference method that can be applied to drive the posterior distribution of hidden variables and model parameters for the proposed model based on training data. And then we have derived the predictive distribution that may be used to classify new action. Third, the paper proposes a process of extracting appropriate spatial-temporal feature vectors that can be used to recognize a wide range of human behaviors from input video image. Finally, we have conducted experiments that can evaluate the performance of the proposed method. The experimental results show that the method presented is more efficient with human action recognition than existing methods.

Keywords: Human action recognition, Bayesian HMM, Dirichlet process mixture model, Gaussian-Wishart emission model, Variational Bayesian inference, Prior distribution and approximate posterior distribution, KTH dataset.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 977
7034 Automation of Fishhooks Objective Measures

Authors: S. Chabrier, G. Molle, E. Conte, C. Carlier

Abstract:

Fishing has always been an essential component of the Polynesians- life. Fishhooks, mostly in pearl shell, found during archaeological excavations are the artifacts related to this activity the most numerous. Thanks to them, we try to reconstruct the ancient techniques of resources exploitation, inside the lagoons and offshore. They can also be used as chronological and cultural indicators. The shapes and dimensions of these artifacts allow comparisons and classifications used in both functional approach and chrono-cultural perspective. Hence it is very important for the ethno-archaeologists to dispose of reliable methods and standardized measurement of these artifacts. Such a reliable objective and standardized method have been previously proposed. But this method cannot be envisaged manually because of the very important time required to measure each fishhook manually and the quantity of fishhooks to measure (many hundreds). We propose in this paper a detailed acquisition protocol of fishhooks and an automation of every step of this method. We also provide some experimental results obtained on the fishhooks coming from three archaeological excavations sites.

Keywords: Automated measures, extraction, fishhook, segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1580
7033 Application Reliability Method for Concrete Dams

Authors: Mustapha Kamel Mihoubi, Mohamed Essadik Kerkar

Abstract:

Probabilistic risk analysis models are used to provide a better understanding of the reliability and structural failure of works, including when calculating the stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of reliability analysis methods including the methods used in engineering. It is in our case, the use of level 2 methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type first order risk method (FORM) and the second order risk method (SORM). By way of comparison, a level three method was used which generates a full analysis of the problem and involves an integration of the probability density function of random variables extended to the field of security using the Monte Carlo simulation method. Taking into account the change in stress following load combinations: normal, exceptional and extreme acting on the dam, calculation of the results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities, thus causing a significant decrease in strength, shear forces then induce a shift that threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case the increase of uplift in a hypothetical default of the drainage system.

Keywords: Dam, failure, limit-state, Monte Carlo simulation, reliability, probability, simulation, sliding, Taylor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1204
7032 Robust Nonlinear Control of Two Links Robot Manipulator and Computing Maximum Load

Authors: Hasanifard Goran, Habib Nejad Korayem Moharam, Nikoobin Amin

Abstract:

A new robust nonlinear control scheme of a manipulator is proposed in this paper which is robust against modeling errors and unknown disturbances. It is based on the principle of variable structure control, with sliding mode control (SMC) method. The variable structure control method is a robust method that appears to be well suited for robotic manipulators because it requers only bounds on the robotic arm parameters. But there is no single systematic procedure that is guaranteed to produce a suitable control law. Also, to reduce chattring of the control signal, we replaced the sgn function in the control law by a continuous approximation such as tangant function. We can compute the maximum load with regard to applied torque into joints. The effectivness of the proposed approach has been evaluated analitically demonstrated through computer simulations for the cases of variable load and robot arm parameters.

Keywords: Variable structure control, robust control, switching surface, robot manipulator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1697
7031 Enhancement of Low Contrast Satellite Images using Discrete Cosine Transform and Singular Value Decomposition

Authors: A. K. Bhandari, A. Kumar, P. K. Padhy

Abstract:

In this paper, a novel contrast enhancement technique for contrast enhancement of a low-contrast satellite image has been proposed based on the singular value decomposition (SVD) and discrete cosine transform (DCT). The singular value matrix represents the intensity information of the given image and any change on the singular values change the intensity of the input image. The proposed technique converts the image into the SVD-DCT domain and after normalizing the singular value matrix; the enhanced image is reconstructed by using inverse DCT. The visual and quantitative results suggest that the proposed SVD-DCT method clearly shows the increased efficiency and flexibility of the proposed method over the exiting methods such as Linear Contrast Stretching technique, GHE technique, DWT-SVD technique, DWT technique, Decorrelation Stretching technique, Gamma Correction method based techniques.

Keywords: Singular Value Decomposition (SVD), discretecosine transforms (DCT), image equalization and satellite imagecontrast enhancement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3817
7030 A Dirty Page Migration Method in Process of Memory Migration Based on Pre-copy Technology

Authors: Kang Zijian, Zhang Tingyu, Burra Venkata Durga Kumar

Abstract:

This article investigates the challenges in memory migration during the live migration of virtual machines. We found three challenges probably existing in pre-copy technology. One of the main challenges is the challenge of downtime migration. Decreasing the downtime could promise the normal work for a virtual machine. Although pre-copy technology is greatly decreasing the downtime, we still need to shut down the machine in order to finish the last round of data transfer. This paper provides an optimization scheme for the problems existing in pro-copy technology, mainly the optimization of the dirty page migration mechanism. The typical pre-copy technology copies n-1th’s dirty pages in nth turn. However, our idea is to create a double iteration method to solve this problem.

Keywords: Virtual machine, pre-copy technology, memory migration process, downtime, dirty pages migration method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 106
7029 A Conservative Multi-block Algorithm for Two-dimensional Numerical Model

Authors: Yaoxin Zhang, Yafei Jia, Sam S.Y. Wang

Abstract:

A multi-block algorithm and its implementation in two-dimensional finite element numerical model CCHE2D are presented. In addition to a conventional Lagrangian Interpolation Method (LIM), a novel interpolation method, called Consistent Interpolation Method (CIM), is proposed for more accurate information transfer across the interfaces. The consistent interpolation solves the governing equations over the auxiliary elements constructed around the interpolation nodes using the same numerical scheme used for the internal computational nodes. With the CIM, the momentum conservation can be maintained as well as the mass conservation. An imbalance correction scheme is used to enforce the conservation laws (mass and momentum) across the interfaces. Comparisons of the LIM and the CIM are made using several flow simulation examples. It is shown that the proposed CIM is physically more accurate and produces satisfactory results efficiently.

Keywords: Multi-block algorithm, conservation, interpolation, numerical model, flow simulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1773
7028 The Application of Adaptive Tabu Search Algorithm and Averaging Model to the Optimal Controller Design of Buck Converters

Authors: T. Sopapirm, K-N. Areerak, K-L. Areerak, A. Srikaew

Abstract:

The paper presents the applications of artificial intelligence technique called adaptive tabu search to design the controller of a buck converter. The averaging model derived from the DQ and generalized state-space averaging methods is applied to simulate the system during a searching process. The simulations using such averaging model require the faster computational time compared with that of the full topology model from the software packages. The reported model is suitable for the work in the paper in which the repeating calculation is needed for searching the best solution. The results will show that the proposed design technique can provide the better output waveforms compared with those designed from the classical method.

Keywords: Buck converter, adaptive tabu search, DQ method, generalized state-space averaging method, modeling and simulation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1812
7027 Clustering based Voltage Control Areas for Localized Reactive Power Management in Deregulated Power System

Authors: Saran Satsangi, Ashish Saini, Amit Saraswat

Abstract:

In this paper, a new K-means clustering based approach for identification of voltage control areas is developed. Voltage control areas are important for efficient reactive power management in power systems operating under deregulated environment. Although, voltage control areas are formed using conventional hierarchical clustering based method, but the present paper investigate the capability of K-means clustering for the purpose of forming voltage control areas. The proposed method is tested and compared for IEEE 14 bus and IEEE 30 bus systems. The results show that this K-means based method is competing with conventional hierarchical approach

Keywords: Voltage control areas, reactive power management, K-means clustering algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2374
7026 An Evaluation of Requirements Management and Traceability Tools

Authors: Muhammad Shahid, Suhaimi Ibrahim, Mohd Naz'ri Mahrin

Abstract:

Requirements management is critical to software delivery success and project lifecycle. Requirements management and their traceability provide assistance for many software engineering activities like impact analysis, coverage analysis, requirements validation and regression testing. In addition requirements traceability is the recognized component of many software process improvement initiatives. Requirements traceability also helps to control and manage evolution of a software system. This paper aims to provide an evaluation of current requirements management and traceability tools. Management and test managers require an appropriate tool for the software under test. We hope, evaluation identified here will help to select the efficient and effective tool.

Keywords: Requirements Traceability, Requirements TraceabilityTools; Requirements Management, Requirement Engineering

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3996
7025 Comparative Study of Complexity in Streetscape Composition

Authors: Ahmed Mansouri, Naoji Matsumoto

Abstract:

This research is a comparative study of complexity, as a multidimensional concept, in the context of streetscape composition in Algeria and Japan. 80 streetscapes visual arrays have been collected and then presented to 20 participants, with different cultural backgrounds, in order to be categorized and classified according to their degrees of complexity. Three analysis methods have been used in this research: cluster analysis, ranking method and Hayashi Quantification method (Method III). The results showed that complexity, disorder, irregularity and disorganization are often conflicting concepts in the urban context. Algerian daytime streetscapes seem to be balanced, ordered and regular, and Japanese daytime streetscapes seem to be unbalanced, regular and vivid. Variety, richness and irregularity with some aspects of order and organization seem to characterize Algerian night streetscapes. Japanese night streetscapes seem to be more related to balance, regularity, order and organization with some aspects of confusion and ambiguity. Complexity characterized mainly Algerian avenues with green infrastructure. Therefore, for Japanese participants, Japanese traditional night streetscapes were complex. And for foreigners, Algerian and Japanese avenues nightscapes were the most complex visual arrays.

Keywords: Streetscape, Nightscape, Complexity, Visual Array, Affordance, Cluster Analysis, Hayashi Quantification Method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2317
7024 Sensor Optimisation via H∞ Applied to a MAGLEV Suspension System

Authors: Konstantinos Michail, Argyrios Zolotas, Roger Goodall, John Pearson

Abstract:

In this paper a systematic method via H∞ control design is proposed to select a sensor set that satisfies a number of input criteria for a MAGLEV suspension system. The proposed method recovers a number of optimised controllers for each possible sensor set that satisfies the performance and constraint criteria using evolutionary algorithms.

Keywords: H-infinity, Sensor optimisation, Genetic algorithms, MAGLEV vehicles

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1459
7023 Subarray Based Multiuser Massive MIMO Design Adopting Large Transmit and Receive Arrays

Authors: Tetsuki Taniguchi, Yoshio Karasawa

Abstract:

This paper describes a subarray based low computational design method of multiuser massive multiple input multiple output (MIMO) system. In our previous works, use of large array is assumed only in transmitter, but this study considers the case both of transmitter and receiver sides are equipped with large array antennas. For this aim, receive arrays are also divided into several subarrays, and the former proposed method is modified for the synthesis of a large array from subarrays in both ends. Through computer simulations, it is verified that the performance of the proposed method is degraded compared with the original approach, but it can achieve the improvement in the aspect of complexity, namely, significant reduction of the computational load to the practical level.

Keywords: Massive multiple input multiple output (MIMO), multiuser, large array, subarray, zero forcing, singular value decomposition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959
7022 Dynamic Network Routing Method Based on Chromosome Learning

Authors: Xun Liang

Abstract:

In this paper, we probe into the traffic assignment problem by the chromosome-learning-based path finding method in simulation, which is to model the driver' behavior in the with-in-a-day process. By simply making a combination and a change of the traffic route chromosomes, the driver at the intersection chooses his next route. The various crossover and mutation rules are proposed with extensive examples.

Keywords: Chromosome learning, crossover, mutation, traffic path finding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1319
7021 An Optimal Control of Water Pollution in a Stream Using a Finite Difference Method

Authors: Nopparat Pochai, Rujira Deepana

Abstract:

Water pollution assessment problems arise frequently in environmental science. In this research, a finite difference method for solving the one-dimensional steady convection-diffusion equation with variable coefficients is proposed; it is then used to optimize water treatment costs.

Keywords: Finite difference, One-dimensional, Steady state, Waterpollution control, Optimization, Convection-diffusion equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1728
7020 Factorial Design Analysis for Quality of Video on MANET

Authors: Hyoup-Sang Yoon

Abstract:

The quality of video transmitted by mobile ad hoc networks (MANETs) can be influenced by several factors, including protocol layers; parameter settings of each protocol. In this paper, we are concerned with understanding the functional relationship between these influential factors and objective video quality in MANETs. We illustrate a systematic statistical design of experiments (DOE) strategy can be used to analyze MANET parameters and performance. Using a 2k factorial design, we quantify the main and interactive effects of 7 factors on a response metric (i.e., mean opinion score (MOS) calculated by PSNR with Evalvid package) we then develop a first-order linear regression model between the influential factors and the performance metric.

Keywords: Evalvid, full factorial design, mobile ad hoc networks, ns-2.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2071
7019 A Novel Approach for Scheduling Rescue Robot Mission Using Decision Analysis

Authors: Rana Soltani-Zarrin, Sohrab Khanmohammadi

Abstract:

In this paper, a new method for multi criteria decision making is represented whichspecifies a trajectory satisfying desired criteria including minimization of time. A rescue robot is defined to perform certain tasks before the arrival of rescue team, including evaluation of the probability of explosion in the area, detecting human-beings, and providing preliminary aidsin case of identifying signs of life, so that the security of the surroundings will have enhanced significantly for the individuals inside the disaster zone as well as the rescue team. The main idea behind our technique is using the Program Evaluation and Review Technique analysis along with Critical Path Method and use the Multi Criteria Decision Making (MCDM) method to decidewhich set of activities must be performed first. Since the disastrous event in one area may be well contagious to others, it is one of the robot's priorities to evaluate the relative adversity of the situation, using the above methods and prioritize its mission.

Keywords: PERT, CPM, MCDM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1643
7018 An Online Mastery Learning Method Based On a Dynamic Formative Evaluation

Authors: Jeongim Kang, Moon Hee Kim, Seong Baeg Kim

Abstract:

This paper proposes a novel e-learning model that is  based on a dynamic formative evaluation. On evaluating the existing  format of e-learning, conditions regarding repetitive learning to  achieve mastery, causes issues for learners to lose tension and become  neglectful of learning. The dynamic formative evaluation proposed is  able to supplement limitation of the existing approaches. Since a  repetitive learning method does not provide a perfect feedback, this  paper puts an emphasis on the dynamic formative evaluation that is  able to maximize learning achievement. Through the dynamic  formative evaluation, the instructor is able to refer to the evaluation  result when making an estimation about the learner. To show the flow  chart of learning, based on the dynamic formative evaluation, the  model proves its effectiveness and validity.

 

Keywords: Online learning, dynamic formative evaluation, mastery learning, repetitive learning method, learning achievement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1765
7017 The System for Root Canal Length Measurement Based on Multifrequency Impedance Method

Authors: Zheng Zhang, Xin Chen, Guoqing Ding

Abstract:

Electronic apex locators (EAL) has been widely used clinically for measuring root canal working length with high accuracy, which is crucial for successful endodontic treatment. In order to maintain high accuracy in different measurement environments, this study presented a system for root canal length measurement based on multifrequency impedance method. This measuring system can generate a sweep current with frequencies from 100 Hz to 1 MHz through a direct digital synthesizer. Multiple impedance ratios with different combinations of frequencies were obtained and transmitted by an analog-to-digital converter and several of them with representatives will be selected after data process. The system analyzed the functional relationship between these impedance ratios and the distance between the file and the apex with statistics by measuring plenty of teeth. The position of the apical foramen can be determined by the statistical model using these impedance ratios. The experimental results revealed that the accuracy of the system based on multifrequency impedance ratios method to determine the position of the apical foramen was higher than the dual-frequency impedance ratio method. Besides that, for more complex measurement environments, the performance of the system was more stable.

Keywords: Root canal length, apex locator, multifrequency impedance, sweep frequency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 717
7016 Structural Optimization Method for 3D Reinforced Concrete Building Structure with Shear Wall

Authors: H. Nikzad, S. Yoshitomi

Abstract:

In this paper, an optimization procedure is applied for 3D Reinforced concrete building structure with shear wall.  In the optimization problem, cross sections of beams, columns and shear wall dimensions are considered as design variables and the optimal cross sections can be derived to minimize the total cost of the structure. As for final design application, the most suitable sections are selected to satisfy ACI 318-14 code provision based on static linear analysis. The validity of the method is examined through numerical example of 15 storied 3D RC building with shear wall.  This optimization method is expected to assist in providing a useful reference in design early stage, and to be an effective and powerful tool for structural design of RC shear wall structures.

Keywords: Structural optimization, linear static analysis, ETABS, MATLAB, RC moment frame, RC shear wall structures.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1314
7015 Robot Vision Application based on Complex 3D Pose Computation

Authors: F. Rotaru, S. Bejinariu, C. D. Niţâ, R. Luca, I. Pâvâloi, C. Lazâr

Abstract:

The paper presents a technique suitable in robot vision applications where it is not possible to establish the object position from one view. Usually, one view pose calculation methods are based on the correspondence of image features established at a training step and exactly the same image features extracted at the execution step, for a different object pose. When such a correspondence is not feasible because of the lack of specific features a new method is proposed. In the first step the method computes from two views the 3D pose of feature points. Subsequently, using a registration algorithm, the set of 3D feature points extracted at the execution phase is aligned with the set of 3D feature points extracted at the training phase. The result is a Euclidean transform which have to be used by robot head for reorientation at execution step.

Keywords: features correspondence, registration algorithm, robot vision, triangulation method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1439
7014 Perceptual Framework for a Modern Left-Turn Collision Warning System

Authors: E. Dabbour, S. M. Easa

Abstract:

Most of the collision warning systems currently available in the automotive market are mainly designed to warn against imminent rear-end and lane-changing collisions. No collision warning system is commercially available to warn against imminent turning collisions at intersections, especially for left-turn collisions when a driver attempts to make a left-turn at either a signalized or non-signalized intersection, conflicting with the path of other approaching vehicles traveling on the opposite-direction traffic stream. One of the major factors that lead to left-turn collisions is the human error and misjudgment of the driver of the turning vehicle when perceiving the speed and acceleration of other vehicles traveling on the opposite-direction traffic stream; therefore, using a properly-designed collision warning system will likely reduce, or even eliminate, this type of collisions by reducing human error. This paper introduces perceptual framework for a proposed collision warning system that can detect imminent left-turn collisions at intersections. The system utilizes a commercially-available detection sensor (either a radar sensor or a laser detector) to detect approaching vehicles traveling on the opposite-direction traffic stream and calculate their speeds and acceleration rates to estimate the time-tocollision and compare that time to the time required for the turning vehicle to clear the intersection. When calculating the time required for the turning vehicle to clear the intersection, consideration is given to the perception-reaction time of the driver of the turning vehicle, which is the time required by the driver to perceive the message given by the warning system and react to it by engaging the throttle. A regression model was developed to estimate perception-reaction time based on age and gender of the driver of the host vehicle. Desired acceleration rate selected by the driver of the turning vehicle, when making the left-turn movement, is another human factor that is considered by the system. Another regression model was developed to estimate the acceleration rate selected by the driver of the turning vehicle based on driver-s age and gender as well as on the location and speed of the nearest approaching vehicle along with the maximum acceleration rate provided by the mechanical characteristics of the turning vehicle. By comparing time-to-collision with the time required for the turning vehicle to clear the intersection, the system displays a message to the driver of the turning vehicle when departure is safe. An application example is provided to illustrate the logic algorithm of the proposed system.

Keywords: Collision warning systems, intelligent transportationsystems, vehicle safety.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2030
7013 Total Quality Management: The Socio- Demographic and Operational-Financial Determinants for Users- Perception of the Services Quality

Authors: H. Silvestre

Abstract:

The aim of this paper is to know the sociodemographic and operational-financial determinants of the services quality perceived by users of the national health services. Through the use of an inquiry conducted by the Ministry of Health, comprehending 16.936 interviews in 2006, we intend to find out if there is any characteristic that determines the 2006 inquiry results. With the revision of the literature we also want to know if the operational-financial results have implications in hospitals users- perception on the quality of the received services. In order to achieve our main goals we will make use of the regression analysis to find out the possible dimensions that determine those results.

Keywords: Management by Results, Quality Approach, Tableau de Bord, Total Quality Management, Services quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1389
7012 Risk Level Evaluation for Power System Facilities in Smart Grid

Authors: Sung-Hun Lee, Yun-Seong Lee, Jin-O Kim

Abstract:

Reliability Centered Maintenance(RCM) is one of most widely used methods in the modern power system to schedule a maintenance cycle and determine the priority of inspection. In order to apply the RCM method to the Smart Grid, a precedence study for the new structure of rearranged system should be performed due to introduction of additional installation such as renewable and sustainable energy resources, energy storage devices and advanced metering infrastructure. This paper proposes a new method to evaluate the priority of maintenance and inspection of the power system facilities in the Smart Grid using the Risk Priority Number. In order to calculate that risk index, it is required that the reliability block diagram should be analyzed for the Smart Grid system. Finally, the feasible technical method is discussed to estimate the risk potential as part of the RCM procedure.

Keywords: Expert System, FMECA, Fuzzy Theory, Reliability Centered Maintenance, Risk Priority Number

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1769
7011 A New Method for Estimating the Mass Recession Rate for Ablator Systems

Authors: Bianca A. Szasz, Keiichi Okuyama

Abstract:

As the human race will continue to explore the space by creating new space transportation means and sending them to other planets, the enhance of atmospheric reentry study is crucial. In this context, an analysis of mass recession rate of ablative materials for thermal shields of reentry spacecrafts is important to be carried out. The paper describes a new estimation method for calculating the mass recession of an ablator system made of carbon fiber reinforced plastic materials. This method is based on Arrhenius equation for low temperatures and, for high temperatures, on a theory applied for the recession phenomenon of carbon fiber reinforced plastic materials, theory which takes into account the presence of the resin inside the materials. The space mission of USERS spacecraft is considered as a case study.

Keywords: Ablator system, mass recession, spacecraft, atmospheric reentry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2095
7010 Numerical Approximation to the Performance of CUSUM Charts for EMA (1) Process

Authors: K. Petcharat, Y. Areepong, S. Sukparungsri, G. Mititelu

Abstract:

These paper, we approximate the average run length (ARL) for CUSUM chart when observation are an exponential first order moving average sequence (EMA1). We used Gauss-Legendre numerical scheme for integral equations (IE) method for approximate ARL0 and ARL1, where ARL in control and out of control, respectively. We compared the results from IE method and exact solution such that the two methods perform good agreement.

Keywords: Cumulative Sum Chart, Moving Average Observation, Average Run Length, Numerical Approximations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2140
7009 Extended Intuitionistic Fuzzy VIKOR Method in Group Decision Making: The Case of Vendor Selection Decision

Authors: Nastaran Hajiheydari, Mohammad Soltani Delgosha

Abstract:

Vendor (supplier) selection is a group decision-making (GDM) process, in which, based on some predetermined criteria, the experts’ preferences are provided in order to rank and choose the most desirable suppliers. In the real business environment, our attitudes or our choices would be made in an uncertain and indecisive situation could not be expressed in a crisp framework. Intuitionistic fuzzy sets (IFSs) could handle such situations in the best way. VIKOR method was developed to solve multi-criteria decision-making (MCDM) problems. This method, which is used to determine the compromised feasible solution with respect to the conflicting criteria, introduces a multi-criteria ranking index based on the particular measure of 'closeness' to the 'ideal solution'. Until now, there has been a little investigation of VIKOR with IFS, therefore we extended the intuitionistic fuzzy (IF) VIKOR to solve vendor selection problem under IF GDM environment. The present study intends to develop an IF VIKOR method in a GDM situation. Therefore, a model is presented to calculate the criterion weights based on entropy measure. Then, the interval-valued intuitionistic fuzzy weighted geometric (IFWG) operator utilized to obtain the total decision matrix. In the next stage, an approach based on the positive idle intuitionistic fuzzy number (PIIFN) and negative idle intuitionistic fuzzy number (NIIFN) was developed. Finally, the application of the proposed method to solve a vendor selection problem illustrated.

Keywords: Group decision making, intuitionistic fuzzy entropy measure, intuitionistic fuzzy set, vendor selection VIKOR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 707
7008 Influence of Optical Fluence Distribution on Photoacoustic Imaging

Authors: Mohamed K. Metwally, Sherif H. El-Gohary, Kyung Min Byun, Seung Moo Han, Soo Yeol Lee, Min Hyoung Cho, Gon Khang, Jinsung Cho, Tae-Seong Kim

Abstract:

Photoacoustic imaging (PAI) is a non-invasive and non-ionizing imaging modality that combines the absorption contrast of light with ultrasound resolution. Laser is used to deposit optical energy into a target (i.e., optical fluence). Consequently, the target temperature rises, and then thermal expansion occurs that leads to generating a PA signal. In general, most image reconstruction algorithms for PAI assume uniform fluence within an imaging object. However, it is known that optical fluence distribution within the object is non-uniform. This could affect the reconstruction of PA images. In this study, we have investigated the influence of optical fluence distribution on PA back-propagation imaging using finite element method. The uniform fluence was simulated as a triangular waveform within the object of interest. The non-uniform fluence distribution was estimated by solving light propagation within a tissue model via Monte Carlo method. The results show that the PA signal in the case of non-uniform fluence is wider than the uniform case by 23%. The frequency spectrum of the PA signal due to the non-uniform fluence has missed some high frequency components in comparison to the uniform case. Consequently, the reconstructed image with the non-uniform fluence exhibits a strong smoothing effect.

Keywords: Finite Element Method, Fluence Distribution, Monte Carlo Method, Photoacoustic Imaging.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2649
7007 An Earth Mover’s Distance Algorithm Based DDoS Detection Mechanism in SDN

Authors: Yang Zhou, Kangfeng Zheng, Wei Ni, Ren Ping Liu

Abstract:

Software-defined networking (SDN) provides a solution for scalable network framework with decoupled control and data plane. However, this architecture also induces a particular distributed denial-of-service (DDoS) attack that can affect or even overwhelm the SDN network. DDoS attack detection problem has to date been mostly researched as entropy comparison problem. However, this problem lacks the utilization of SDN, and the results are not accurate. In this paper, we propose a DDoS attack detection method, which interprets DDoS detection as a signature matching problem and is formulated as Earth Mover’s Distance (EMD) model. Considering the feasibility and accuracy, we further propose to define the cost function of EMD to be a generalized Kullback-Leibler divergence. Simulation results show that our proposed method can detect DDoS attacks by comparing EMD values with the ones computed in the case without attacks. Moreover, our method can significantly increase the true positive rate of detection.

Keywords: DDoS detection, EMD, relative entropy, SDN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 735