Search results for: mixed method analysis
12716 Symbolic Model Checking of Interactions in Sequence Diagrams with Combined Fragments by SMV
Authors: Yuka Kawakami, Tomoyuki Yokogawa, Hisashi Miyazaki, Sousuke Amasaki, Yoichiro Sato, Michiyoshi Hayase
Abstract:
In this paper, we proposed a method for detecting consistency violation between state machine diagrams and a sequence diagram defined in UML 2.0 using SMV. We extended a method expressing these diagrams defined in UML 1.0 with boolean formulas so that it can express a sequence diagram with combined fragments introduced in UML 2.0. This extension made it possible to represent three types of combined fragment: alternative, option and parallel. As a result of experiment, we confirmed that the proposed method could detect consistency violation correctly with SMV.
Keywords: UML, model checking, SMV, sequence diagram.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 146912715 T-DOF PID Controller Design using Characteristic Ratio Assignment Method for Quadruple Tank Process
Authors: Tianchai Suksri, U-thai Sritheeravirojana, Arjin Numsomran, Viriya Kongrattana, Thongchai Werataweemart
Abstract:
A control system design with Characteristic Ratio Assignment (CRA) is proven that effective for SISO control design. But the control system design for MIMO via CRA is not concrete procedure. In this paper presents the control system design method for quadruple-tank process via CRA. By using the decentralized method for both minimum phase and non-minimum phase are made. The results from PI and PID controller design via CRA can be illustrated the validity of our approach by MATLAB.Keywords: CRA, Quadruple-Tank.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 158812714 Sound Absorption of Arenga Pinnata Natural Fiber
Authors: Lindawati Ismail, Mohd. Imran Ghazali, Shahruddin Mahzan, Ahmad Mujahid Ahmad Zaidi
Abstract:
Arenga pinnata is an abundantly natural fiber that can be used for sound proof material. However, the scientific data of acoustics properties of Arenga pinnata was not available yet. In this study the sound absorption of pure arenga pinnata was measured. The thickness of Arenga pinnata was varied in 10 mm, 20 mm, 30mm, and 40mm. This work was carried out to investigate the potential of using Arenga pinnata fiber as raw material for sound absorbing material. Impedance Tube Method was used to measure sound absorption coefficient (α). The Measurements was done in accordance with ASTM E1050-98, that is the standard test method for impedance and absorption of acoustical materials using a tube, two microphones and a digital frequency analysis system . The results showed that sound absorption coefficients of Arenga pinnata were good from 2000 Hz to 5000 Hz within the range of 0.75 – 0.90. The optimum sound absorption coefficient was obtained from the thickness of 40 mm. These results indicated that Arenga pinnata fiber is promising to be used as raw material of sound absorbing material with low cost, light, and biodegradable.
Keywords: Sound absorption, Arenga pinnata, raw material.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 361912713 Characteristics of Wall Thickness Increase in Pipe Reduction Process using Planetary Rolls
Authors: Yuji Kotani, Shunsuke Kanai, Hisaki Watari
Abstract:
In recent years, global warming has become a worldwide problem. The reduction of carbon dioxide emissions is a top priority for many companies in the manufacturing industry. In the automobile industry as well, the reduction of carbon dioxide emissions is one of the most important issues. Technology to reduce the weight of automotive parts improves the fuel economy of automobiles, and is an important technology for reducing carbon dioxide. Also, even if this weight reduction technology is applied to electric automobiles rather than gasoline automobiles, reducing energy consumption remains an important issue. Plastic processing of hollow pipes is one important technology for realizing the weight reduction of automotive parts. Ohashi et al. [1],[2] present an example of research on pipe formation in which a process was carried out to enlarge a pipe diameter using a lost core, achieving the suppression of wall thickness reduction and greater pipe expansion than hydroforming. In this study, we investigated a method to increase the wall thickness of a pipe through pipe compression using planetary rolls. The establishment of a technology whereby the wall thickness of a pipe can be controlled without buckling the pipe is an important technology for the weight reduction of products. Using the finite element analysis method, we predicted that it would be possible to increase the compression of an aluminum pipe with a 3mm wall thickness by approximately 20%, and wall thickness by approximately 20% by pressing the hollow pipe with planetary rolls.Keywords: Pipe-Forming, Wall Thickness, Finite-element-method
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 298512712 Laser Doppler Flowmetry in Diagnostics of Vascular Lesions in Lower Extremities
Authors: Petr V. Vasilev, Eduard V. Volkov, Alexej N. Godok, Alexej A, Grischschuk, Vitalij A. Rybalchenko
Abstract:
Laser Doppler flowmetry is a modern method of noninvasive microcirculation investigation. The aim of our study was to use this method in the examination of patients with secondary lymphedema of the lower extremities and obliterating atherosclerosis of lower extremities. In the analysis of the amplitude-frequency spectrum of secondary lymphedema patients we have identified remarkable changes. To describe the changes we used a special amplitude rate. In both of patients groups this rate was significally (p<0.05) different with the control group. So the marker phenomena of the amplitude-frequency spectrum of the LDF signal were identified. It is suggested that there is a limfodynamics contribution to the formation of the output signal of laser Doppler flowmetry. These data have fundamental meaning and are interesting for practical medicine, as they give an opportunity to further developments for the use of laser Doppler flowmetry in the diagnostics and monitoring the effectiveness of the treatment.Keywords: laser Doppler flowmetry, secondary lymphedema of the lower extremities, obliterating atherosclerosis, non-invasive diagnostics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 173212711 GPU Implementation for Solving in Compressible Two-Phase Flows
Authors: Sheng-Hsiu Kuo, Pao-Hsiung Chiu, Reui-Kuo Lin, Yan-Ting Lin
Abstract:
A one-step conservative level set method, combined with a global mass correction method, is developed in this study to simulate the incompressible two-phase flows. The present framework do not need to solve the conservative level set scheme at two separated steps, and the global mass can be exactly conserved. The present method is then more efficient than two-step conservative level set scheme. The dispersion-relation-preserving schemes are utilized for the advection terms. The pressure Poisson equation solver is applied to GPU computation using the pCDR library developed by National Center for High-Performance Computing, Taiwan. The SMP parallelization is used to accelerate the rest of calculations. Three benchmark problems were done for the performance evaluation. Good agreements with the referenced solutions are demonstrated for all the investigated problems.
Keywords: Conservative level set method, two-phase flow, dispersion-relation-preserving, graphics processing unit (GPU), multi-threading.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 206512710 Integrating Security Indifference Curve to Formal Decision Evaluation
Authors: Anon Yantarasri, Yachai Limpiyakorn
Abstract:
Decisions are regularly made during a project or daily life. Some decisions are critical and have a direct impact on project or human success. Formal evaluation is thus required, especially for crucial decisions, to arrive at the optimal solution among alternatives to address issues. According to microeconomic theory, all people-s decisions can be modeled as indifference curves. The proposed approach supports formal analysis and decision by constructing indifference curve model from the previous experts- decision criteria. These knowledge embedded in the system can be reused or help naïve users select alternative solution of the similar problem. Moreover, the method is flexible to cope with unlimited number of factors influencing the decision-making. The preliminary experimental results of the alternative selection are accurately matched with the expert-s decisions.Keywords: Decision Analysis and Resolution, Indifference Curve, Multi-criteria Decision Making.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162112709 A New Method for Estimation of the Source Coherency Structure of Wideband Sources
Authors: Yong-jun Zhao, Heng-li Zhang, Zong-yun Hu
Abstract:
Based on the sources- smoothed rank profile (SRP) and modified minimum description length (MMDL) principle, a method for estimation of the source coherency structure (SCS) and the number of wideband sources is proposed in this paper. Instead of focusing, we first use a spatial smoothing technique to pre-process the array covariance matrix of each frequency for de-correlating the sources and then use smoothed rank profile to determine the SCS and the number of wideband sources. We demonstrate the availability of the method by numerical simulations.
Keywords: Wideband sources, source coherency structure (SCS), smoothed rank profile (SRP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 131712708 A Self Organized Map Method to Classify Auditory-Color Synesthesia from Frontal Lobe Brain Blood Volume
Authors: Takashi Kaburagi, Takamasa Komura, Yosuke Kurihara
Abstract:
Absolute pitch is the ability to identify a musical note without a reference tone. Training for absolute pitch often occurs in preschool education. It is necessary to clarify how well the trainee can make use of synesthesia in order to evaluate the effect of the training. To the best of our knowledge, there are no existing methods for objectively confirming whether the subject is using synesthesia. Therefore, in this study, we present a method to distinguish the use of color-auditory synesthesia from the separate use of color and audition during absolute pitch training. This method measures blood volume in the prefrontal cortex using functional Near-infrared spectroscopy (fNIRS) and assumes that the cognitive step has two parts, a non-linear step and a linear step. For the linear step, we assume a second order ordinary differential equation. For the non-linear part, it is extremely difficult, if not impossible, to create an inverse filter of such a complex system as the brain. Therefore, we apply a method based on a self-organizing map (SOM) and are guided by the available data. The presented method was tested using 15 subjects, and the estimation accuracy is reported.
Keywords: Absolute pitch, functional near-infrared spectroscopy, prefrontal cortex, synesthesia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 97812707 Estimation of Real Power Transfer Allocation Using Intelligent Systems
Authors: H. Shareef, A. Mohamed, S. A. Khalid, Aziah Khamis
Abstract:
This paper presents application artificial intelligent (AI) techniques, namely artificial neural network (ANN), adaptive neuro fuzzy interface system (ANFIS), to estimate the real power transfer between generators and loads. Since these AI techniques adopt supervised learning, it first uses modified nodal equation method (MNE) to determine real power contribution from each generator to loads. Then the results of MNE method and load flow information are utilized to estimate the power transfer using AI techniques. The 25-bus equivalent system of south Malaysia is utilized as a test system to illustrate the effectiveness of both AI methods compared to that of the MNE method. The mean squared error of the estimate of ANN and ANFIS power transfer allocation methods are 1.19E-05 and 2.97E-05, respectively. Furthermore, when compared to MNE method, ANN and ANFIS methods computes generator contribution to loads within 20.99 and 39.37msec respectively whereas the MNE method took 360msec for the calculation of same real power transfer allocation.
Keywords: Artificial intelligence, Power tracing, Artificial neural network, ANFIS, Power system deregulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 258412706 Alternative Convergence Analysis for a Kind of Singularly Perturbed Boundary Value Problems
Authors: Jiming Yang
Abstract:
A kind of singularly perturbed boundary value problems is under consideration. In order to obtain its approximation, simple upwind difference discretization is applied. We use a moving mesh iterative algorithm based on equi-distributing of the arc-length function of the current computed piecewise linear solution. First, a maximum norm a posteriori error estimate on an arbitrary mesh is derived using a different method from the one carried out by Chen [Advances in Computational Mathematics, 24(1-4) (2006), 197-212.]. Then, basing on the properties of discrete Green-s function and the presented posteriori error estimate, we theoretically prove that the discrete solutions computed by the algorithm are first-order uniformly convergent with respect to the perturbation parameter ε.
Keywords: Convergence analysis, green's function, singularly perturbed, equi-distribution, moving mesh.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 169612705 Comparison of the Parameter using ECG with Bisepctrum Parameter using EEG during General Anesthesia
Authors: Seong-wan Baik, Soo-young Ye, Byeong-cheol Choi, Gye-rok Jeon
Abstract:
The measurement of anesthetic depth is necessary in anesthesiology. NN10 is very simple method among the RR intervals analysis methods. NN10 parameter means the numbers of above the 10 ms intervals of the normal to normal RR intervals. Bispectrum analysis is defined as 2D FFT. EEG signal reflected the non-linear peristalsis phenomena according to the change brain function. After analyzing the bispectrum of the 2 dimension, the most significant power spectrum density peaks appeared abundantly at the specific area in awakening and anesthesia state. These points are utilized to create the new index since many peaks appeared at the specific area in the frequency coordinate. The measured range of an index was 0-100. An index is 20-50 at an anesthesia, while the index is 90-60 at the awake. In this paper, the relation between NN10 parameter using ECG and bisepctrum index using EEG is observed to estimate the depth of anesthesia during anesthesia and then we estimated the utility of the anesthetic.Keywords: Anesthesia, Bispectrum index, ECG, EEG
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 157412704 Application of a Time-Frequency-Based Blind Source Separation to an Instantaneous Mixture of Secondary Radar Sources
Authors: M. Tria, M. Benidir, E. Chaumette
Abstract:
In Secondary Surveillance Radar (SSR) systems, it is more difficult to locate and recognise aircrafts in the neighbourhood of civil airports since aerial traffic becomes greater. Here, we propose to apply a recent Blind Source Separation (BSS) algorithm based on Time-Frequency Analysis, in order to separate messages sent by different aircrafts and falling in the same radar beam in reception. The above source separation method involves joint-diagonalization of a set of smoothed version of spatial Wigner-Ville distributions. The technique makes use of the difference in the t-f signatures of the nonstationary sources to be separated. Consequently, as the SSR sources emit different messages at different frequencies, the above fitted to this new application. We applied the technique in simulation to separate SSR replies. Results are provided at the end of the paper.Keywords: Blind Source Separation, Time-Frequency Analysis, Secondary Radar
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 169612703 Person-Environment Fit (PE Fit): Evidence from Brazil
Authors: Jucelia Appio, Danielle Deimling De Carli, Bruno Henrique Rocha Fernandes, Nelson Natalino Frizon
Abstract:
The purpose of this paper is to investigate if there are positive and significant correlations between the dimensions of Person-Environment Fit (Person-Job, Person-Organization, Person-Group and Person-Supervisor) at the “Best Companies to Work for” in Brazil in 2017. For that, a quantitative approach was used with a descriptive method being defined as a research sample the "150 Best Companies to Work for", according to data base collected in 2017 and provided by Fundação Instituto of Administração (FIA) of the University of São Paulo (USP). About the data analysis procedures, asymmetry and kurtosis, factorial analysis, Kaiser-Meyer-Olkin (KMO) tests, Bartlett sphericity and Cronbach's alpha were used for the 69 research variables, and as a statistical technique for the purpose of analyzing the hypothesis, Pearson's correlation analysis was performed. As a main result, we highlight that there was a positive and significant correlation between the dimensions of Person-Environment Fit, corroborating the H1 hypothesis that there is a positive and significant correlation between Person-Job Fit, Person-Organization Fit, Person-Group Fit and Person-Supervisor Fit.
Keywords: Human resource management, person-environment fit, strategic people management, best companies to work for.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 100012702 Reliability Evaluation of Composite Electric Power System Based On Latin Hypercube Sampling
Authors: R. Ashok Bakkiyaraj, N. Kumarappan
Abstract:
This paper investigates the suitability of Latin Hypercube sampling (LHS) for composite electric power system reliability analysis. Each sample generated in LHS is mapped into an equivalent system state and used for evaluating the annualized system and load point indices. DC loadflow based state evaluation model is solved for each sampled contingency state. The indices evaluated are loss of load probability, loss of load expectation, expected demand not served and expected energy not supplied. The application of the LHS is illustrated through case studies carried out using RBTS and IEEE-RTS test systems. Results obtained are compared with non-sequential Monte Carlo simulation and state enumeration analytical approaches. An error analysis is also carried out to check the LHS method’s ability to capture the distributions of the reliability indices. It is found that LHS approach estimates indices nearer to actual value and gives tighter bounds of indices than non-sequential Monte Carlo simulation.
Keywords: Composite power system, Latin Hypercube sampling, Monte Carlo simulation, Reliability evaluation, Variance analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 310912701 H.264 Video Privacy Protection Method Using Regions of Interest Encryption
Authors: Taekyun Doo, Cheongmin Ji, Manpyo Hong
Abstract:
Like a closed-circuit television (CCTV), video surveillance system is widely placed for gathering video from unspecified people to prevent crime, surveillance, or many other purposes. However, abuse of CCTV brings about concerns of personal privacy invasions. In this paper, we propose an encryption method to protect personal privacy system in H.264 compressed video bitstream with encrypting only regions of interest (ROI). There is no need to change the existing video surveillance system. In addition, encrypting ROI in compressed video bitstream is a challenging work due to spatial and temporal drift errors. For this reason, we propose a novel drift mitigation method when ROI is encrypted. The proposed method was implemented by using JM reference software based on the H.264 compressed videos, and experimental results show the verification of our proposed methods and its effectiveness.
Keywords: H.264/AVC, video encryption, privacy protection, post compression, region of interest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156812700 Numerical Modeling of Wave Run-Up in Shallow Water Flows Using Moving Wet/Dry Interfaces
Authors: Alia Alghosoun, Michael Herty, Mohammed Seaid
Abstract:
We present a new class of numerical techniques to solve shallow water flows over dry areas including run-up. Many recent investigations on wave run-up in coastal areas are based on the well-known shallow water equations. Numerical simulations have also performed to understand the effects of several factors on tsunami wave impact and run-up in the presence of coastal areas. In all these simulations the shallow water equations are solved in entire domain including dry areas and special treatments are used for numerical solution of singularities at these dry regions. In the present study we propose a new method to deal with these difficulties by reformulating the shallow water equations into a new system to be solved only in the wetted domain. The system is obtained by a change in the coordinates leading to a set of equations in a moving domain for which the wet/dry interface is the reconstructed using the wave speed. To solve the new system we present a finite volume method of Lax-Friedrich type along with a modified method of characteristics. The method is well-balanced and accurately resolves dam-break problems over dry areas.Keywords: Run-up waves, Shallow water equations, finite volume method, wet/dry interface, dam-break problem.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 71012699 Derivation of Darcy’s Law using Homogenization Method
Authors: Kannanut Chamsri
Abstract:
Darcy’s Law is a well-known constitutive equation describing the flow of a fluid through a porous medium. The equation shows a relationship between the superficial or Darcy velocity and the pressure gradient which was first experimentally observed by Henry Darcy in 1855-1856. In this study, we apply homogenization method to Stokes equation in order to derive Darcy’s Law. The process of deriving the equation is complicated, especially in multidimensional domain. Thus, for the sake of simplicity, we use the indicial notation as well as the homogenization. This combination provides a smooth, simple and easy technique to derive Darcy’s Law.
Keywords: Darcy’s Law, Homogenization method, Indicial notation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 501912698 Scheduling a Flexible Flow Shops Problem using DEA
Authors: Fatemeh Dadkhah, Hossein Ali Akbarpour
Abstract:
This paper considers a scheduling problem in flexible flow shops environment with the aim of minimizing two important criteria including makespan and cumulative tardiness of jobs. Since the proposed problem is known as an Np-hard problem in literature, we have to develop a meta-heuristic to solve it. We considered general structure of Genetic Algorithm (GA) and developed a new version of that based on Data Envelopment Analysis (DEA). Two objective functions assumed as two different inputs for each Decision Making Unit (DMU). In this paper we focused on efficiency score of DMUs and efficient frontier concept in DEA technique. After introducing the method we defined two different scenarios with considering two types of mutation operator. Also we provided an experimental design with some computational results to show the performance of algorithm. The results show that the algorithm implements in a reasonable time.Keywords: Data envelopment analysis, Efficiency, Flexible flow shops, Genetic algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 181412697 New Multisensor Data Fusion Method Based on Probabilistic Grids Representation
Authors: Zhichao Zhao, Yi Liu, Shunping Xiao
Abstract:
A new data fusion method called joint probability density matrix (JPDM) is proposed, which can associate and fuse measurements from spatially distributed heterogeneous sensors to identify the real target in a surveillance region. Using the probabilistic grids representation, we numerically combine the uncertainty regions of all the measurements in a general framework. The NP-hard multisensor data fusion problem has been converted to a peak picking problem in the grids map. Unlike most of the existing data fusion method, the JPDM method dose not need association processing, and will not lead to combinatorial explosion. Its convergence to the CRLB with a diminishing grid size has been proved. Simulation results are presented to illustrate the effectiveness of the proposed technique.
Keywords: Cramer-Rao lower bound (CRLB), data fusion, probabilistic grids, joint probability density matrix, localization, sensor network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 180312696 Youth Friendly Health Services for Rural Thai Teenagers
Authors: C. Sridawruang
Abstract:
Young people today has sexual activities differing from those of earlier generations, in that teenagers are likely to have multiple partners, and are frequently in short-term relationships or with partners that are not well known to them. The proportion of teenage mothers in Thailand has increased. Young people were not specifically addressed during the overall very successful HIV-prevention campaigns. Because of this missed opportunity, they are still unaware of the risk of unsafe sexual behavior. Aims: To describe the reproductive health care services in perspectives of rural Thai teenagers Methods: This survey was one part of a mixed method approach taken using survey and focus groups with 439 teenagers aged 12-18 years in 5 villages, Udon Thani, Thailand. The standard questionnaire survey had been used for collecting data. The numeric data was checked and analyzed by using descriptive statistics. Results: Most teenager respondents stated that they do not know where sexual reproductive health services provided for them. Most teenagers felt difficult to access and talk with health staff about sexual related issues. They stated that discussing, or consulting with health providers might not be safe. Teenagers might lose opportunities to access and get advice from health care services. The mean knowledge score of contraception and condom reproductive was 6.34 from a total score 11. Most teenagers especially girls expressed a need for counseling services and reported a need for telephone services. Conclusions: The need of appropriate information focusing on sexual relationships and contraception should be designed to help young people make wise decisions and there should be set health care services for Thai teenagers to make sure that teenagers could access easily. Health care providers need to be trained to improve their knowledge, attitudes and skills in reproductive health care practices for Thai teenagers.
Keywords: Youth friendly health services, rural, Thai, teenagers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130412695 Optimum Time Coordination of Overcurrent Relays using Two Phase Simplex Method
Authors: Prashant P. Bedekar, Sudhir R. Bhide, Vijay S. Kale
Abstract:
Overcurrent (OC) relays are the major protection devices in a distribution system. The operating time of the OC relays are to be coordinated properly to avoid the mal-operation of the backup relays. The OC relay time coordination in ring fed distribution networks is a highly constrained optimization problem which can be stated as a linear programming problem (LPP). The purpose is to find an optimum relay setting to minimize the time of operation of relays and at the same time, to keep the relays properly coordinated to avoid the mal-operation of relays. This paper presents two phase simplex method for optimum time coordination of OC relays. The method is based on the simplex algorithm which is used to find optimum solution of LPP. The method introduces artificial variables to get an initial basic feasible solution (IBFS). Artificial variables are removed using iterative process of first phase which minimizes the auxiliary objective function. The second phase minimizes the original objective function and gives the optimum time coordination of OC relays.Keywords: Constrained optimization, LPP, Overcurrent relaycoordination, Two-phase simplex method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 300712694 The Study of the Intelligent Fuzzy Weighted Input Estimation Method Combined with the Experiment Verification for the Multilayer Materials
Authors: Ming-Hui Lee, Tsung-Chien Chen, Tsu-Ping Yu, Horng-Yuan Jang
Abstract:
The innovative intelligent fuzzy weighted input estimation method (FWIEM) can be applied to the inverse heat transfer conduction problem (IHCP) to estimate the unknown time-varying heat flux of the multilayer materials as presented in this paper. The feasibility of this method can be verified by adopting the temperature measurement experiment. The experiment modular may be designed by using the copper sample which is stacked up 4 aluminum samples with different thicknesses. Furthermore, the bottoms of copper samples are heated by applying the standard heat source, and the temperatures on the tops of aluminum are measured by using the thermocouples. The temperature measurements are then regarded as the inputs into the presented method to estimate the heat flux in the bottoms of copper samples. The influence on the estimation caused by the temperature measurement of the sample with different thickness, the processing noise covariance Q, the weighting factor γ , the sampling time interval Δt , and the space discrete interval Δx , will be investigated by utilizing the experiment verification. The results show that this method is efficient and robust to estimate the unknown time-varying heat input of the multilayer materials.Keywords: Multilayer Materials, Input Estimation Method, IHCP, Heat Flux.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 123812693 3D Liver Segmentation from CT Images Using a Level Set Method Based on a Shape and Intensity Distribution Prior
Authors: Nuseiba M. Altarawneh, Suhuai Luo, Brian Regan, Guijin Tang
Abstract:
Liver segmentation from medical images poses more challenges than analogous segmentations of other organs. This contribution introduces a liver segmentation method from a series of computer tomography images. Overall, we present a novel method for segmenting liver by coupling density matching with shape priors. Density matching signifies a tracking method which operates via maximizing the Bhattacharyya similarity measure between the photometric distribution from an estimated image region and a model photometric distribution. Density matching controls the direction of the evolution process and slows down the evolving contour in regions with weak edges. The shape prior improves the robustness of density matching and discourages the evolving contour from exceeding liver’s boundaries at regions with weak boundaries. The model is implemented using a modified distance regularized level set (DRLS) model. The experimental results show that the method achieves a satisfactory result. By comparing with the original DRLS model, it is evident that the proposed model herein is more effective in addressing the over segmentation problem. Finally, we gauge our performance of our model against matrices comprising of accuracy, sensitivity, and specificity.
Keywords: Bhattacharyya distance, distance regularized level set (DRLS) model, liver segmentation, level set method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 233812692 The Application of Homotopy Method In Solving Electrical Circuit Design Problem
Authors: Talib Hashim Hasan
Abstract:
This paper describes simple implementation of homotopy (also called continuation) algorithm for determining the proper resistance of the resistor to dissipate energy at a specified rate of an electric circuit. Homotopy algorithm can be considered as a developing of the classical methods in numerical computing such as Newton-Raphson and fixed point methods. In homoptopy methods, an embedding parameter is used to control the convergence. The method purposed in this work utilizes a special homotopy called Newton homotopy. Numerical example solved in MATLAB is given to show the effectiveness of the purposed methodKeywords: electrical circuit homotopy, methods, MATLAB, Newton homotopy
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 303312691 Preconditioned Generalized Accelerated Overrelaxation Methods for Solving Certain Nonsingular Linear System
Authors: Deyu Sun, Guangbin Wang
Abstract:
In this paper, we present preconditioned generalized accelerated overrelaxation (GAOR) methods for solving certain nonsingular linear system. We compare the spectral radii of the iteration matrices of the preconditioned and the original methods. The comparison results show that the preconditioned GAOR methods converge faster than the GAOR method whenever the GAOR method is convergent. Finally, we give two numerical examples to confirm our theoretical results.
Keywords: Preconditioned, GAOR method, linear system, convergence, comparison.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156712690 Finite Element Analysis of Thermally-Induced Bistable Plate Using Four Plate Elements
Authors: Jixiao Tao, Xiaoqiao He
Abstract:
The present study deals with the finite element (FE) analysis of thermally-induced bistable plate using various plate elements. The quadrilateral plate elements include the 4-node conforming plate element based on the classical laminate plate theory (CLPT), the 4-node and 9-node Mindlin plate element based on the first-order shear deformation laminated plate theory (FSDT), and a displacement-based 4-node quadrilateral element (RDKQ-NL20). Using the von-Karman’s large deflection theory and the total Lagrangian (TL) approach, the nonlinear FE governing equations for plate under thermal load are derived. Convergence analysis for four elements is first conducted. These elements are then used to predict the stable shapes of thermally-induced bistable plate. Numerical test shows that the plate element based on FSDT, namely the 4-node and 9-node Mindlin, and the RDKQ-NL20 plate element can predict two stable cylindrical shapes while the 4-node conforming plate predicts a saddles shape. Comparing the simulation results with ABAQUS, the RDKQ-NL20 element shows the best accuracy among all the elements.
Keywords: Finite element method, geometrical nonlinearity, bistable, quadrilateral plate elements.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 78812689 Using Simulation Modeling Approach to Predict USMLE Steps 1 and 2 Performances
Authors: Chau-Kuang Chen, John Hughes, Jr., A. Dexter Samuels
Abstract:
The prediction models for the United States Medical Licensure Examination (USMLE) Steps 1 and 2 performances were constructed by the Monte Carlo simulation modeling approach via linear regression. The purpose of this study was to build robust simulation models to accurately identify the most important predictors and yield the valid range estimations of the Steps 1 and 2 scores. The application of simulation modeling approach was deemed an effective way in predicting student performances on licensure examinations. Also, sensitivity analysis (a/k/a what-if analysis) in the simulation models was used to predict the magnitudes of Steps 1 and 2 affected by changes in the National Board of Medical Examiners (NBME) Basic Science Subject Board scores. In addition, the study results indicated that the Medical College Admission Test (MCAT) Verbal Reasoning score and Step 1 score were significant predictors of the Step 2 performance. Hence, institutions could screen qualified student applicants for interviews and document the effectiveness of basic science education program based on the simulation results.Keywords: Prediction Model, Sensitivity Analysis, Simulation Method, USMLE.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 146112688 An Improved Method on Static Binary Analysis to Enhance the Context-Sensitive CFI
Authors: Qintao Shen, Lei Luo, Jun Ma, Jie Yu, Qingbo Wu, Yongqi Ma, Zhengji Liu
Abstract:
Control Flow Integrity (CFI) is one of the most promising technique to defend Code-Reuse Attacks (CRAs). Traditional CFI Systems and recent Context-Sensitive CFI use coarse control flow graphs (CFGs) to analyze whether the control flow hijack occurs, left vast space for attackers at indirect call-sites. Coarse CFGs make it difficult to decide which target to execute at indirect control-flow transfers, and weaken the existing CFI systems actually. It is an unsolved problem to extract CFGs precisely and perfectly from binaries now. In this paper, we present an algorithm to get a more precise CFG from binaries. Parameters are analyzed at indirect call-sites and functions firstly. By comparing counts of parameters prepared before call-sites and consumed by functions, targets of indirect calls are reduced. Then the control flow would be more constrained at indirect call-sites in runtime. Combined with CCFI, we implement our policy. Experimental results on some popular programs show that our approach is efficient. Further analysis show that it can mitigate COOP and other advanced attacks.Keywords: Contex-sensitive, CFI, binary analysis, code reuse attack.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 94312687 Introducing a Platform for Encryption Algorithms
Authors: Ahmad Habibizad Navin, Yasaman Hashemi, Omid Mirmotahari
Abstract:
In this paper, we introduce a novel platform encryption method, which modify its keys and random number generators step by step during encryption algorithms. According to complexity of the proposed algorithm, it was safer than any other method.Keywords: Decryption, Encryption, Algorithm, security.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1435