Search results for: Derived calculation methods
5164 Comparison of Methods of Estimation for Use in Goodness of Fit Tests for Binary Multilevel Models
Authors: I. V. Pinto, M. R. Sooriyarachchi
Abstract:
It can be frequently observed that the data arising in our environment have a hierarchical or a nested structure attached with the data. Multilevel modelling is a modern approach to handle this kind of data. When multilevel modelling is combined with a binary response, the estimation methods get complex in nature and the usual techniques are derived from quasi-likelihood method. The estimation methods which are compared in this study are, marginal quasi-likelihood (order 1 & order 2) (MQL1, MQL2) and penalized quasi-likelihood (order 1 & order 2) (PQL1, PQL2). A statistical model is of no use if it does not reflect the given dataset. Therefore, checking the adequacy of the fitted model through a goodness-of-fit (GOF) test is an essential stage in any modelling procedure. However, prior to usage, it is also equally important to confirm that the GOF test performs well and is suitable for the given model. This study assesses the suitability of the GOF test developed for binary response multilevel models with respect to the method used in model estimation. An extensive set of simulations was conducted using MLwiN (v 2.19) with varying number of clusters, cluster sizes and intra cluster correlations. The test maintained the desirable Type-I error for models estimated using PQL2 and it failed for almost all the combinations of MQL. Power of the test was adequate for most of the combinations in all estimation methods except MQL1. Moreover, models were fitted using the four methods to a real-life dataset and performance of the test was compared for each model.
Keywords: Goodness-of-fit test, marginal quasi-likelihood, multilevel modelling, type-I error, penalized quasi-likelihood, power, quasi-likelihood.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7335163 Improving Digital Image Edge Detection by Fuzzy Systems
Authors: Begol, Moslem, Maghooli, Keivan
Abstract:
Image Edge Detection is one of the most important parts of image processing. In this paper, by fuzzy technique, a new method is used to improve digital image edge detection. In this method, a 3x3 mask is employed to process each pixel by means of vicinity. Each pixel is considered a fuzzy input and by examining fuzzy rules in its vicinity, the edge pixel is specified and by utilizing calculation algorithms in image processing, edges are displayed more clearly. This method shows significant improvement compared to different edge detection methods (e.g. Sobel, Canny).Keywords: Fuzzy Systems, Edge Detection, Fuzzy edgedetection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20875162 Large-Eddy Simulation of Hypersonic Configuration Aerodynamics
Authors: Huang Shengqin, Xiao Hong
Abstract:
LES with mixed subgrid-scale model has been used to simulate aerodynamic performance of hypersonic configuration. The simulation was conducted to replicate conditions and geometry of a model which has been previously tested. LES Model has been successful in predict pressure coefficient with the max error 1.5% besides afterbody. But in the high Mach number condition, it is poor in predict ability and product 12.5% error. The calculation error are mainly conducted by the distribution swirling. The fact of poor ability in the high Mach number and afterbody region indicated that the mixed subgrid-scale model should be improved in large eddied especially in hypersonic separate region. In the condition of attach and sideslip flight, the calculation results have waves. LES are successful in the prediction the pressure wave in hypersonic flow.Keywords: Hypersonic, LES, mixed Subgrid-scale model, experiment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15725161 Describing the Fine Electronic Structure and Predicting Properties of Materials with ATOMIC MATTERS Computation System
Authors: Rafal Michalski, Jakub Zygadlo
Abstract:
We present the concept and scientific methods and algorithms of our computation system called ATOMIC MATTERS. This is the first presentation of the new computer package, that allows its user to describe physical properties of atomic localized electron systems subject to electromagnetic interactions. Our solution applies to situations where an unclosed electron 2p/3p/3d/4d/5d/4f/5f subshell interacts with an electrostatic potential of definable symmetry and external magnetic field. Our methods are based on Crystal Electric Field (CEF) approach, which takes into consideration the electrostatic ligands field as well as the magnetic Zeeman effect. The application allowed us to predict macroscopic properties of materials such as: Magnetic, spectral and calorimetric as a result of physical properties of their fine electronic structure. We emphasize the importance of symmetry of charge surroundings of atom/ion, spin-orbit interactions (spin-orbit coupling) and the use of complex number matrices in the definition of the Hamiltonian. Calculation methods, algorithms and convention recalculation tools collected in ATOMIC MATTERS were chosen to permit the prediction of magnetic and spectral properties of materials in isostructural series.Keywords: Atomic matters, crystal electric field, spin-orbit coupling, localized states, electron subshell, fine electronic structure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12085160 Queen-bee Algorithm for Energy Efficient Clusters in Wireless Sensor Networks
Authors: Z. Pooranian, A. Barati, A. Movaghar
Abstract:
Wireless sensor networks include small nodes which have sensing ability; calculation and connection extend themselves everywhere soon. Such networks have source limitation on connection, calculation and energy consumption. So, since the nodes have limited energy in sensor networks, the optimized energy consumption in these networks is of more importance and has created many challenges. The previous works have shown that by organizing the network nodes in a number of clusters, the energy consumption could be reduced considerably. So the lifetime of the network would be increased. In this paper, we used the Queen-bee algorithm to create energy efficient clusters in wireless sensor networks. The Queen-bee (QB) is similar to nature in that the queen-bee plays a major role in reproduction process. The QB is simulated with J-sim simulator. The results of the simulation showed that the clustering by the QB algorithm decreases the energy consumption with regard to the other existing algorithms and increases the lifetime of the network.Keywords: Queen-bee, sensor network, energy efficient, clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19745159 Hydrochemical Contamination Profiling and Spatial-Temporal Mapping with the Support of Multivariate and Cluster Statistical Analysis
Authors: S. Barbosa, M. Pinto, J. A. Almeida, E. Carvalho, C. Diamantino
Abstract:
The aim of this work was to test a methodology able to generate spatial-temporal maps that can synthesize simultaneously the trends of distinct hydrochemical indicators in an old radium-uranium tailings dam deposit. Multidimensionality reduction derived from principal component analysis and subsequent data aggregation derived from clustering analysis allow to identify distinct hydrochemical behavioral profiles and generate synthetic evolutionary hydrochemical maps.
Keywords: Contamination plume migration, K-means of PCA scores, groundwater and mine water monitoring, spatial-temporal hydrochemical trends.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6255158 Numerical Modeling of Temperature Fields in Aviation Gas Turbine Elements
Authors: A. M. Pashaev, R. A. Sadihov, A. S. Samedov, C. Ardil
Abstract:
A mathematical model and a numerical method for computing the temperature field of the profile part of convectionally cooled blades are developed. The theoretical substantiation of the method is proved by corresponding theorems. To this end, convergent quadrature processes were developed and error estimates were obtained in terms of the Zygmund continuity moduli. The boundary conditions for heat exchange are determined from the solution of the corresponding integral equations and empirical relations. The reliability of the developed methods is confirmed by calculation and experimental studies of the thermohydraulic characteristics of the nozzle apparatus of the first stage of the gas turbine.
Keywords: Aviation gas turbine, temperature field, cooled blades, numerical modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7515157 Obstacle Classification Method Based On 2D LIDAR Database
Authors: Moohyun Lee, Soojung Hur, Yongwan Park
Abstract:
We propose obstacle classification method based on 2D LIDAR Database. The existing obstacle classification method based on 2D LIDAR, has an advantage in terms of accuracy and shorter calculation time. However, it was difficult to classifier the type of obstacle and therefore accurate path planning was not possible. In order to overcome this problem, a method of classifying obstacle type based on width data of obstacle was proposed. However, width data was not sufficient to improve accuracy. In this paper, database was established by width and intensity data; the first classification was processed by the width data; the second classification was processed by the intensity data; classification was processed by comparing to database; result of obstacle classification was determined by finding the one with highest similarity values. An experiment using an actual autonomous vehicle under real environment shows that calculation time declined in comparison to 3D LIDAR and it was possible to classify obstacle using single 2D LIDAR.
Keywords: Obstacle, Classification, LIDAR, Segmentation, Width, Intensity, Database.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34455156 Exponentiated Transmuted Weibull Distribution A Generalization of the Weibull Distribution
Authors: Abd El Hady N. Ebraheim
Abstract:
This paper introduces a new generalization of the two parameter Weibull distribution. To this end, the quadratic rank transmutation map has been used. This new distribution is named exponentiated transmuted Weibull (ETW) distribution. The ETW distribution has the advantage of being capable of modeling various shapes of aging and failure criteria. Furthermore, eleven lifetime distributions such as the Weibull, exponentiated Weibull, Rayleigh and exponential distributions, among others follow as special cases. The properties of the new model are discussed and the maximum likelihood estimation is used to estimate the parameters. Explicit expressions are derived for the quantiles. The moments of the distribution are derived, and the order statistics are examined.
Keywords: Exponentiated, Inversion Method, Maximum Likelihood Estimation, Transmutation Map.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34715155 Some Issues of Measurement of Impairment of Non-Financial Assets in the Public Sector
Authors: Mariam Vardiashvili
Abstract:
The economic value of the asset impairment process is quite large. Impairment reflects the reduction of future economic benefits or service potentials itemized in the asset. The assets owned by public sector entities bring economic benefits or are used for delivery of the free-of-charge services. Consequently, they are classified as cash-generating and non-cash-generating assets. IPSAS 21 - Impairment of non-cash-generating assets, and IPSAS 26 - Impairment of cash-generating assets, have been designed considering this specificity. When measuring impairment of assets, it is important to select the relevant methods. For measurement of the impaired Non-Cash-Generating Assets, IPSAS 21 recommends three methods: Depreciated Replacement Cost Approach, Restoration Cost Approach, and Service Units Approach. Impairment of Value in Use of Cash-Generating Assets (according to IPSAS 26) is measured by discounted value of the money sources to be received in future. Value in use of the cash-generating asserts (as per IPSAS 26) is measured by the discounted value of the money sources to be received in the future. The article provides classification of the assets in the public sector as non-cash-generating assets and cash-generating assets and, deals also with the factors which should be considered when evaluating impairment of assets. An essence of impairment of the non-financial assets and the methods of measurement thereof evaluation are formulated according to IPSAS 21 and IPSAS 26. The main emphasis is put on different methods of measurement of the value in use of the impaired Cash-Generating Assets and Non-Cash-Generation Assets and the methods of their selection. The traditional and the expected cash flow approaches for calculation of the discounted value are reviewed. The article also discusses the issues of recognition of impairment loss and its reflection in the financial reporting. The article concludes that despite a functional purpose of the impaired asset, whichever method is used for measuring the asset, presentation of realistic information regarding the value of the assets should be ensured in the financial reporting. In the theoretical development of the issue, the methods of scientific abstraction, analysis and synthesis were used. The research was carried out with a systemic approach. The research process uses international standards of accounting, theoretical researches and publications of Georgian and foreign scientists.
Keywords: Non-cash-generating assets, cash-generating assets, recoverable value, recoverable service amount, value in use.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6985154 An Approach to Correlate the Statistical-Based Lorenz Method, as a Way of Measuring Heterogeneity, with Kozeny-Carman Equation
Authors: H. Khanfari, M. Johari Fard
Abstract:
Dealing with carbonate reservoirs can be mind-boggling for the reservoir engineers due to various digenetic processes that cause a variety of properties through the reservoir. A good estimation of the reservoir heterogeneity which is defined as the quality of variation in rock properties with location in a reservoir or formation, can better help modeling the reservoir and thus can offer better understanding of the behavior of that reservoir. Most of reservoirs are heterogeneous formations whose mineralogy, organic content, natural fractures, and other properties vary from place to place. Over years, reservoir engineers have tried to establish methods to describe the heterogeneity, because heterogeneity is important in modeling the reservoir flow and in well testing. Geological methods are used to describe the variations in the rock properties because of the similarities of environments in which different beds have deposited in. To illustrate the heterogeneity of a reservoir vertically, two methods are generally used in petroleum work: Dykstra-Parsons permeability variations (V) and Lorenz coefficient (L) that are reviewed briefly in this paper. The concept of Lorenz is based on statistics and has been used in petroleum from that point of view. In this paper, we correlated the statistical-based Lorenz method to a petroleum concept, i.e. Kozeny-Carman equation and derived the straight line plot of Lorenz graph for a homogeneous system. Finally, we applied the two methods on a heterogeneous field in South Iran and discussed each, separately, with numbers and figures. As expected, these methods show great departure from homogeneity. Therefore, for future investment, the reservoir needs to be treated carefully.
Keywords: Carbonate reservoirs, heterogeneity, homogeneous system, Dykstra-Parsons permeability variations (V), Lorenz coefficient (L).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17905153 Perturbation Based Modelling of Differential Amplifier Circuit
Authors: Rahul Bansal, Sudipta Majumdar
Abstract:
This paper presents the closed form nonlinear expressions of bipolar junction transistor (BJT) differential amplifier (DA) using perturbation method. Circuit equations have been derived using Kirchhoff’s voltage law (KVL) and Kirchhoff’s current law (KCL). The perturbation method has been applied to state variables for obtaining the linear and nonlinear terms. The implementation of the proposed method is simple. The closed form nonlinear expressions provide better insights of physical systems. The derived equations can be used for signal processing applications.Keywords: Differential amplifier, perturbation method, Taylor series.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10175152 Fundamental Problems in the Operation of the Automotive Parts Industry Small and Medium Businesses in Bangkok and Surrounding Provinces
Authors: P. Thepnarintra
Abstract:
The purposes of this study were to: 1) investigate operation conditions of SME automotive part industry in Bangkok and vicinity and 2) to compare operation problem levels of SME automotive part industry in Bangkok and vicinity according to the sizes of the enterprises. Samples in this study included 196 entrepreneurs of SME automotive part industry in Bangkok and vicinity derived from simple random sampling and calculation from R. V. Krejcie and D. W. Morgan’s tables. Research statistics included frequency, percentage, mean, standard deviation, and T-test. The results revealed that in general the problem levels of SME automotive part industry in Bangkok and vicinity were high. When considering in details, it was found that the problem levels were high at every aspect, i.e. personal, production, export, finance, and marketing respectively. The comparison of the problem levels according to the sizes of the enterprises revealed statistically significant differences at .05. When considering on each aspect, it was found that the aspect with the statistical difference at .05 included 5 aspects, i.e. production, marketing, finance, personal, and export. The findings also showed that small enterprises faced more severe problems than those of medium enterprises.Keywords: Automotive part industry, operation problems, SME, perimeter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15955151 Analyzing Defects with Failure Assessment Diagrams of Gas Pipelines
Authors: Alfred Hasanaj, Ardit Gjeta, Miranda Kullolli
Abstract:
The approach in analyzing defects on different pipe lines is conducted through Failure Assessment Diagram (FAD). These methods of analyses have further extended in recent years. This approach is used to identify and stress out a solution for the defects which randomly occur with gas pipes such are corrosion defects, gauge defects, and combination of defects where gauge and dents are included. Few of the defects are to be analyzed in this paper where our main focus will be the fracture of cast Iron pipes, elastic-plastic failure and plastic collapse of X52 steel pipes for gas transport. We need to conduct a calculation of probability of the defects in order to predict and avoid such costly defects.
Keywords: Defects, Failure Assessment Diagrams, Safety Factor Steel Pipes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29475150 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm
Authors: Ameur Abdelkader, Abed Bouarfa Hafida
Abstract:
Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.
Keywords: Predictive analysis, big data, predictive analysis algorithms. CART algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10755149 Accurate Calculation of Free Frequencies of Beams and Rectangular Plates
Authors: R .Lassoued, M. Guenfoud
Abstract:
An accurate procedure to determine free vibrations of beams and plates is presented. The natural frequencies are exact solutions of governing vibration equations witch load to a nonlinear homogeny system. The bilinear and linear structures considered simulate a bridge. The dynamic behavior of this one is analyzed by using the theory of the orthotropic plate simply supported on two sides and free on the two others. The plate can be excited by a convoy of constant or harmonic loads. The determination of the dynamic response of the structures considered requires knowledge of the free frequencies and the shape modes of vibrations. Our work is in this context. Indeed, we are interested to develop a self-consistent calculation of the Eigen frequencies. The formulation is based on the determination of the solution of the differential equations of vibrations. The boundary conditions corresponding to the shape modes permit to lead to a homogeneous system. Determination of the noncommonplace solutions of this system led to a nonlinear problem in Eigen frequencies. We thus, develop a computer code for the determination of the eigenvalues. It is based on a method of bisection with interpolation whose precision reaches 10 -12. Moreover, to determine the corresponding modes, the calculation algorithm that we develop uses the method of Gauss with a partial optimization of the "pivots" combined with an inverse power procedure. The Eigen frequencies of a plate simply supported along two opposite sides while considering the two other free sides are thus analyzed. The results could be generalized with the case of a beam by regarding it as a plate with low width. We give, in this paper, some examples of treated cases. The comparison with results presented in the literature is completely satisfactory.Keywords: Free frequencies, beams, rectangular plates.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21935148 Investigation on Mesh Sensitivity of a Transient Model for Nozzle Clogging
Authors: H. Barati, M. Wu, A. Kharicha, A. Ludwig
Abstract:
A transient model for nozzle clogging has been developed and successfully validated against a laboratory experiment. Key steps of clogging are considered: transport of particles by turbulent flow towards the nozzle wall; interactions between fluid flow and nozzle wall, and the adhesion of the particle on the wall; the growth of the clog layer and its interaction with the flow. The current paper is to investigate the mesh (size and type) sensitivity of the model in both two and three dimensions. It is found that the algorithm for clog growth alone excluding the flow effect is insensitive to the mesh type and size, but the calculation including flow becomes sensitive to the mesh quality. The use of 2D meshes leads to overestimation of the clog growth because the 3D nature of flow in the boundary layer cannot be properly solved by 2D calculation. 3D simulation with tetrahedron mesh can also lead to an error estimation of the clog growth. A mesh-independent result can be achieved with hexahedral mesh, or at least with triangular prism (inflation layer) for near-wall regions.
Keywords: Clogging, nozzle, numerical model, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8415147 Novel Anti-leukemia Calanone Compounds by Quantitative Structure-Activity Relationship AM1 Semiempirical Method
Authors: Ponco Iswanto, Mochammad Chasani, Muhammad Hanafi, Iqmal Tahir, Eva Vaulina YD, Harjono, Lestari Solikhati, Winkanda S. Putra, Yayuk Yuliantini
Abstract:
Quantitative Structure-Activity Relationship (QSAR) approach for discovering novel more active Calanone derivative as anti-leukemia compound has been conducted. There are 6 experimental activities of Calanone compounds against leukemia cell L1210 that are used as material of the research. Calculation of theoretical predictors (independent variables) was performed by AM1 semiempirical method. The QSAR equation is determined by Principle Component Regression (PCR) analysis, with Log IC50 as dependent variable and the independent variables are atomic net charges, dipole moment (μ), and coefficient partition of noctanol/ water (Log P). Three novel Calanone derivatives that obtained by this research have higher activity against leukemia cell L1210 than pure Calanone.Keywords: AM1 semiempirical calculation, Calanone, Principle Component Regression, QSAR approach.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14785146 Improving Injection Moulding Processes Using Experimental Design
Authors: Yousef Amer, Mehdi Moayyedian, Zeinab Hajiabolhasani, Lida Moayyedian
Abstract:
Moulded parts contribute to more than 70% of components in products. However, common defects particularly in plastic injection moulding exist such as: warpage, shrinkage, sink marks, and weld lines. In this paper Taguchi experimental design methods are applied to reduce the warpage defect of thin plate Acrylonitrile Butadiene Styrene (ABS) and are demonstrated in two levels; namely, orthogonal arrays of Taguchi and the Analysis of Variance (ANOVA). Eight trials have been run in which the optimal parameters that can minimize the warpage defect in factorial experiment are obtained. The results obtained from ANOVA approach analysis with respect to those derived from MINITAB illustrate the most significant factors which may cause warpage in injection moulding process. Moreover, ANOVA approach in comparison with other approaches like S/N ratio is more accurate and with the interaction of factors it is possible to achieve higher and the better outcomes.Keywords: Analysis of variance, ANOVA, plastic injection mould, Taguchi methods, Warpage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 38975145 On a Way for Constructing Numerical Methods on the Joint of Multistep and Hybrid Methods
Authors: G.Mehdiyeva, M.Imanova, V.Ibrahimov
Abstract:
Taking into account that many problems of natural sciences and engineering are reduced to solving initial-value problem for ordinary differential equations, beginning from Newton, the scientists investigate approximate solution of ordinary differential equations. There are papers of different authors devoted to the solution of initial value problem for ODE. The Euler-s known method that was developed under the guidance of the famous scientists Adams, Runge and Kutta is the most popular one among these methods. Recently the scientists began to construct the methods preserving some properties of Adams and Runge-Kutta methods and called them hybrid methods. The constructions of such methods are investigated from the middle of the XX century. Here we investigate one generalization of multistep and hybrid methods and on their base we construct specific methods of accuracy order p = 5 and p = 6 for k = 1 ( k is the order of the difference method).Keywords: Multistep and hybrid methods, initial value problem, degree and stability of hybrid methods
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15975144 An Approximation Method for Exact Boundary Controllability of Euler-Bernoulli System
Authors: Abdelaziz Khernane, Naceur Khelil, Leila Djerou
Abstract:
The aim of this work is to study the numerical implementation of the Hilbert Uniqueness Method for the exact boundary controllability of Euler-Bernoulli beam equation. This study may be difficult. This will depend on the problem under consideration (geometry, control and dimension) and the numerical method used. Knowledge of the asymptotic behaviour of the control governing the system at time T may be useful for its calculation. This idea will be developed in this study. We have characterized as a first step, the solution by a minimization principle and proposed secondly a method for its resolution to approximate the control steering the considered system to rest at time T.Keywords: Boundary control, exact controllability, finite difference methods, functional optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14875143 FEA-Based Calculation of Performances of IPM Machines with Five Topologies for Hybrid- Electric Vehicle Traction
Authors: Aimeng Wang, Dejun Ma, Hui Wang
Abstract:
The paper presents a detailed calculation of characteristic of five different topology permanent magnet machines for high performance traction including hybrid -electric vehicles using finite element analysis (FEA) method. These machines include V-shape single layer interior PM, W-shape single-layer interior PM, Segment interior PM and surface PM on the rotor and with distributed winding on the stator. The performance characteristics which include the back-emf voltage and its harmonic, magnet mass, iron loss and ripple torque are compared and analyzed. One of a 7.5kW IPM prototype was tested and verified finite-element analysis results. The aim of the paper is given some guidance and reference for machine designer which are interested in IPM machine selection for high performance traction application.
Keywords: Interior permanent magnet machine, finite-element analysis (FEA), five topologies, electric vehicle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39255142 Bit-Error-Rate Performance Analysis of an Overlap-based CSS System
Authors: Taeung Yoon, Dahae Chong, Sangho Ahn, Seokho Yoon
Abstract:
In a chirp spread spectrum (CSS) system, the overlap technique is used for increasing bit rate. More overlaps can offer higher data throughput; however, they may cause more intersymbol interference (ISI) at the same time, resulting in serious bit error rate (BER) performance degradation. In this paper, we perform the BER analysis and derive a closed form BER expression for the overlap-based CSS system. The derived BER expression includes the number of overlaps as a parameter, and thus, would be very useful in determining the number of overlaps for a specified BER. The numerical results demonstrate that the BER derived in a closed form closely agrees with the simulated BER.Keywords: CSS, DM, chirp, overlap.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21005141 Study on Performance of Wigner Ville Distribution for Linear FM and Transient Signal Analysis
Authors: Azeemsha Thacham Poyil, Nasimudeen KM
Abstract:
This research paper presents some methods to assess the performance of Wigner Ville Distribution for Time-Frequency representation of non-stationary signals, in comparison with the other representations like STFT, Spectrogram etc. The simultaneous timefrequency resolution of WVD is one of the important properties which makes it preferable for analysis and detection of linear FM and transient signals. There are two algorithms proposed here to assess the resolution and to compare the performance of signal detection. First method is based on the measurement of area under timefrequency plot; in case of a linear FM signal analysis. A second method is based on the instantaneous power calculation and is used in case of transient, non-stationary signals. The implementation is explained briefly for both methods with suitable diagrams. The accuracy of the measurements is validated to show the better performance of WVD representation in comparison with STFT and Spectrograms.
Keywords: WVD: Wigner Ville Distribution, STFT: Short Time Fourier Transform, FT: Fourier Transform, TFR: Time-Frequency Representation, FM: Frequency Modulation, LFM Signal: Linear FM Signal, JTFA: Joint time frequency analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24235140 Investigation of Electromagnetic Force in 3P5W Busbar System under Peak Short-Circuit Current
Authors: Farhana Mohamad Yusop, Syafrudin Masri, Dahaman Ishak, Mohamad Kamarol
Abstract:
Electromagnetic forces on three-phase five-wire (3P5W) busbar system is investigated under three-phase short-circuits current. The conductor busbar placed in compact galvanized steel enclosure is in the rectangular shape. Transient analysis from Opera-2D is carried out to develop the model of three-phase short-circuits current in the system. The result of the simulation is compared with the calculation result, which is obtained by applying the theories of Biot Savart’s law and Laplace equation. Under this analytical approach, the moment of peak short-circuit current is taken into account. The effect upon geometrical arrangement of the conductor and the present of the steel enclosure are considered by the theory of image. The result depict that the electromagnetic force due to the transient short-circuit from simulation is agreed with the calculation.
Keywords: Busbar, electromagnetic force, short-circuit current, transient analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39945139 Dark and Bright Envelopes for Dehazing Images
Authors: Zihan Yu, Kohei Inoue, Kiichi Urahama
Abstract:
We present a method for dehazing images. A dark envelope image is derived with the bilateral minimum filter and a bright envelope is derived with the bilateral maximum filter. The ambient light and transmission of the scene are estimated from these two envelope images. An image without haze is reconstructed from the estimated ambient light and transmission.
Keywords: Image dehazing, bilateral minimum filter, bilateral maximum filter, local contrast.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20025138 Parallel Priority Region Approach to Detect Background
Authors: Sallama Athab, Hala Bahjat, Zhang Yinghui
Abstract:
Background detection is essential in video analyses; optimization is often needed in order to achieve real time calculation. Information gathered by dual cameras placed in the front and rear part of an Autonomous Vehicle (AV) is integrated for background detection. In this paper, real time calculation is achieved on the proposed technique by using Priority Regions (PR) and Parallel Processing together where each frame is divided into regions then and each region process is processed in parallel. PR division depends upon driver view limitations. A background detection system is built on the Temporal Difference (TD) and Gaussian Filtering (GF). Temporal Difference and Gaussian Filtering with multi threshold and sigma (weight) value are be based on PR characteristics. The experiment result is prepared on real scene. Comparison of the speed and accuracy with traditional background detection techniques, the effectiveness of PR and parallel processing are also discussed in this paper.
Keywords: Autonomous Vehicle, Background Detection, Dual Camera, Gaussian Filtering, Parallel Processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16975137 Fully Parameterizable FPGA based Crypto-Accelerator
Authors: Iqbalur Rahman, Miftahur Rahman, Abul L Haque, Mostafizur Rahman,
Abstract:
In this paper, RSA encryption algorithm and its hardware implementation in Xilinx-s Virtex Field Programmable Gate Arrays (FPGA) is analyzed. The issues of scalability, flexible performance, and silicon efficiency for the hardware acceleration of public key crypto systems are being explored in the present work. Using techniques based on the interleaved math for exponentiation, the proposed RSA calculation architecture is compared to existing FPGA-based solutions for speed, FPGA utilization, and scalability. The paper covers the RSA encryption algorithm, interleaved multiplication, Miller Rabin algorithm for primality test, extended Euclidean math, basic FPGA technology, and the implementation details of the proposed RSA calculation architecture. Performance of several alternative hardware architectures is discussed and compared. Finally, conclusion is drawn, highlighting the advantages of a fully flexible & parameterized design.Keywords: Crypto Accelerator, FPGA, Public Key Cryptography, RSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27695136 Best Option for Countercyclical Capital Buffer Implementation - Scenarios for Baltic States
Authors: Ģirts Brasliņš, Ilja Arefjevs, Nadežda Tarakanova
Abstract:
The objective of countercyclical capital buffer is to encourage banks to build up buffers in good times that can be drawn down in bad times. The aim of the report is to assess such decisions by banks derived from three approaches. The approaches are the aggregate credit-to-GDP ratio, credit growth as well as banking sector profits. The approaches are implemented for Estonia, Latvia and Lithuania for the time period 2000-2012. The report compares three approaches and analyses their relevance to the Baltic States by testing the correlation between a growth in studied variables and a growth of corresponding gaps. Methods used in the empirical part of the report are econometric analysis as well as economic analysis, development indicators, relative and absolute indicators and other methods. The research outcome is a cross-Baltic comparison of two alternative approaches to establish or release a countercyclical capital buffer by banks and their implications for each Baltic country.
Keywords: Basel III, countercyclical capital buffer, banks, credit growth, Baltic States.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18715135 Relative Injective Modules and Relative Flat Modules
Authors: Jianmin Xing, Rufeng Xing
Abstract:
Let R be a ring, n a fixed nonnegative integer. The concepts of (n, 0)-FI-injective and (n, 0)-FI-flat modules, and then give some characterizations of these modules over left n-coherent rings are introduced . In addition, we investigate the left and right n-FI-resolutions of R-modules by left (right) derived functors Extn(−,−) (Torn(−,−) ) over a left n-coherent ring, where n-FI stands for the categories of all (n, 0)- injective left R-modules. These modules together with the left or right derived functors are used to study the (n, 0)-injective dimensions of modules and rings.
Keywords: (n, 0)-injective module, (n, 0)-injective dimension, (n, 0)-FI-injective(flat) module, (Pre)cover, (Pre)envelope.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1896