Search results for: Probabilistic methods
3915 A Dual Method for Solving General Convex Quadratic Programs
Authors: Belkacem Brahmi, Mohand Ouamer Bibi
Abstract:
In this paper, we present a new method for solving quadratic programming problems, not strictly convex. Constraints of the problem are linear equalities and inequalities, with bounded variables. The suggested method combines the active-set strategies and support methods. The algorithm of the method and numerical experiments are presented, while comparing our approach with the active set method on randomly generated problems.
Keywords: Convex quadratic programming, dual support methods, active set methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18943914 A Parametric Study of an Inverse Electrostatics Problem (IESP) Using Simulated Annealing, Hooke & Jeeves and Sequential Quadratic Programming in Conjunction with Finite Element and Boundary Element Methods
Authors: Ioannis N. Koukoulis, Clio G. Vossou, Christopher G. Provatidis
Abstract:
The aim of the current work is to present a comparison among three popular optimization methods in the inverse elastostatics problem (IESP) of flaw detection within a solid. In more details, the performance of a simulated annealing, a Hooke & Jeeves and a sequential quadratic programming algorithm was studied in the test case of one circular flaw in a plate solved by both the boundary element (BEM) and the finite element method (FEM). The proposed optimization methods use a cost function that utilizes the displacements of the static response. The methods were ranked according to the required number of iterations to converge and to their ability to locate the global optimum. Hence, a clear impression regarding the performance of the aforementioned algorithms in flaw identification problems was obtained. Furthermore, the coupling of BEM or FEM with these optimization methods was investigated in order to track differences in their performance.
Keywords: Elastostatic, inverse problem, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18743913 Assessment of Slope Stability by Continuum and Discontinuum Methods
Authors: Taleb Hosni Abderrahmane, Berga Abdelmadjid
Abstract:
The development of numerical analysis and its application to geomechanics problems have provided geotechnical engineers with extremely powerful tools. One of the most important problems in geotechnical engineering is the slope stability assessment. It is a very difficult task due to several aspects such the nature of the problem, experimental consideration, monitoring, controlling, and assessment. The main objective of this paper is to perform a comparative numerical study between the following methods: The Limit Equilibrium (LEM), Finite Element (FEM), Limit Analysis (LAM) and Distinct Element (DEM). The comparison is conducted in terms of the safety factors and the critical slip surfaces. Through the results, we see the feasibility to analyse slope stability by many methods.Keywords: Comparison, factor of safety, geomechanics, numerical methods, slope analysis, slip surfaces.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22773912 Semiconvergence of Alternating Iterative Methods for Singular Linear Systems
Authors: Jing Wu
Abstract:
In this paper, we discuss semiconvergence of the alternating iterative methods for solving singular systems. The semiconvergence theories for the alternating methods are established when the coefficient matrix is a singular matrix. Furthermore, the corresponding comparison theorems are obtained.
Keywords: Alternating iterative method, Semiconvergence, Singular matrix.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16553911 Convergence Analysis of Training Two-Hidden-Layer Partially Over-Parameterized ReLU Networks via Gradient Descent
Authors: Zhifeng Kong
Abstract:
Over-parameterized neural networks have attracted a great deal of attention in recent deep learning theory research, as they challenge the classic perspective of over-fitting when the model has excessive parameters and have gained empirical success in various settings. While a number of theoretical works have been presented to demystify properties of such models, the convergence properties of such models are still far from being thoroughly understood. In this work, we study the convergence properties of training two-hidden-layer partially over-parameterized fully connected networks with the Rectified Linear Unit activation via gradient descent. To our knowledge, this is the first theoretical work to understand convergence properties of deep over-parameterized networks without the equally-wide-hidden-layer assumption and other unrealistic assumptions. We provide a probabilistic lower bound of the widths of hidden layers and proved linear convergence rate of gradient descent. We also conducted experiments on synthetic and real-world datasets to validate our theory.Keywords: Over-parameterization, Rectified Linear Units (ReLU), convergence, gradient descent, neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8963910 Syntactic Recognition of Distorted Patterns
Authors: Marek Skomorowski
Abstract:
In syntactic pattern recognition a pattern can be represented by a graph. Given an unknown pattern represented by a graph g, the problem of recognition is to determine if the graph g belongs to a language L(G) generated by a graph grammar G. The so-called IE graphs have been defined in [1] for a description of patterns. The IE graphs are generated by so-called ETPL(k) graph grammars defined in [1]. An efficient, parsing algorithm for ETPL(k) graph grammars for syntactic recognition of patterns represented by IE graphs has been presented in [1]. In practice, structural descriptions may contain pattern distortions, so that the assignment of a graph g, representing an unknown pattern, to a graph language L(G) generated by an ETPL(k) graph grammar G is rejected by the ETPL(k) type parsing. Therefore, there is a need for constructing effective parsing algorithms for recognition of distorted patterns. The purpose of this paper is to present a new approach to syntactic recognition of distorted patterns. To take into account all variations of a distorted pattern under study, a probabilistic description of the pattern is needed. A random IE graph approach is proposed here for such a description ([2]).Keywords: Syntactic pattern recognition, Distorted patterns, Random graphs, Graph grammars.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13953909 A Study of Methods for Preservation of River Banks in order to Reduce Sediments
Authors: Emad Hazbkhah. Abdolreza Zahiri, Hossein Ghorbanizade Kharazi
Abstract:
Reducing river sediments through path correction and preservation of river walls leads to considerable reduction of sedimentation at the pumping stations. Path correction and preservation of walls is not limited to one particular method but, depending on various conditions, a combination of several methods can be employed. In this article, we try to review and evaluate methods for preservation of river banks in order to reduce sediments.Keywords: Erosion , River Banks , Sediments
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16403908 Application of Soft Computing Methods for Economic Dispatch in Power Systems
Authors: Jagabondhu Hazra, Avinash Sinha
Abstract:
Economic dispatch problem is an optimization problem where objective function is highly non linear, non-convex, non-differentiable and may have multiple local minima. Therefore, classical optimization methods may not converge or get trapped to any local minima. This paper presents a comparative study of four different evolutionary algorithms i.e. genetic algorithm, bacteria foraging optimization, ant colony optimization and particle swarm optimization for solving the economic dispatch problem. All the methods are tested on IEEE 30 bus test system. Simulation results are presented to show the comparative performance of these methods.
Keywords: Ant colony optimization, bacteria foraging optimization, economic dispatch, evolutionary algorithm, genetic algorithm, particle swarm optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24813907 Evaluation of Clustering Based on Preprocessing in Gene Expression Data
Authors: Seo Young Kim, Toshimitsu Hamasaki
Abstract:
Microarrays have become the effective, broadly used tools in biological and medical research to address a wide range of problems, including classification of disease subtypes and tumors. Many statistical methods are available for analyzing and systematizing these complex data into meaningful information, and one of the main goals in analyzing gene expression data is the detection of samples or genes with similar expression patterns. In this paper, we express and compare the performance of several clustering methods based on data preprocessing including strategies of normalization or noise clearness. We also evaluate each of these clustering methods with validation measures for both simulated data and real gene expression data. Consequently, clustering methods which are common used in microarray data analysis are affected by normalization and degree of noise and clearness for datasets.
Keywords: Gene expression, clustering, data preprocessing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17403906 The Effect of X-Ray on Plasma and Erythrocyte Concentration of Zn and Cu in Radiology Staff of Tehran Oil Hospital
Authors: L. Nekoozad, M. Salehi Barough, B. Salmasian
Abstract:
Introduction: Some parameters should be considered to investigate the chronic effects of radiation absorption in radiation workers. Trace elements are parameters which small changes in them can cause significant effects on live systems. The role of trace element concentration in human health is significant. These elements play an important role in the developing and functioning of the immune system, cellular respiration, and oxidation processes. Considering the importance and necessity of this issue and few studies, measurements of concentration changes of these elements due to the absorbed dose are important. Purpose: This study aimed to determine the biological effects of occupational dose absorption on plasma and erythrocyte concentration of Zn and Cu in the radiology staff of Tehran Oil Hospital. Material and methods: In this analytical-comparative study, 72 people have entered. 36 people (18 males and 18 females) were selected as radiology staff in the diagnostic and therapeutic departments of Tehran Oil Hospital. And 36 people (18 males and 18 females) were selected as general section staff in the same hospital as a control group. Radiology and control groups’ age and sex were matched. 10 ml of venous blood was taken from all people. An atomic absorption spectrometer was used to obtain zinc and copper plasma concentrations. Levine test was used to compare these results validity. Results: The mean concentrations of copper and zinc were measured as 0.951 and 0.754 mg/L in the plasma phase and 3.2 and 0.401 mg/L in the RBC phase for the radiology group. Copper and zinc average concentrations, respectively 0.976 and 0.813 mg/L in the plasma phase and 2.906 and 0.476 mg/L in the RBC phase, were measured for the control group. These elements Concentrations in the plasma phase were significantly different from that of the control group, but the concentrations in the red blood cell phase did not show a significant difference compared to the control group. In comparison, a separate comparison between men and women in the experimental and control groups showed a significant difference in the values of the elements mentioned. With a significant increase in samples, a better justification than the available statistical results can be extracted. Conclusions: Within this study results, chronic occupational probabilistic absorption destructive effects (even within the permitted range) on blood trace element concentration have been confirmed.
Keywords: Chronic absorption, atomic absorption spectrometry, radiology staff, trace element concentration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7123905 Protein Secondary Structure Prediction
Authors: Manpreet Singh, Parvinder Singh Sandhu, Reet Kamal Kaur
Abstract:
Protein structure determination and prediction has been a focal research subject in the field of bioinformatics due to the importance of protein structure in understanding the biological and chemical activities of organisms. The experimental methods used by biotechnologists to determine the structures of proteins demand sophisticated equipment and time. A host of computational methods are developed to predict the location of secondary structure elements in proteins for complementing or creating insights into experimental results. However, prediction accuracies of these methods rarely exceed 70%.Keywords: Protein, Secondary Structure, Prediction, DNA, RNA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13893904 Review and Classification of the Indicators and Trends Used in Bridge Performance Modeling
Authors: S. Rezaei, Z. Mirzaei, M. Khalighi, J. Bahrami
Abstract:
Bridges, as an essential part of road infrastructures, are affected by various deterioration mechanisms over time due to the changes in their performance. As changes in performance can have many negative impacts on society, it is essential to be able to evaluate and measure the performance of bridges throughout their life. This evaluation includes the development or the choice of the appropriate performance indicators, which, in turn, are measured based on the selection of appropriate models for the existing deterioration mechanism. The purpose of this article is a statistical study of indicators and deterioration mechanisms of bridges in order to discover further research capacities in bridges performance assessment. For this purpose, some of the most common indicators of bridge performance, including reliability, risk, vulnerability, robustness, and resilience, were selected. The researches performed on each index based on the desired deterioration mechanisms and hazards were comprehensively reviewed. In addition, the formulation of the indicators and their relationship with each other were studied. The research conducted on the mentioned indicators were classified from the point of view of deterministic or probabilistic method, the level of study (element level, object level, etc.), and the type of hazard and the deterioration mechanism of interest. For each of the indicators, a number of challenges and recommendations were presented according to the review of previous studies.
Keywords: Bridge, deterioration mechanism, lifecycle, performance indicator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4583903 Reliability Based Investigation on the Choice of Characteristic Soil Properties
Authors: Jann-Eike Saathoff, Kirill Alexander Schmoor, Martin Achmus, Mauricio Terceros
Abstract:
By using partial factors of safety, uncertainties due to the inherent variability of the soil properties and loads are taken into account in the geotechnical design process. According to the reliability index concept in Eurocode-0 in conjunction with Eurocode-7 a minimum safety level of β = 3.8 for reliability class RC2 shall be established. The reliability of the system depends heavily on the choice of the prespecified safety factor and the choice of the characteristic soil properties. The safety factors stated in the standards are mainly based on experience. However, no general accepted method for the calculation of a characteristic value within the current design practice exists. In this study, a laterally loaded monopile is investigated and the influence of the chosen quantile values of the deterministic system, calculated with p-y springs, will be presented. Monopiles are the most common foundation concepts for offshore wind energy converters. Based on the calculations for non-cohesive soils, a recommendation for an appropriate quantile value for the necessary safety level according to the standards for a deterministic design is given.
Keywords: Asymptotic sampling, characteristic value, monopile foundation, probabilistic design, quantile values.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6723902 Research of a Multistep Method Applied to Numerical Solution of Volterra Integro-Differential Equation
Authors: M.Imanova, G.Mehdiyeva, V.Ibrahimov
Abstract:
Solution of some practical problems is reduced to the solution of the integro-differential equations. But for the numerical solution of such equations basically quadrature methods or its combination with multistep or one-step methods are used. The quadrature methods basically is applied to calculation of the integral participating in right hand side of integro-differential equations. As this integral is of Volterra type, it is obvious that at replacement with its integrated sum the upper limit of the sum depends on a current point in which values of the integral are defined. Thus we receive the integrated sum with variable boundary, to work with is hardly. Therefore multistep method with the constant coefficients, which is free from noted lack and gives the way for finding it-s coefficients is present.Keywords: Volterra integro-differential equations, multistepmethods, finite-difference methods, initial value problem
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15043901 Dynamic-Stochastic Influence Diagrams: Integrating Time-Slices IDs and Discrete Event Systems Modeling
Authors: Xin Zhao, Yin-fan Zhu, Wei-ping Wang, Qun Li
Abstract:
The Influence Diagrams (IDs) is a kind of Probabilistic Belief Networks for graphic modeling. The usage of IDs can improve the communication among field experts, modelers, and decision makers, by showing the issue frame discussed from a high-level point of view. This paper enhances the Time-Sliced Influence Diagrams (TSIDs, or called Dynamic IDs) based formalism from a Discrete Event Systems Modeling and Simulation (DES M&S) perspective, for Exploring Analysis (EA) modeling. The enhancements enable a modeler to specify times occurred of endogenous events dynamically with stochastic sampling as model running and to describe the inter- influences among them with variable nodes in a dynamic situation that the existing TSIDs fails to capture. The new class of model is named Dynamic-Stochastic Influence Diagrams (DSIDs). The paper includes a description of the modeling formalism and the hiberarchy simulators implementing its simulation algorithm, and shows a case study to illustrate its enhancements.
Keywords: Time-sliced influence diagrams, discrete event systems, dynamic-stochastic influence diagrams, modeling formalism, simulation algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14323900 A Study on the Interlaminar Shear Strength of Carbon Fiber Reinforced Plastics Depending on the Lamination Methods
Authors: Min Sang Lee, Hee Jae Shin, In Pyo Cha, Sun Ho Ko, Hyun Kyung Yoon, Hong Gun Kim, Lee Ku Kwac
Abstract:
The prepreg process among the CFRP (Carbon Fiber Reinforced Plastic) forming methods is the short term of ‘Pre-impregnation’, which is widely used for aerospace composites that require a high quality property such as a fiber-reinforced woven fabric, in which an epoxy hardening resin is impregnated the reality. However, that this process requires continuous researches and developments for its commercialization because the delamination characteristically develops between the layers when a great weight is loaded from outside to supplement such demerit, three lamination methods among the prepreg lamination methods of CFRP were designed to minimize the delamination between the layers due to external impacts. Further, the newly designed methods and the existing lamination methods were analyzed through a mechanical characteristic test, Interlaminar Shear Strength test. The Interlaminar Shear Strength test result confirmed that the newly proposed three lamination methods, i.e. the Roll, Half and Zigzag laminations, presented more excellent strengths compared to the conventional Ply lamination. The interlaminar shear strength in the roll method with relatively dense fiber distribution was approximately 1.75% higher than that in the existing ply lamination method, and in the half method, it was approximately 0.78% higher.
Keywords: Carbon Fiber Reinforced Plastic (CFRP), Pre-Impregnation, Laminating Method, Interlaminar Shear Strength (ILSS).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29103899 Application of Formal Methods for Designing a Separation Kernel for Embedded Systems
Authors: Kei Kawamorita, Ryouta Kasahara, Yuuki Mochizuki, Kenichiro Noguchi
Abstract:
A separation-kernel-based operating system (OS) has been designed for use in secure embedded systems by applying formal methods to the design of the separation-kernel part. The separation kernel is a small OS kernel that provides an abstract distributed environment on a single CPU. The design of the separation kernel was verified using two formal methods, the B method and the Spin model checker. A newly designed semi-formal method, the extended state transition method, was also applied. An OS comprising the separation-kernel part and additional OS services on top of the separation kernel was prototyped on the Intel IA-32 architecture. Developing and testing of a prototype embedded application, a point-of-sale application, on the prototype OS demonstrated that the proposed architecture and the use of formal methods to design its kernel part are effective for achieving a secure embedded system having a high-assurance separation kernel.
Keywords: B method, embedded systems, extended state transition, formal methods, separation kernel, Spin.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19253898 Forensic Speaker Verification in Noisy Environmental by Enhancing the Speech Signal Using ICA Approach
Authors: Ahmed Kamil Hasan Al-Ali, Bouchra Senadji, Ganesh Naik
Abstract:
We propose a system to real environmental noise and channel mismatch for forensic speaker verification systems. This method is based on suppressing various types of real environmental noise by using independent component analysis (ICA) algorithm. The enhanced speech signal is applied to mel frequency cepstral coefficients (MFCC) or MFCC feature warping to extract the essential characteristics of the speech signal. Channel effects are reduced using an intermediate vector (i-vector) and probabilistic linear discriminant analysis (PLDA) approach for classification. The proposed algorithm is evaluated by using an Australian forensic voice comparison database, combined with car, street and home noises from QUT-NOISE at a signal to noise ratio (SNR) ranging from -10 dB to 10 dB. Experimental results indicate that the MFCC feature warping-ICA achieves a reduction in equal error rate about (48.22%, 44.66%, and 50.07%) over using MFCC feature warping when the test speech signals are corrupted with random sessions of street, car, and home noises at -10 dB SNR.Keywords: Noisy forensic speaker verification, ICA algorithm, MFCC, MFCC feature warping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9903897 Interactive Methods of Design Education as the Principles of Social Implications of Modern Communities
Authors: Pelin Yildiz
Abstract:
The term interactive education indicates the meaning related with multidisciplinary aspects of distance education following contemporary means around a common basis with different functional requirements. The aim of this paper is to reflect the new techniques in education with the new methods and inventions. These methods are better supplied by interactivity. The integration of interactive facilities in the discipline of education with distance learning is not a new concept but in addition the usage of these methods on design issue is newly being adapted to design education. In this paper the general approach of this method and after the analysis of different samples, the advantages and disadvantages of these approaches are being identified. The method of this paper is to evaluate the related samples and then analyzing the main hypothesis. The main focus is to mention the formation processes of this education. Technological developments in education should be filtered around the necessities of the design education and the structure of the system could then be formed or renewed. The conclusion indicates that interactive methods of education in design issue is a meaning capturing not only technical and computational intelligence aspects but also aesthetical and artistic approaches coming together around the same purpose.Keywords: Interactive education, distance learning, designeducation, computational intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16423896 Assessment of EU Competitiveness Factors by Multivariate Methods
Authors: L. Melecký
Abstract:
Measurement of competitiveness between countries or regions is an important topic of many economic analysis and scientific papers. In European Union (EU), there is no mainstream approach of competitiveness evaluation and measuring. There are many opinions and methods of measurement and evaluation of competitiveness between states or regions at national and European level. The methods differ in structure of using the indicators of competitiveness and ways of their processing. The aim of the paper is to analyze main sources of competitive potential of the EU Member States with the help of Factor analysis (FA) and to classify the EU Member States to homogeneous units (clusters) according to the similarity of selected indicators of competitiveness factors by Cluster analysis (CA) in reference years 2000 and 2011. The theoretical part of the paper is devoted to the fundamental bases of competitiveness and the methodology of FA and CA methods. The empirical part of the paper deals with the evaluation of competitiveness factors in the EU Member States and cluster comparison of evaluated countries by cluster analysis.
Keywords: Competitiveness, cluster analysis, EU, factor analysis, multivariate methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20453895 Manifold Analysis by Topologically Constrained Isometric Embedding
Authors: Guy Rosman, Alexander M. Bronstein, Michael M. Bronstein, Ron Kimmel
Abstract:
We present a new algorithm for nonlinear dimensionality reduction that consistently uses global information, and that enables understanding the intrinsic geometry of non-convex manifolds. Compared to methods that consider only local information, our method appears to be more robust to noise. Unlike most methods that incorporate global information, the proposed approach automatically handles non-convexity of the data manifold. We demonstrate the performance of our algorithm and compare it to state-of-the-art methods on synthetic as well as real data.
Keywords: Dimensionality reduction, manifold learning, multidimensional scaling, geodesic distance, boundary detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14543894 A Comparative Analysis of Fuzzy, Neuro-Fuzzy and Fuzzy-GA Based Approaches for Software Reusability Evaluation
Authors: Parvinder Singh Sandhu, Dalwinder Singh Salaria, Hardeep Singh
Abstract:
Software Reusability is primary attribute of software quality. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. In this paper, we have devised the framework of metrics that uses McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component as input attributes and calculated reusability of the software component. Here, comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA approaches is performed to evaluate the reusability of software components and Fuzzy-GA results outperform the other used approaches. The developed reusability model has produced high precision results as expected by the human experts.Keywords: Software Reusability, Software Metrics, Neural Networks, Genetic Algorithm, Fuzzy Logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18163893 Sixth-Order Two-Point Efficient Family of Super-Halley Type Methods
Authors: Ramandeep Behl, S. S. Motsa
Abstract:
The main focus of this manuscript is to provide a highly efficient two-point sixth-order family of super-Halley type methods that do not require any second-order derivative evaluation for obtaining simple roots of nonlinear equations, numerically. Each member of the proposed family requires two evaluations of the given function and two evaluations of the first-order derivative per iteration. By using Mathematica-9 with its high precision compatibility, a variety of concrete numerical experiments and relevant results are extensively treated to confirm t he t heoretical d evelopment. From their basins of attraction, it has been observed that the proposed methods have better stability and robustness as compared to the other sixth-order methods available in the literature.Keywords: Basins of attraction, nonlinear equations, simple roots, Super-Halley.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6983892 Analysis of Web User Identification Methods
Authors: Renáta Iváncsy, Sándor Juhász
Abstract:
Web usage mining has become a popular research area, as a huge amount of data is available online. These data can be used for several purposes, such as web personalization, web structure enhancement, web navigation prediction etc. However, the raw log files are not directly usable; they have to be preprocessed in order to transform them into a suitable format for different data mining tasks. One of the key issues in the preprocessing phase is to identify web users. Identifying users based on web log files is not a straightforward problem, thus various methods have been developed. There are several difficulties that have to be overcome, such as client side caching, changing and shared IP addresses and so on. This paper presents three different methods for identifying web users. Two of them are the most commonly used methods in web log mining systems, whereas the third on is our novel approach that uses a complex cookie-based method to identify web users. Furthermore we also take steps towards identifying the individuals behind the impersonal web users. To demonstrate the efficiency of the new method we developed an implementation called Web Activity Tracking (WAT) system that aims at a more precise distinction of web users based on log data. We present some statistical analysis created by the WAT on real data about the behavior of the Hungarian web users and a comprehensive analysis and comparison of the three methodsKeywords: Data preparation, Tracking individuals, Web useridentification, Web usage mining
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43923891 Using Pattern Search Methods for Minimizing Clustering Problems
Authors: Parvaneh Shabanzadeh, Malik Hj Abu Hassan, Leong Wah June, Maryam Mohagheghtabar
Abstract:
Clustering is one of an interesting data mining topics that can be applied in many fields. Recently, the problem of cluster analysis is formulated as a problem of nonsmooth, nonconvex optimization, and an algorithm for solving the cluster analysis problem based on nonsmooth optimization techniques is developed. This optimization problem has a number of characteristics that make it challenging: it has many local minimum, the optimization variables can be either continuous or categorical, and there are no exact analytical derivatives. In this study we show how to apply a particular class of optimization methods known as pattern search methods to address these challenges. These methods do not explicitly use derivatives, an important feature that has not been addressed in previous studies. Results of numerical experiments are presented which demonstrate the effectiveness of the proposed method.Keywords: Clustering functions, Non-smooth Optimization, Nonconvex Optimization, Pattern Search Method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16403890 A New Method for Rapid DNA Extraction from Artemia (Branchiopoda, Crustacea)
Authors: R. Manaffar, R. Maleki, S. Zare, N. Agh, S. Soltanian, B. Sehatnia, P. Sorgeloos, P. Bossier, G. Van Stappen
Abstract:
Artemia is one of the most conspicuous invertebrates associated with aquaculture. It can be considered as a model organism, offering numerous advantages for comprehensive and multidisciplinary studies using morphologic or molecular methods. Since DNA extraction is an important step of any molecular experiment, a new and a rapid method of DNA extraction from adult Artemia was described in this study. Besides, the efficiency of this technique was compared with two widely used alternative techniques, namely Chelex® 100 resin and SDS-chloroform methods. Data analysis revealed that the new method is the easiest and the most cost effective method among the other methods which allows a quick and efficient extraction of DNA from the adult animal.Keywords: APD, Artemia, DNA extraction, Molecularexperiments
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31923889 Second Order Statistics of Dynamic Response of Structures Using Gamma Distributed Damping Parameters
Authors: B. Chemali, B. Tiliouine
Abstract:
This article presents the main results of a numerical investigation on the uncertainty of dynamic response of structures with statistically correlated random damping Gamma distributed. A computational method based on a Linear Statistical Model (LSM) is implemented to predict second order statistics for the response of a typical industrial building structure. The significance of random damping with correlated parameters and its implications on the sensitivity of structural peak response in the neighborhood of a resonant frequency are discussed in light of considerable ranges of damping uncertainties and correlation coefficients. The results are compared to those generated using Monte Carlo simulation techniques. The numerical results obtained show the importance of damping uncertainty and statistical correlation of damping coefficients when obtaining accurate probabilistic estimates of dynamic response of structures. Furthermore, the effectiveness of the LSM model to efficiently predict uncertainty propagation for structural dynamic problems with correlated damping parameters is demonstrated.Keywords: Correlated random damping, linear statistical model, Monte Carlo simulation, uncertainty of dynamic response.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18353888 Young’s Modulus Variability: Influence on Masonry Vault Behavior
Authors: A. Zanaz, S. Yotte, F. Fouchal, A. Chateauneuf
Abstract:
This paper presents a methodology for probabilistic assessment of bearing capacity and prediction of failure mechanism of masonry vaults at the ultimate state with consideration of the natural variability of Young’s modulus of stones. First, the computation model is explained. The failure mode corresponds to the four-hinge mechanism. Based on this consideration, the study of a vault composed of 16 segments is presented. The Young’s modulus of the segments is considered as random variable defined by a mean value and a coefficient of variation. A relationship linking the vault bearing capacity to the voussoirs modulus variation is proposed. The most probable failure mechanisms, in addition to that observed in the deterministic case, are identified for each variability level as well as their probability of occurrence. The results show that the mechanism observed in the deterministic case has decreasing probability of occurrence in terms of variability, while the number of other mechanisms and their probability of occurrence increases with the coefficient of variation of Young’s modulus. This means that if a significant change in the Young’s modulus of the segments is proven, taking it into account in computations becomes mandatory, both for determining the vault bearing capacity and for predicting its failure mechanism.Keywords: Masonry, mechanism, probability, variability, vault.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20053887 An Integrated Framework for the Realtime Investigation of State Space Exploration
Authors: Jörg Lassig, Stefanie Thiem
Abstract:
The objective of this paper is the introduction to a unified optimization framework for research and education. The OPTILIB framework implements different general purpose algorithms for combinatorial optimization and minimum search on standard continuous test functions. The preferences of this library are the straightforward integration of new optimization algorithms and problems as well as the visualization of the optimization process of different methods exploring the search space exclusively or for the real time visualization of different methods in parallel. Further the usage of several implemented methods is presented on the basis of two use cases, where the focus is especially on the algorithm visualization. First it is demonstrated how different methods can be compared conveniently using OPTILIB on the example of different iterative improvement schemes for the TRAVELING SALESMAN PROBLEM. A second study emphasizes how the framework can be used to find global minima in the continuous domain.Keywords: Global Optimization Heuristics, Particle Swarm Optimization, Ensemble Based Threshold Accepting, Ruin and Recreate
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13833886 Superior Performances of the Neural Network on the Masses Lesions Classification through Morphological Lesion Differences
Authors: U. Bottigli, R.Chiarucci, B. Golosio, G.L. Masala, P. Oliva, S.Stumbo, D.Cascio, F. Fauci, M. Glorioso, M. Iacomi, R. Magro, G. Raso
Abstract:
Purpose of this work is to develop an automatic classification system that could be useful for radiologists in the breast cancer investigation. The software has been designed in the framework of the MAGIC-5 collaboration. In an automatic classification system the suspicious regions with high probability to include a lesion are extracted from the image as regions of interest (ROIs). Each ROI is characterized by some features based generally on morphological lesion differences. A study in the space features representation is made and some classifiers are tested to distinguish the pathological regions from the healthy ones. The results provided in terms of sensitivity and specificity will be presented through the ROC (Receiver Operating Characteristic) curves. In particular the best performances are obtained with the Neural Networks in comparison with the K-Nearest Neighbours and the Support Vector Machine: The Radial Basis Function supply the best results with 0.89 ± 0.01 of area under ROC curve but similar results are obtained with the Probabilistic Neural Network and a Multi Layer Perceptron.
Keywords: Neural Networks, K-Nearest Neighbours, Support Vector Machine, Computer Aided Detection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614