Search results for: Neural function.
3166 Comparison between Beta Wavelets Neural Networks, RBF Neural Networks and Polynomial Approximation for 1D, 2DFunctions Approximation
Authors: Wajdi Bellil, Chokri Ben Amar, Adel M. Alimi
Abstract:
This paper proposes a comparison between wavelet neural networks (WNN), RBF neural network and polynomial approximation in term of 1-D and 2-D functions approximation. We present a novel wavelet neural network, based on Beta wavelets, for 1-D and 2-D functions approximation. Our purpose is to approximate an unknown function f: Rn - R from scattered samples (xi; y = f(xi)) i=1....n, where first, we have little a priori knowledge on the unknown function f: it lives in some infinite dimensional smooth function space and second the function approximation process is performed iteratively: each new measure on the function (xi; f(xi)) is used to compute a new estimate f as an approximation of the function f. Simulation results are demonstrated to validate the generalization ability and efficiency of the proposed Beta wavelet network.
Keywords: Beta wavelets networks, RBF neural network, training algorithms, MSE, 1-D, 2D function approximation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19273165 Application of Wavelet Neural Networks in Optimization of Skeletal Buildings under Frequency Constraints
Authors: Mohammad Reza Ghasemi, Amin Ghorbani
Abstract:
The main goal of the present work is to decrease the computational burden for optimum design of steel frames with frequency constraints using a new type of neural networks called Wavelet Neural Network. It is contested to train a suitable neural network for frequency approximation work as the analysis program. The combination of wavelet theory and Neural Networks (NN) has lead to the development of wavelet neural networks. Wavelet neural networks are feed-forward networks using wavelet as activation function. Wavelets are mathematical functions within suitable inner parameters, which help them to approximate arbitrary functions. WNN was used to predict the frequency of the structures. In WNN a RAtional function with Second order Poles (RASP) wavelet was used as a transfer function. It is shown that the convergence speed was faster than other neural networks. Also comparisons of WNN with the embedded Artificial Neural Network (ANN) and with approximate techniques and also with analytical solutions are available in the literature.Keywords: Weight Minimization, Frequency Constraints, Steel Frames, ANN, WNN, RASP Function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17443164 A Combined Neural Network Approach to Soccer Player Prediction
Authors: Wenbin Zhang, Hantian Wu, Jian Tang
Abstract:
An artificial neural network is a mathematical model inspired by biological neural networks. There are several kinds of neural networks and they are widely used in many areas, such as: prediction, detection, and classification. Meanwhile, in day to day life, people always have to make many difficult decisions. For example, the coach of a soccer club has to decide which offensive player to be selected to play in a certain game. This work describes a novel Neural Network using a combination of the General Regression Neural Network and the Probabilistic Neural Networks to help a soccer coach make an informed decision.
Keywords: General Regression Neural Network, Probabilistic Neural Networks, Neural function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37713163 Performance Evaluation of Complex Valued Neural Networks Using Various Error Functions
Authors: Anita S. Gangal, P. K. Kalra, D. S. Chauhan
Abstract:
The backpropagation algorithm in general employs quadratic error function. In fact, most of the problems that involve minimization employ the Quadratic error function. With alternative error functions the performance of the optimization scheme can be improved. The new error functions help in suppressing the ill-effects of the outliers and have shown good performance to noise. In this paper we have tried to evaluate and compare the relative performance of complex valued neural network using different error functions. During first simulation for complex XOR gate it is observed that some error functions like Absolute error, Cauchy error function can replace Quadratic error function. In the second simulation it is observed that for some error functions the performance of the complex valued neural network depends on the architecture of the network whereas with few other error functions convergence speed of the network is independent of architecture of the neural network.Keywords: Complex backpropagation algorithm, complex errorfunctions, complex valued neural network, split activation function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24303162 Complex-Valued Neural Network in Image Recognition: A Study on the Effectiveness of Radial Basis Function
Authors: Anupama Pande, Vishik Goel
Abstract:
A complex valued neural network is a neural network, which consists of complex valued input and/or weights and/or thresholds and/or activation functions. Complex-valued neural networks have been widening the scope of applications not only in electronics and informatics, but also in social systems. One of the most important applications of the complex valued neural network is in image and vision processing. In Neural networks, radial basis functions are often used for interpolation in multidimensional space. A Radial Basis function is a function, which has built into it a distance criterion with respect to a centre. Radial basis functions have often been applied in the area of neural networks where they may be used as a replacement for the sigmoid hidden layer transfer characteristic in multi-layer perceptron. This paper aims to present exhaustive results of using RBF units in a complex-valued neural network model that uses the back-propagation algorithm (called 'Complex-BP') for learning. Our experiments results demonstrate the effectiveness of a Radial basis function in a complex valued neural network in image recognition over a real valued neural network. We have studied and stated various observations like effect of learning rates, ranges of the initial weights randomly selected, error functions used and number of iterations for the convergence of error on a neural network model with RBF units. Some inherent properties of this complex back propagation algorithm are also studied and discussed.
Keywords: Complex valued neural network, Radial BasisFunction, Image recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24163161 Spline Basis Neural Network Algorithm for Numerical Integration
Authors: Lina Yan, Jingjing Di, Ke Wang
Abstract:
A new basis function neural network algorithm is proposed for numerical integration. The main idea is to construct neural network model based on spline basis functions, which is used to approximate the integrand by training neural network weights. The convergence theorem of the neural network algorithm, the theorem for numerical integration and one corollary are presented and proved. The numerical examples, compared with other methods, show that the algorithm is effective and has the characteristics such as high precision and the integrand not required known. Thus, the algorithm presented in this paper can be widely applied in many engineering fields.
Keywords: Numerical integration, Spline basis function, Neural network algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29353160 Implementation of a New Neural Network Function Block to Programmable Logic Controllers Library Function
Authors: Hamid Abdi, Abolfazl Salami, Abolfazl Ahmadi
Abstract:
Programmable logic controllers are the main controllers in the today's industries; they are used for several applications in industrial control systems and there are lots of examples exist from the PLC applications in industries especially in big companies and plants such as refineries, power plants, petrochemical companies, steel companies, and food and production companies. In the PLCs there are some functions in the function library in software that can be used in PLC programs as basic program elements. The aim of this project are introducing and implementing a new function block of a neural network to the function library of PLC. This block can be applied for some control applications or nonlinear functions calculations after it has been trained for these applications. The implemented neural network is a Perceptron neural network with three layers, three input nodes and one output node. The block can be used in manual or automatic mode. In this paper the structure of the implemented function block, the parameters and the training method of the network are presented by considering the especial method of PLC programming and its complexities. Finally the application of the new block is compared with a classic simulated block and the results are presented.Keywords: Programmable Logic Controller, PLC Programming, Neural Networks, Perception Network, Intelligent Control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 38173159 Identification of Nonlinear Systems Using Radial Basis Function Neural Network
Authors: C. Pislaru, A. Shebani
Abstract:
This paper uses the radial basis function neural network (RBFNN) for system identification of nonlinear systems. Five nonlinear systems are used to examine the activity of RBFNN in system modeling of nonlinear systems; the five nonlinear systems are dual tank system, single tank system, DC motor system, and two academic models. The feed forward method is considered in this work for modelling the non-linear dynamic models, where the KMeans clustering algorithm used in this paper to select the centers of radial basis function network, because it is reliable, offers fast convergence and can handle large data sets. The least mean square method is used to adjust the weights to the output layer, and Euclidean distance method used to measure the width of the Gaussian function.
Keywords: System identification, Nonlinear system, Neural networks, RBF neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28683158 Training Radial Basis Function Networks with Differential Evolution
Authors: Bing Yu , Xingshi He
Abstract:
In this paper, Differential Evolution (DE) algorithm, a new promising evolutionary algorithm, is proposed to train Radial Basis Function (RBF) network related to automatic configuration of network architecture. Classification tasks on data sets: Iris, Wine, New-thyroid, and Glass are conducted to measure the performance of neural networks. Compared with a standard RBF training algorithm in Matlab neural network toolbox, DE achieves more rational architecture for RBF networks. The resulting networks hence obtain strong generalization abilities.
Keywords: differential evolution, neural network, Rbf function
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20533157 Function Approximation with Radial Basis Function Neural Networks via FIR Filter
Authors: Kyu Chul Lee, Sung Hyun Yoo, Choon Ki Ahn, Myo Taeg Lim
Abstract:
Recent experimental evidences have shown that because of a fast convergence and a nice accuracy, neural networks training via extended kalman filter (EKF) method is widely applied. However, as to an uncertainty of the system dynamics or modeling error, the performance of the method is unreliable. In order to overcome this problem in this paper, a new finite impulse response (FIR) filter based learning algorithm is proposed to train radial basis function neural networks (RBFN) for nonlinear function approximation. Compared to the EKF training method, the proposed FIR filter training method is more robust to those environmental conditions. Furthermore , the number of centers will be considered since it affects the performance of approximation.
Keywords: Extended kalmin filter (EKF), classification problem, radial basis function networks (RBFN), finite impulse response (FIR)filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24033156 Learning Flexible Neural Networks for Pattern Recognition
Authors: A. Mirzaaghazadeh, H. Motameni, M. Karshenas, H. Nematzadeh
Abstract:
Learning the gradient of neuron's activity function like the weight of links causes a new specification which is flexibility. In flexible neural networks because of supervising and controlling the operation of neurons, all the burden of the learning is not dedicated to the weight of links, therefore in each period of learning of each neuron, in fact the gradient of their activity function, cooperate in order to achieve the goal of learning thus the number of learning will be decreased considerably. Furthermore, learning neurons parameters immunes them against changing in their inputs and factors which cause such changing. Likewise initial selecting of weights, type of activity function, selecting the initial gradient of activity function and selecting a fixed amount which is multiplied by gradient of error to calculate the weight changes and gradient of activity function, has a direct affect in convergence of network for learning.Keywords: Back propagation, Flexible, Gradient, Learning, Neural network, Pattern recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14993155 Use of Radial Basis Function Neural Network for Bearing Pressure Prediction of Strip Footing on Reinforced Granular Bed Overlying Weak Soil
Authors: Srinath Shetty K., Shivashankar R., Rashmi P. Shetty
Abstract:
Earth reinforcing techniques have become useful and economical to solve problems related to difficult grounds and provide satisfactory foundation performance. In this context, this paper uses radial basis function neural network (RBFNN) for predicting the bearing pressure of strip footing on reinforced granular bed overlying weak soil. The inputs for the neural network models included plate width, thickness of granular bed and number of layers of reinforcements, settlement ratio, water content, dry density, cohesion and angle of friction. The results indicated that RBFNN model exhibited more than 84 % prediction accuracy, thereby demonstrating its application in a geotechnical problem.
Keywords: Bearing pressure, granular bed, radial basis function neural network, strip footing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19493154 Binary Mixture of Copper-Cobalt Ions Uptake by Zeolite using Neural Network
Authors: John Kabuba, Antoine Mulaba-Bafubiandi, Kim Battle
Abstract:
In this study a neural network (NN) was proposed to predict the sorption of binary mixture of copper-cobalt ions into clinoptilolite as ion-exchanger. The configuration of the backpropagation neural network giving the smallest mean square error was three-layer NN with tangent sigmoid transfer function at hidden layer with 10 neurons, linear transfer function at output layer and Levenberg-Marquardt backpropagation training algorithm. Experiments have been carried out in the batch reactor to obtain equilibrium data of the individual sorption and the mixture of coppercobalt ions. The obtained modeling results have shown that the used of neural network has better adjusted the equilibrium data of the binary system when compared with the conventional sorption isotherm models.Keywords: Adsorption isotherm, binary system, neural network; sorption
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20473153 Blind Image Deconvolution by Neural Recursive Function Approximation
Authors: Jiann-Ming Wu, Hsiao-Chang Chen, Chun-Chang Wu, Pei-Hsun Hsu
Abstract:
This work explores blind image deconvolution by recursive function approximation based on supervised learning of neural networks, under the assumption that a degraded image is linear convolution of an original source image through a linear shift-invariant (LSI) blurring matrix. Supervised learning of neural networks of radial basis functions (RBF) is employed to construct an embedded recursive function within a blurring image, try to extract non-deterministic component of an original source image, and use them to estimate hyper parameters of a linear image degradation model. Based on the estimated blurring matrix, reconstruction of an original source image from a blurred image is further resolved by an annealed Hopfield neural network. By numerical simulations, the proposed novel method is shown effective for faithful estimation of an unknown blurring matrix and restoration of an original source image.
Keywords: Blind image deconvolution, linear shift-invariant(LSI), linear image degradation model, radial basis functions (rbf), recursive function, annealed Hopfield neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20663152 Complex-Valued Neural Networks for Blind Equalization of Time-Varying Channels
Authors: Rajoo Pandey
Abstract:
Most of the commonly used blind equalization algorithms are based on the minimization of a nonconvex and nonlinear cost function and a neural network gives smaller residual error as compared to a linear structure. The efficacy of complex valued feedforward neural networks for blind equalization of linear and nonlinear communication channels has been confirmed by many studies. In this paper we present two neural network models for blind equalization of time-varying channels, for M-ary QAM and PSK signals. The complex valued activation functions, suitable for these signal constellations in time-varying environment, are introduced and the learning algorithms based on the CMA cost function are derived. The improved performance of the proposed models is confirmed through computer simulations.
Keywords: Blind Equalization, Neural Networks, Constant Modulus Algorithm, Time-varying channels.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18943151 Image Mapping with Cumulative Distribution Function for Quick Convergence of Counter Propagation Neural Networks in Image Compression
Authors: S. Anna Durai, E. Anna Saro
Abstract:
In general the images used for compression are of different types like dark image, high intensity image etc. When these images are compressed using Counter Propagation Neural Network, it takes longer time to converge. The reason for this is that the given image may contain a number of distinct gray levels with narrow difference with their neighborhood pixels. If the gray levels of the pixels in an image and their neighbors are mapped in such a way that the difference in the gray levels of the neighbor with the pixel is minimum, then compression ratio as well as the convergence of the network can be improved. To achieve this, a Cumulative Distribution Function is estimated for the image and it is used to map the image pixels. When the mapped image pixels are used the Counter Propagation Neural Network yield high compression ratio as well as it converges quickly.Keywords: Correlation, Counter Propagation Neural Networks, Cummulative Distribution Function, Image compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16773150 A Model-following Adaptive Controller for Linear/Nonlinear Plantsusing Radial Basis Function Neural Networks
Authors: Yuichi Masukake, Yoshihisa Ishida
Abstract:
In this paper, we proposed a method to design a model-following adaptive controller for linear/nonlinear plants. Radial basis function neural networks (RBF-NNs), which are known for their stable learning capability and fast training, are used to identify linear/nonlinear plants. Simulation results show that the proposed method is effective in controlling both linear and nonlinear plants with disturbance in the plant input.Keywords: Linear/nonlinear plants, neural networks, radial basisfunction networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14873149 Parameter Sensitivity Analysis of Artificial Neural Network for Predicting Water Turbidity
Authors: Chia-Ling Chang, Chung-Sheng Liao
Abstract:
The present study focuses on the discussion over the parameter of Artificial Neural Network (ANN). Sensitivity analysis is applied to assess the effect of the parameters of ANN on the prediction of turbidity of raw water in the water treatment plant. The result shows that transfer function of hidden layer is a critical parameter of ANN. When the transfer function changes, the reliability of prediction of water turbidity is greatly different. Moreover, the estimated water turbidity is less sensitive to training times and learning velocity than the number of neurons in the hidden layer. Therefore, it is important to select an appropriate transfer function and suitable number of neurons in the hidden layer in the process of parameter training and validation.Keywords: Artificial Neural Network (ANN), sensitivity analysis, turbidity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28183148 Improved Stability Criteria for Neural Networks with Two Additive Time-Varying Delays
Authors: Miaomiao Yang, Shouming Zhong
Abstract:
This paper studies the problem of stability criteria for neural networks with two additive time-varying delays.A new Lyapunov-Krasovskii function is constructed and some new delay dependent stability criterias are derived in the terms of linear matrix inequalities(LMI), zero equalities and reciprocally convex approach.The several stability criterion proposed in this paper is simpler and effective. Finally,numerical examples are provided to demonstrate the feasibility and effectiveness of our results.
Keywords: Stability, Neural networks, Linear Matrix Inequalities (LMI) , Lyapunov function, Time-varying delays
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14543147 Designing Intelligent Adaptive Controller for Nonlinear Pendulum Dynamical System
Authors: R. Ghasemi, M. R. Rahimi Khoygani
Abstract:
This paper proposes the designing direct adaptive neural controller to apply for a class of a nonlinear pendulum dynamic system. The radial basis function (RBF) neural adaptive controller is robust in presence of external and internal uncertainties. Both the effectiveness of the controller and robustness against disturbances are importance of this paper. The simulation results show the promising performance of the proposed controller.
Keywords: Adaptive Neural Controller, Nonlinear Dynamical, Neural Network, RBF, Driven Pendulum, Position Control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25963146 Neural Network Learning Based on Chaos
Authors: Truong Quang Dang Khoa, Masahiro Nakagawa
Abstract:
Chaos and fractals are novel fields of physics and mathematics showing up a new way of universe viewpoint and creating many ideas to solve several present problems. In this paper, a novel algorithm based on the chaotic sequence generator with the highest ability to adapt and reach the global optima is proposed. The adaptive ability of proposal algorithm is flexible in 2 steps. The first one is a breadth-first search and the second one is a depth-first search. The proposal algorithm is examined by 2 functions, the Camel function and the Schaffer function. Furthermore, the proposal algorithm is applied to optimize training Multilayer Neural Networks.
Keywords: learning and evolutionary computing, Chaos Optimization Algorithm, Artificial Neural Networks, nonlinear optimization, intelligent computational technologies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17893145 Exponential Stability Analysis for Uncertain Neural Networks with Discrete and Distributed Time-Varying Delays
Authors: Miaomiao Yang, Shouming Zhong
Abstract:
This paper studies the problem of exponential stability analysis for uncertain neural networks with discrete and distributed time-varying delays. Together with a suitable augmented Lyapunov Krasovskii function, zero equalities, reciprocally convex approach and a novel sufficient condition to guarantee the exponential stability of the considered system. The several exponential stability criterion proposed in this paper is simpler and effective. Finally,numerical examples are provided to demonstrate the feasibility and effectiveness of our results.
Keywords: Exponential stability, Uncertain Neural networks, LMI approach, Lyapunov-Krasovskii function, Time-varying.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14523144 Recurrent Radial Basis Function Network for Failure Time Series Prediction
Authors: Ryad Zemouri, Paul Ciprian Patic
Abstract:
An adaptive software reliability prediction model using evolutionary connectionist approach based on Recurrent Radial Basis Function architecture is proposed. Based on the currently available software failure time data, Fuzzy Min-Max algorithm is used to globally optimize the number of the k Gaussian nodes. The corresponding optimized neural network architecture is iteratively and dynamically reconfigured in real-time as new actual failure time data arrives. The performance of our proposed approach has been tested using sixteen real-time software failure data. Numerical results show that our proposed approach is robust across different software projects, and has a better performance with respect to next-steppredictability compared to existing neural network model for failure time prediction.Keywords: Neural network, Prediction error, Recurrent RadialBasis Function Network, Reliability prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18243143 Prediction of Slump in Concrete using Artificial Neural Networks
Authors: V. Agrawal, A. Sharma
Abstract:
High Strength Concrete (HSC) is defined as concrete that meets special combination of performance and uniformity requirements that cannot be achieved routinely using conventional constituents and normal mixing, placing, and curing procedures. It is a highly complex material, which makes modeling its behavior a very difficult task. This paper aimed to show possible applicability of Neural Networks (NN) to predict the slump in High Strength Concrete (HSC). Neural Network models is constructed, trained and tested using the available test data of 349 different concrete mix designs of High Strength Concrete (HSC) gathered from a particular Ready Mix Concrete (RMC) batching plant. The most versatile Neural Network model is selected to predict the slump in concrete. The data used in the Neural Network models are arranged in a format of eight input parameters that cover the Cement, Fly Ash, Sand, Coarse Aggregate (10 mm), Coarse Aggregate (20 mm), Water, Super-Plasticizer and Water/Binder ratio. Furthermore, to test the accuracy for predicting slump in concrete, the final selected model is further used to test the data of 40 different concrete mix designs of High Strength Concrete (HSC) taken from the other batching plant. The results are compared on the basis of error function (or performance function).Keywords: Artificial Neural Networks, Concrete, prediction ofslump, slump in concrete
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36013142 Classification of Initial Stripe Height Patterns using Radial Basis Function Neural Network for Proportional Gain Prediction
Authors: Prasit Wonglersak, Prakarnkiat Youngkong, Ittipon Cheowanish
Abstract:
This paper aims to improve a fine lapping process of hard disk drive (HDD) lapping machines by removing materials from each slider together with controlling the strip height (SH) variation to minimum value. The standard deviation is the key parameter to evaluate the strip height variation, hence it is minimized. In this paper, a design of experiment (DOE) with factorial analysis by twoway analysis of variance (ANOVA) is adopted to obtain a statistically information. The statistics results reveal that initial stripe height patterns affect the final SH variation. Therefore, initial SH classification using a radial basis function neural network is implemented to achieve the proportional gain prediction.Keywords: Stripe height variation, Two-way analysis ofvariance (ANOVA), Radial basis function neural network, Proportional gain prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16513141 Novel Approach for Promoting the Generalization Ability of Neural Networks
Authors: Naiqin Feng, Fang Wang, Yuhui Qiu
Abstract:
A new approach to promote the generalization ability of neural networks is presented. It is based on the point of view of fuzzy theory. This approach is implemented through shrinking or magnifying the input vector, thereby reducing the difference between training set and testing set. It is called “shrinking-magnifying approach" (SMA). At the same time, a new algorithm; α-algorithm is presented to find out the appropriate shrinking-magnifying-factor (SMF) α and obtain better generalization ability of neural networks. Quite a few simulation experiments serve to study the effect of SMA and α-algorithm. The experiment results are discussed in detail, and the function principle of SMA is analyzed in theory. The results of experiments and analyses show that the new approach is not only simpler and easier, but also is very effective to many neural networks and many classification problems. In our experiments, the proportions promoting the generalization ability of neural networks have even reached 90%.Keywords: Fuzzy theory, generalization, misclassification rate, neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15383140 Low Resolution Single Neural Network Based Face Recognition
Authors: Jahan Zeb, Muhammad Younus Javed, Usman Qayyum
Abstract:
This research paper deals with the implementation of face recognition using neural network (recognition classifier) on low-resolution images. The proposed system contains two parts, preprocessing and face classification. The preprocessing part converts original images into blurry image using average filter and equalizes the histogram of those image (lighting normalization). The bi-cubic interpolation function is applied onto equalized image to get resized image. The resized image is actually low-resolution image providing faster processing for training and testing. The preprocessed image becomes the input to neural network classifier, which uses back-propagation algorithm to recognize the familiar faces. The crux of proposed algorithm is its beauty to use single neural network as classifier, which produces straightforward approach towards face recognition. The single neural network consists of three layers with Log sigmoid, Hyperbolic tangent sigmoid and Linear transfer function respectively. The training function, which is incorporated in our work, is Gradient descent with momentum (adaptive learning rate) back propagation. The proposed algorithm was trained on ORL (Olivetti Research Laboratory) database with 5 training images. The empirical results provide the accuracy of 94.50%, 93.00% and 90.25% for 20, 30 and 40 subjects respectively, with time delay of 0.0934 sec per image.Keywords: Average filtering, Bicubic Interpolation, Neurons, vectorization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17553139 Affine Radial Basis Function Neural Networks for the Robust Control of Hyperbolic Distributed Parameter Systems
Authors: Eleni Aggelogiannaki, Haralambos Sarimveis
Abstract:
In this work, a radial basis function (RBF) neural network is developed for the identification of hyperbolic distributed parameter systems (DPSs). This empirical model is based only on process input-output data and used for the estimation of the controlled variables at specific locations, without the need of online solution of partial differential equations (PDEs). The nonlinear model that is obtained is suitably transformed to a nonlinear state space formulation that also takes into account the model mismatch. A stable robust control law is implemented for the attenuation of external disturbances. The proposed identification and control methodology is applied on a long duct, a common component of thermal systems, for a flow based control of temperature distribution. The closed loop performance is significantly improved in comparison to existing control methodologies.
Keywords: Hyperbolic Distributed Parameter Systems, Radial Basis Function Neural Networks, H∞ control, Thermal systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14243138 Facial Emotion Recognition with Convolutional Neural Network Based Architecture
Authors: Koray U. Erbas
Abstract:
Neural networks are appealing for many applications since they are able to learn complex non-linear relationships between input and output data. As the number of neurons and layers in a neural network increase, it is possible to represent more complex relationships with automatically extracted features. Nowadays Deep Neural Networks (DNNs) are widely used in Computer Vision problems such as; classification, object detection, segmentation image editing etc. In this work, Facial Emotion Recognition task is performed by proposed Convolutional Neural Network (CNN)-based DNN architecture using FER2013 Dataset. Moreover, the effects of different hyperparameters (activation function, kernel size, initializer, batch size and network size) are investigated and ablation study results for Pooling Layer, Dropout and Batch Normalization are presented.
Keywords: Convolutional Neural Network, Deep Learning, Deep Learning Based FER, Facial Emotion Recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13773137 Distance Transmission Line Protection Based on Radial Basis Function Neural Network
Authors: Anant Oonsivilai, Sanom Saichoomdee
Abstract:
To determine the presence and location of faults in a transmission by the adaptation of protective distance relay based on the measurement of fixed settings as line impedance is achieved by several different techniques. Moreover, a fast, accurate and robust technique for real-time purposes is required for the modern power systems. The appliance of radial basis function neural network in transmission line protection is demonstrated in this paper. The method applies the power system via voltage and current signals to learn the hidden relationship presented in the input patterns. It is experiential that the proposed technique is competent to identify the particular fault direction more speedily. System simulations studied show that the proposed approach is able to distinguish the direction of a fault on a transmission line swiftly and correctly, therefore suitable for the real-time purposes.
Keywords: radial basis function neural network, transmission lines protection, relaying, power system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2369