CHAPTER IX Radial Basis Function Networks

Similar documents
Perceptron as a graph

Support Vector Machines

CSE 5526: Introduction to Neural Networks Radial Basis Function (RBF) Networks

Support Vector Machines

COMPUTATIONAL INTELLIGENCE

CHAPTER 6 IMPLEMENTATION OF RADIAL BASIS FUNCTION NEURAL NETWORK FOR STEGANALYSIS

Image Compression: An Artificial Neural Network Approach

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Instance-based Learning

IMPLEMENTATION OF RBF TYPE NETWORKS BY SIGMOIDAL FEEDFORWARD NEURAL NETWORKS

Channel Performance Improvement through FF and RBF Neural Network based Equalization

Neural Network Learning. Today s Lecture. Continuation of Neural Networks. Artificial Neural Networks. Lecture 24: Learning 3. Victor R.

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

A Dendrogram. Bioinformatics (Lec 17)

Radial Basis Function Neural Network Classifier

A Massively-Parallel SIMD Processor for Neural Network and Machine Vision Applications

Dynamic Analysis of Structures Using Neural Networks

3 Nonlinear Regression

Function Approximation Using Artificial Neural Networks

Automatic basis selection for RBF networks using Stein s unbiased risk estimator

RADIAL BASIS FUNCTIONS NETWORK FOR DEFECT SIZING. S. Nair, S. Udpa and L. Udpa Center for Non-Destructive Evaluation Scholl Road Ames, IA 50011

Machine Learning 13. week

Optimization. Industrial AI Lab.

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

Radial Basis Function Networks

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Chap.12 Kernel methods [Book, Chap.7]

2. Neural network basics

Gauss-Sigmoid Neural Network

CHAPTER VI BACK PROPAGATION ALGORITHM

3 Nonlinear Regression

International Journal of Emerging Technologies in Computational and Applied Sciences (IJETCAS)

Kernels + K-Means Introduction to Machine Learning. Matt Gormley Lecture 29 April 25, 2018

Face Detection Using Radial Basis Function Neural Networks With Fixed Spread Value

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Investigating the Effect of Different Kernel Functions on the Performance of SVM for Recognizing Arabic Characters

Bumptrees for Efficient Function, Constraint, and Classification Learning

5 Learning hypothesis classes (16 points)

Accurate modeling of SiGe HBT using artificial neural networks: Performance Comparison of the MLP and RBF Networks

DEVELOPMENT OF NEURAL NETWORK TRAINING METHODOLOGY FOR MODELING NONLINEAR SYSTEMS WITH APPLICATION TO THE PREDICTION OF THE REFRACTIVE INDEX

Data mining with Support Vector Machine

CHAPTER 5 RADIAL BASIS FUNCTION (RBF) NEURAL NETWORKS FOR TOOL WEAR MONITORING

A Learning Algorithm for Piecewise Linear Regression

Width optimization of the Gaussian kernels in Radial Basis Function Networks

Nelder-Mead Enhanced Extreme Learning Machine

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

4. Feedforward neural networks. 4.1 Feedforward neural network structure

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

CS6220: DATA MINING TECHNIQUES

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Neural networks for variable star classification

Learning with Regularization Networks

Basis Functions. Volker Tresp Summer 2017

COMPEL 17,1/2/3. This work was supported by the Greek General Secretariat of Research and Technology through the PENED 94 research project.

Chapter 4. The Classification of Species and Colors of Finished Wooden Parts Using RBFNs

Evolvable Hardware for Generalized Neural Networks

FAST EDDY CURRENT FORWARD MODELS USING ARTIFICIAL NEURAL NETWORKS

Research on Evaluation Method of Product Style Semantics Based on Neural Network

Face Detection Using Radial Basis Function Neural Networks with Fixed Spread Value

Radial Basis Function Networks: Algorithms

In the Name of God. Lecture 17: ANFIS Adaptive Network-Based Fuzzy Inference System

Machine Learning for NLP

Function approximation using RBF network. 10 basis functions and 25 data points.

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

CPSC 340: Machine Learning and Data Mining. More Regularization Fall 2017

Multi-Layer Perceptrons with B-SpIine Receptive Field Functions

Neural Networks Laboratory EE 329 A

Neural Networks and Deep Learning

Autoencoders, denoising autoencoders, and learning deep networks

Neural Network and Adaptive Feature Extraction Technique for Pattern Recognition

Locally Weighted Learning for Control. Alexander Skoglund Machine Learning Course AASS, June 2005

Data Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs)

A Systematic Overview of Data Mining Algorithms

Neuro-Fuzzy Inverse Forward Models

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Machine Learning Classifiers and Boosting

Feature scaling in support vector data description

SVM Classification in -Arrays

ADVANCED MACHINE LEARNING MACHINE LEARNING. Kernel for Clustering kernel K-Means

Designing Radial Basis Neural Networks using a Distributed Architecture

CS 8520: Artificial Intelligence. Machine Learning 2. Paula Matuszek Fall, CSC 8520 Fall Paula Matuszek

Introduction to ANSYS DesignXplorer

Support Vector Machines

Locally Weighted Learning

Artificial Neural Networks MLP, RBF & GMDH

CMPT 882 Week 3 Summary

Machine Learning in Biology

Artificial Neural Networks Lecture Notes Part 5. Stephen Lucci, PhD. Part 5

Understanding Andrew Ng s Machine Learning Course Notes and codes (Matlab version)

Artificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Learning from Data: Adaptive Basis Functions

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Efficient Object Tracking Using K means and Radial Basis Function

Optimization. 1. Optimization. by Prof. Seungchul Lee Industrial AI Lab POSTECH. Table of Contents

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM

Learning from Data Linear Parameter Models

Deep Generative Models Variational Autoencoders

Transcription:

CHAPTER IX Radial Basis Function Networks Radial basis function (RBF) networks are feed-forward networks trained using a supervised training algorithm. They are typically configured with a single hidden layer of units whose activation function is selected from a class of functions called basis functions. While similar to back propagation in many respects, radial basis function networks have several advantages. They usually train much faster than back propagation networks. They are less susceptible to problems with non-stationary inputs because of the behavior of the radial basis function hidden units. Popularized by Moody and Darken (989), RBF networks have proven to be a useful neural network architecture. The major difference between RBF networks and back propagation networks (that is, multi layer perceptron trained by Back Propagation algorithm) is the behavior of the single hidden layer. Rather than using the sigmoidal or S-shaped activation function as in back propagation, the hidden units in RBF networks use a Gaussian or some other basis kernel function. Each hidden unit acts as a locally tuned processor that computes a score for the match between the input vector and its connection weights or centers. In effect, the basis units are highly specialized pattern detectors. The weights connecting the basis units to the outputs are used to take linear combinations of the hidden units to product the final classification or output. In this chapter first the structure of the network will be introduced and it will be explained how it can be used for function approximation and data interpolation. Then it will be explained how it can be trained. EE543 ECTURE NOTES. METU EEE. ANKARA 39

9. The Structure of the RBF Networks Radial Basis Functions are first introduced in the solution of the real multivariable interpolation problems. Broomhead and owe (988), and Moody and Darken (989) were the first to exploit the use of radial basis functions in the design of neural networks. The structure of an RBF networks in its most basic form involves three entirely different layers (Figure 9.). u c w x u 2 c 2 w 2 x 2 u l c i w j x j w u N c x M Fig. 9.. Structure of the Standart RBF network The input layer is made up of source nodes (sensory units) whose number is equal to the dimension p of the input vector u. EE543 ECTURE NOTES. METU EEE. ANKARA 40

9.. Hidden layer The second layer is the hidden layer which is composed of nonlinear units that are connected directly to all of the nodes in the input layer. It is of high enough dimension, which serves a different purpose from that in a multilayer perceptron. Each hidden unit takes its input from all the nodes at the components at the input layer. As mentioned above the hidden units contains a basis function, which has the parameters center and width. The center of the basis function for a node i at the hidden layer is a vector c i whose size is the as the input vector u and there is normally a different center for each unit in the network. First, the radial distance d i, between the input vector u and the center of the basis function c i is computed for each unit i in the hidden layer as d i = u c (9..) i using the Euclidean distance. The output h i of each hidden unit i is then computed by applying the basis function G to this distance h i = G(d i,σ i ) (9..2) As it is shown in Figure 9.2, the basis function is a curve (typically a Gaussian function, the width corresponding to the variance,σ i ) which has a peak at zero distance and it decreases as the distance from the center increases. EE543 ECTURE NOTES. METU EEE. ANKARA 4

Figure 9.2. The response region of an RBF hidden node around its center as a function of the distance from this center. For an input space u R 2, that is M=2, this corresponds to the two dimensional Gaussian centered at c i on the input space, where also c i R 2, as it is shown in Figure 9.3 Figure 9.3 Response of a hidden unit on the input space for u R 2 EE543 ECTURE NOTES. METU EEE. ANKARA 42

9..2 Output layer The transformation from the input space to the hidden unit space is nonlinear, whereas the transformation to the hidden unit space to the output space is linear. The j th output is computed as x j = f j ( u ) = w0 j + wijhi j =,2,.., M (9..3) 9..3 Mathematical model In summary, the mathematical model of the RBF network can be expressed as: x=f(u), f:r N R M x j = f j ( u ) = w0 j + wijg( u ci ) j =,2,.., M (9..5) where is the the Euclidean distance between u and ci 9.2 Function approximation et y=g(u) be a given function of u, y R, u R, g:r R, and let G i.., be a finite set of basis functions. The function g can be written in terms of the given basis functions as y = g( u ) = wi Gi ( u) + r( u) (9.2.) EE543 ECTURE NOTES. METU EEE. ANKARA 43

where r(u) is the residual. The function y can be approximated as y = g( u ) w G ( u) (9.2.2) i i The aim is to minimize the error by setting the parameters of G i appropriately. A possible choice for the error definition is the 2 norm of the residual function r(u) which is defined as 2 r( u) = 2 r( u) (9.2.3) 9.2. Approximation by RBFNN Now, consider the single input single output RBF network shown in Figure 9.4. Then x can be written as x = f( u ) = w i G i ( u ci ) (9.2.4) By the use of such a network, y can be written as y = wi G( u ci ) + r( u) = f( u) + r( u) (9.2.5) where f(u) is the output of the RBFNN given in Figure 9.4 and r(u) is the residual. By setting the center c i, the variance σ i, and the weight w i the error appropriately, the error can be minimized. EE543 ECTURE NOTES. METU EEE. ANKARA 44

c w 0 c 2 w u w 2 x c i w j w c Figure 9.4 Single input, single output RBF network Whatever we discussed here for g:r R, can be generalized to g:r N R M easily by using an N input, M output RBFNN given in figure 9. previously. 9.2.2 Data Interpolation Given input output training patterns (u k,y k ), k=,2,..k, the aim of data interpolation is to approximate the function y from which the data is generated. Since the function y is unknown, the problem can be stated as a minimization problem which takes only the sample points into consideration: Choose wi,j and ci,,2..., j=,2...m so as to minimize K k k J ( w, c) = y f ( u ) (9.2.6) k = 2 EE543 ECTURE NOTES. METU EEE. ANKARA 45

As an example, the output of an RBF network trained to fit the data points given in Table 9. is given in Figure 9.5. TABE I: 3 data points generated by using sum of three gaussians with c =0.2000 c 2 =0.6000 c 3 =0.9000 w =0.2000 w 2 =0.5000 w 3 =0.3000 σ=0.000 data no 2 3 4 5 6 7 9 0 2 3 x 0.0500 0.2000 0.2500 0.3000 0.4000 0.4300 0.4800 0.6000 0.7000 0.8000 0.9000 0.9500 f(x) 0.0863 0.2662 0.2362 0.687 0.260 0.756 0.3290 0.6694 0.4573 0.3320 0.4063 0.3535 Figure 9.5 Output of the RBF network trained to fit the datapoints given in Table 9. 9.3 Training RBF Networks The training of a RBF network can be formulated as a nonlinear unconstrained optimization problem given below: Given input output training patterns (u k,y k ), k=,2,..k, choose wi,j and ci,,,2..., j=,2...m so as to minimize K k k J ( w, c) = y f ( u ) (9.3.) k = 2 EE543 ECTURE NOTES. METU EEE. ANKARA 46

(.3) Note that the training problem becomes quadratic once if ci s (radial basis function centers) are known. 9.3. Adjusting the widths In its simplest form, all hidden units in the RBF network have the same width or degree of sensitivity to inputs. However, in portions of the input space where there are few patterns, it is sometime desirable to have hidden units with a wide area of reception. ikewise, in portions of the input space, which are crowded, it might be desirable to have very highly tuned processors with narrow reception fields. Computing these individual widths increases the performance of the RBF network at the expense of a more complicated training process. 9.3.2 Adjusting the centers Remember that in a back propagation network, all weights in all of the layers are adjusted at the same time. In radial basis function networks, however, the weights into the hidden layer basis units are usually set before the second layer of weights is adjusted. As the input moves away from the connection weights, the activation value falls off. This behavior leads to the use of the term center for the first-layer weights. These center weights can be computed using Kohonen feature maps, statistical methods such as K-Means clustering, or some other means. In any case, they are then used to set the areas of sensitivity for the RBF network s hidden units, which then remain fixed. 9.3.3 Adjusting the weights Once the hidden layer weights are set, a second phase of training is used to adjust the output weights. This process typically uses the standard steepest descent algorithm. Note that the training problem becomes quadratic once if ci s (radial basis function centers) are known. EE543 ECTURE NOTES. METU EEE. ANKARA 47