COMPUTATIONAL INTELLIGENCE

Similar documents
CHAPTER 6 IMPLEMENTATION OF RADIAL BASIS FUNCTION NEURAL NETWORK FOR STEGANALYSIS

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Support Vector Machines

Radial Basis Function Neural Network Classifier

Radial Basis Function Networks: Algorithms

Supervised Learning in Neural Networks (Part 2)

Function approximation using RBF network. 10 basis functions and 25 data points.

CSE 5526: Introduction to Neural Networks Radial Basis Function (RBF) Networks

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes

Radial Basis Function Networks

Neural Networks and Deep Learning

CHAPTER IX Radial Basis Function Networks

CS6220: DATA MINING TECHNIQUES

Face Detection Using Radial Basis Function Neural Networks with Fixed Spread Value

Support Vector Machines

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

Artificial Neural Networks MLP, RBF & GMDH

Multilayer Feed-forward networks

COMPUTATIONAL INTELLIGENCE

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Learning to Learn: additional notes

Neural Network Neurons

IN recent years, neural networks have attracted considerable attention

COMPUTATIONAL INTELLIGENCE (INTRODUCTION TO MACHINE LEARNING) SS18. Lecture 2: Linear Regression Gradient Descent Non-linear basis functions

Locally Weighted Learning for Control. Alexander Skoglund Machine Learning Course AASS, June 2005

Face Detection Using Radial Basis Function Neural Networks With Fixed Spread Value

Some fast and compact neural network solutions for artificial intelligence applications

Backpropagation + Deep Learning

Radial Basis Function (RBF) Neural Networks Based on the Triple Modular Redundancy Technology (TMR)

Dynamic Analysis of Structures Using Neural Networks

ImageNet Classification with Deep Convolutional Neural Networks

A Dendrogram. Bioinformatics (Lec 17)

Experimental Analysis of GTM

Supervised vs unsupervised clustering

Neural Network Weight Selection Using Genetic Algorithms

3 Nonlinear Regression

MACHINE LEARNING: CLUSTERING, AND CLASSIFICATION. Steve Tjoa June 25, 2014

Learning from Data: Adaptive Basis Functions

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Machine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

3 Nonlinear Regression

Neural Bag-of-Features Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

SE 263 R. Venkatesh Babu. Object Tracking. R. Venkatesh Babu

Efficient Object Tracking Using K means and Radial Basis Function

Generative and discriminative classification techniques

COMPUTATIONAL INTELLIGENCE (CS) (INTRODUCTION TO MACHINE LEARNING) SS16. Lecture 2: Linear Regression Gradient Descent Non-linear basis functions

Unsupervised Learning

Learning via Optimization

Hybrid Training Algorithm for RBF Network

Unsupervised Learning: Clustering

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2018

Machine Learning. MGS Lecture 3: Deep Learning

Ensemble methods in machine learning. Example. Neural networks. Neural networks

CMPT 882 Week 3 Summary

Review Article Using Radial Basis Function Networks for Function Approximation and Classification

SUPPORT VECTOR MACHINES

Classification and Regression using Linear Networks, Multilayer Perceptrons and Radial Basis Functions

CHAPTER 5 RADIAL BASIS FUNCTION (RBF) NEURAL NETWORKS FOR TOOL WEAR MONITORING

Lab 8 CSC 5930/ Computer Vision

Image Compression: An Artificial Neural Network Approach

Mahalanobis clustering, with applications to AVO classification and seismic reservoir parameter estimation

Opening the Black Box Data Driven Visualizaion of Neural N

Bioinformatics - Lecture 07

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,

Learning from Data Linear Parameter Models

What is machine learning?

SUPERVISED LEARNING METHODS. Stanley Liang, PhD Candidate, Lassonde School of Engineering, York University Helix Science Engagement Programs 2018

Pattern Classification Algorithms for Face Recognition

Week 3: Perceptron and Multi-layer Perceptron

Report: Privacy-Preserving Classification on Deep Neural Network

Deep Learning. Deep Learning. Practical Application Automatically Adding Sounds To Silent Movies

Chap.12 Kernel methods [Book, Chap.7]

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Neural Networks in Statistica

Machine Learning in Biology

ECE 285 Final Project

Neural Networks: What can a network represent. Deep Learning, Fall 2018

Please write your initials at the top right of each page (e.g., write JS if you are Jonathan Shewchuk). Finish this by the end of your 3 hours.

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

Neural Networks: What can a network represent. Deep Learning, Spring 2018

arxiv: v1 [cs.lg] 25 Jan 2018

ECG782: Multidimensional Digital Signal Processing

Neural Networks for unsupervised learning From Principal Components Analysis to Autoencoders to semantic hashing

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM

Accurate modeling of SiGe HBT using artificial neural networks: Performance Comparison of the MLP and RBF Networks

Machine Learning Classifiers and Boosting

CHAPTER 5 NEURAL NETWORK BASED CLASSIFICATION OF ELECTROGASTROGRAM SIGNALS

Image Warping. Srikumar Ramalingam School of Computing University of Utah. [Slides borrowed from Ross Whitaker] 1

Data Preprocessing. Javier Béjar AMLT /2017 CS - MAI. (CS - MAI) Data Preprocessing AMLT / / 71 BY: $\

Lecture on Modeling Tools for Clustering & Regression

Data Mining and Analytics

Data Preprocessing. Javier Béjar. URL - Spring 2018 CS - MAI 1/78 BY: $\

Lecture 10 CNNs on Graphs

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Hierarchical Learning Algorithm for the Beta Basis Function Neural Network

Transcription:

COMPUTATIONAL INTELLIGENCE Radial Basis Function Networks Adrian Horzyk

Preface Radial Basis Function Networks (RBFN) are a kind of artificial neural networks that use radial basis functions (RBF) as activation functions. Typical RBF networks have three layers: 1. An input layer that forwards input signals/stimuli/data, 2. A hidden layer with a selected non-linear RBF activation function, 3. An output layer with linear activation function, however, some modifications of this architecture are also possible, e.g. the output layer can include neurons with non-linear (e.g.) sigmoidal activation functions or there can be used a few other non-linear layers (in this case we achieve a deep ANN architecture).

Separation of Training Samples In case of classification or clusterization (grouping), training samples have to be separated into groups which define different classes. We can distinguish a few ways how we can do it: Non-linear separation Radial overlapping Rectangular overlapping Irregular overlapping

Separation of Training Samples In case of classification or clusterization (grouping), training samples have to be separated into groups which define different classes. We can distinguish a few ways how we can do it: Linear separation Radial overlapping Rectangular overlapping Irregular overlapping

Separation of Training Samples In case of classification or clusterization (grouping), training samples have to be separated into groups which define different classes. We can distinguish a few ways how we can do it: Linear separation Non-linear separation Rectangular overlapping Irregular overlapping

Separation of Training Samples In case of classification or clusterization (grouping), training samples have to be separated into groups which define different classes. We can distinguish a few ways how we can do it: Linear separation Non-linear separation Radial overlapping Irregular overlapping

Separation of Training Samples In case of classification or clusterization (grouping), training samples have to be separated into groups which define different classes. We can distinguish a few ways how we can do it: Linear separation Non-linear separation Radial overlapping Rectangular overlapping

Separation of Training Samples In case of classification or clusterization (grouping), training samples have to be separated into groups which define different classes. We can distinguish a few ways how we can do it:

Radial Separation of RBF Networks Using RBF Networks we need to answer a few question: How many circles will be appropriate? Which circles are the best? How to compute centers? How to achieve the best generalization property? Is the set of circles adequate? Can we determine them better? Should circle overlap? We will seek to identify the least number of circles which cover elements of each class and simultaneously well generalize them.

Basic RBF Network The input can be modeled as a vector (or a matrix) of real numbers x R n. The outputs of the network is then a function of this input vector φ: R n R m : where y = φ(x) = M i=1 w i ρ x c i M is the number of neurons in the hidden layer that should be much less than the number of training samples N. c i is the center vector for neuron i w i is the weight of neuron i in the linear output neuron y is the output vector of computed results (approximation) ρ: R R is an RBF function that depends on the distance of input vector from a center vector represented by the given neuron, usually taken to be Gaussian: ρ x c i = e β i x c i 2

Basic RBF Network In the basic form of this network, all inputs are connected to each hidden neuron. The norm is typically taken to be: N Euclidean distance: x c i 2 = n=1 x n c 2 in Mahalanobis distance: x c i 2 = x c i T S 1 x c i N Manhattan distance: x c i 1 = n=1 x n c in where S is the covariance matrix that consists of elements, which are defined as covariance between the i-th and the j-th elements of a random vector. The parameters w i, c i, and β i are determined during the training process in a manner that optimizes the fit between φ and the data.

Radial Basis Functions The most often used radial basis functions: Gaussian: ρ x c i = e β i x c 2 i Multi-Quadric: ρ x c i = 1 + β i x c 2 i Inverse Quadratic: ρ x c i = Inverse Multi-Quadric : ρ x c i = 1 1+β i x c 2 i 1 1+β i x c 2 i Polyharmonic Spline: ρ x c i = x c k i for k = 1, 3, 5, 7 ln( x c i ) x c k i for k = 2, 4, 6, 8

Normalized RBN Network We often use a normalized architecture that usually achieves better results. In this case, the outputs is computed using the following function: Where y = φ x = i=1 I w i ρ x c i I = ρ x c i i=1 is called a normalized radial basis function. I i=1 ω x c i = ρ x c i I ρ x c i i=1 w i ω x c i CONCLUSIONS: RBF networks are universal approximators that can be used to various regression or classification tasks. An RBF networks with enough hidden neurons can approximate any continuous function with arbitrary precision.

RBF Network Adaptation and Training The adaptation and training process for RBF networks is usually divided into two steps: 1. Unsupervised step: The center vectors c i of the RBF functions in the hidden layer are chosen (e.g. randomly sampled from the set of training samples, obtained by Orthogonal Least Square Learning, or determined using k-means clustering) and their width determined. The choice of the centers are not obvious. The width are usually fixed to the same value which is proportional to the maximum distance between chosen centers. 2. Supervised step: The backpropagation or gradient descent algorithm is used to fine-tune weights of the output layer neurons. Thus, the weights are adjusted at each time step by moving them in a direction opposite from the gradient of the function.

Bibliography and References