Supervised Learning in Neural Networks (Part 2)

Similar documents
11/14/2010 Intelligent Systems and Soft Computing 1

Multilayer Feed-forward networks

COMPUTATIONAL INTELLIGENCE

Dr. Qadri Hamarsheh Supervised Learning in Neural Networks (Part 1) learning algorithm Δwkj wkj Theoretically practically

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

Neural Network Learning. Today s Lecture. Continuation of Neural Networks. Artificial Neural Networks. Lecture 24: Learning 3. Victor R.

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Deep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES

Perceptrons and Backpropagation. Fabio Zachert Cognitive Modelling WiSe 2014/15

Data Mining. Neural Networks

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

Neural Networks (Overview) Prof. Richard Zanibbi

2. Neural network basics

Week 3: Perceptron and Multi-layer Perceptron

Neural Network Neurons

For Monday. Read chapter 18, sections Homework:

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

CMPT 882 Week 3 Summary

Pattern Classification Algorithms for Face Recognition

Classification and Regression using Linear Networks, Multilayer Perceptrons and Radial Basis Functions

Character Recognition Using Convolutional Neural Networks

In this assignment, we investigated the use of neural networks for supervised classification

CS6220: DATA MINING TECHNIQUES

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

Opening the Black Box Data Driven Visualizaion of Neural N

CHAPTER VI BACK PROPAGATION ALGORITHM

Neuron Selectivity as a Biologically Plausible Alternative to Backpropagation

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska

Lecture 20: Neural Networks for NLP. Zubin Pahuja

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

Multi Layer Perceptron with Back Propagation. User Manual

Technical University of Munich. Exercise 7: Neural Network Basics

Practical Tips for using Backpropagation

Deep Neural Networks Optimization

Notes on Multilayer, Feedforward Neural Networks

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

International Journal of Advanced Research in Computer Science and Software Engineering

6. Backpropagation training 6.1 Background

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM

Artificial Neural Networks Lecture Notes Part 5. Stephen Lucci, PhD. Part 5

Machine Learning Classifiers and Boosting

OMBP: Optic Modified BackPropagation training algorithm for fast convergence of Feedforward Neural Network

COMPUTATIONAL INTELLIGENCE

11/14/2010 Intelligent Systems and Soft Computing 1

RIMT IET, Mandi Gobindgarh Abstract - In this paper, analysis the speed of sending message in Healthcare standard 7 with the use of back

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Introduction to Neural Networks: Structure and Training

Artificial Neural Networks MLP, RBF & GMDH

A Novel Pruning Algorithm for Optimizing Feedforward Neural Network of Classification Problems

Introduction to Multilayer Perceptrons

CP365 Artificial Intelligence

Artificial Neural Networks

Deep Neural Networks with Flexible Activation Function

Multi-Layered Perceptrons (MLPs)

Ensemble methods in machine learning. Example. Neural networks. Neural networks

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Hybrid PSO-SA algorithm for training a Neural Network for Classification

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Supervised Learning (contd) Linear Separation. Mausam (based on slides by UW-AI faculty)

Introduction AL Neuronale Netzwerke. VL Algorithmisches Lernen, Teil 2b. Norman Hendrich

Machine Learning 13. week

Artificial Intellegence

Thomas Nabelek September 22, ECE 7870 Project 1 Backpropagation

Future Image Prediction using Artificial Neural Networks

Chapter 5 Components for Evolution of Modular Artificial Neural Networks

Practice Exam Sample Solutions

5 Learning hypothesis classes (16 points)

Perceptron: This is convolution!

Introduction to Neural Networks

Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach

Image Compression: An Artificial Neural Network Approach

A neural network that classifies glass either as window or non-window depending on the glass chemistry.

Cost Functions in Machine Learning

Biologically inspired object categorization in cluttered scenes

Khmer Character Recognition using Artificial Neural Network

AN NOVEL NEURAL NETWORK TRAINING BASED ON HYBRID DE AND BP

IMPLEMENTING DEEP LEARNING USING CUDNN 이예하 VUNO INC.

Neural Nets for Adaptive Filter and Adaptive Pattern Recognition

Inversion of Fracture Parameters by Using the Artificial Neural Network

Fast Learning for Big Data Using Dynamic Function

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Exercise: Training Simple MLP by Backpropagation. Using Netlab.

An Empirical Study of Software Metrics in Artificial Neural Networks

Neural Networks. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Experimenting with Neural Nets Numbers marked with indicate material you need to write up.

Automatic Adaptation of Learning Rate for Backpropagation Neural Networks

Additive Manufacturing Defect Detection using Neural Networks

arxiv: v1 [cs.lg] 25 Jan 2018

An Improved Backpropagation Method with Adaptive Learning Rate

Neural Networks. By Laurence Squires

New methodology for calculating flight parameters with neural network EGD method

Back propagation Algorithm:

Knowledge Discovery and Data Mining. Neural Nets. A simple NN as a Mathematical Formula. Notes. Lecture 13 - Neural Nets. Tom Kelsey.

Transcription:

Supervised Learning in Neural Networks (Part 2) Multilayer neural networks (back-propagation training algorithm) The input signals are propagated in a forward direction on a layer-bylayer basis. Learning in a multilayer network proceeds the same way as for a perceptron. A training set of input patterns is presented to the network. The network computes its output pattern, and if there is an error - or in other words a difference between actual and desired output patterns - the weights are adjusted to reduce this error. In a back-propagation neural network, the learning algorithm has two phases. o First, a training input pattern is presented to the network input layer. The network propagates the input pattern from layer to layer until the output pattern is generated by the output layer. o Second, if this pattern is different from the desired output, an error is calculated and then propagated backwards through the network from the output layer to the input layer. The weights are modified as the error is propagated. back-propagation neural network 1

The back-propagation training algorithm Backpropagation is a common method for training a neural network, the goal of backpropagation is to optimize the weights so that the neural network can learn how to correctly map arbitrary inputs to outputs. Backpropagation method contains the following steps: Step 1: Initialization; set all the weights and threshold levels of the network to random numbers uniformly distributed inside a range: Where F i is the total number of inputs of neuron i in the network. Step 2: Activation; activate the back-propagation neural network by applying inputs x 1 (p), x 2 (p),, x n (p) and desired outputs y d,1 (p), y d,2 (p),, y d,n (p) (forward pass). o Calculate the actual outputs of the neurons in the hidden layer: Where n is the number of inputs of neuron j in the hidden layer. o Calculate the actual outputs of the neurons in the output layer: Where m is the number of inputs of neuron k in the output layer Step 3: Weight training (back-propagate) o Update the weights in the back-propagation network propagating backward the errors associated with output neurons. Calculate the error gradient for the neurons in the output layer: Where Calculate the weight corrections: Update the weights at the output neurons: o Calculate the error gradient for the neurons in the hidden layer: 2

Calculate the weight corrections: Update the weights at the hidden neurons: Step 4: Iteration; increase iteration p by one, go back to Step 2 and repeat the process until the selected error criterion is satisfied. A Step by Step Backpropagation Example: Exclusive-OR. The initial weights and threshold levels are set randomly as follows: w 13 = 0.5, w 14 = 0.9, w 23 = 0.4, w 24 = 1.0, w 35 = -1.2, w 45 = 1.1, θ 3 = 0.8, θ 4 = -0.1 and θ 5 = 0.3. We consider a training set where inputs x 1 and x 2 are equal to 1 and desired output y d,5 is 0. The actual outputs of neurons 3 and 4 in the hidden layer are calculated as Now the actual output of neuron 5 in the output layer is determined as: Thus, the following error is obtained: 3

The next step is weight training. To update the weights and threshold levels in our network, we propagate the error, e, from the output layer backward to the input layer. First, we calculate the error gradient for neuron 5 in the output layer: Then we determine the weight corrections assuming that the learning rate parameter, α, is equal to 0.1: Next we calculate the error gradients for neurons 3 and 4 in the hidden layer: We then determine the weight corrections: At last, we update all weights and threshold: The training process is repeated until the sum of squared errors is less than 0.001. 4

Learning curve for operation Exclusive-OR Final results of network learning: Network represented by McCulloch-Pitts model for solving the Exclusive- OR operation. 5

Decision boundaries (a) Decision boundary constructed by hidden neuron 3; (b) Decision boundary constructed by hidden neuron 4; (c) Decision boundaries constructed by complete network Problems with Backpropagation 1) Local Minima : This occurs because the algorithm always changes the weights in such a way as to cause the error to decrease. But the error might briefly have to increase, as shown in figure. If this is the case, the algorithm will gets stuck and the error will not decrease further. o Solutions to this problem: Reset the weights to different random numbers and train again. Add momentum to the weight correction: weight correction depends not just on the current error, but also on previous changes, For example W + =W+Current change+(change on previous iteration * constant) Constant is < 1. 0 β < 1; Typically β = 0. 95 This equation is called the generalized delta rule. Learning with momentum for operation Exclusive-OR: 126 Epochs 2) Biological neurons do not work backward to adjust the synaptic weights, so Backpropagation Algorithm can t emulate brain-like learning. 6

3) In Backpropagation Algorithm, calculations are extensive (training is slow). o Solutions to this problem: Use the sigmoidal activation function (hyperbolic tangent). where a and b are constants. Include a momentum term. Adjust the learning rate parameter during training (not constant). Learning with adaptive learning rate: 103 Epochs. Learning with momentum and adaptive learning rate: 85 Epochs. Many variations on the standard Backpropagation Algorithm have been developed over the years to overcome such problems. 7