Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Similar documents
Data Mining. Neural Networks

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Perceptrons and Backpropagation. Fabio Zachert Cognitive Modelling WiSe 2014/15

Supervised Learning in Neural Networks (Part 2)

Supervised Learning with Neural Networks. We now look at how an agent might learn to solve a general problem by seeing examples.

For Monday. Read chapter 18, sections Homework:

Dr. Qadri Hamarsheh Supervised Learning in Neural Networks (Part 1) learning algorithm Δwkj wkj Theoretically practically

Neural Network Learning. Today s Lecture. Continuation of Neural Networks. Artificial Neural Networks. Lecture 24: Learning 3. Victor R.

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

Artificial Neural Networks

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Neural Networks CMSC475/675

11/14/2010 Intelligent Systems and Soft Computing 1

Machine Learning Classifiers and Boosting

CMPT 882 Week 3 Summary

Multilayer Feed-forward networks

Supervised Learning (contd) Linear Separation. Mausam (based on slides by UW-AI faculty)

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

Neural Networks. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Data Mining and Analytics

Logical Rhythm - Class 3. August 27, 2018

Pattern Classification Algorithms for Face Recognition

Neural Network Neurons

Machine Learning in Biology

CP365 Artificial Intelligence

Character Recognition Using Convolutional Neural Networks

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska

Multi-Layered Perceptrons (MLPs)

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

In this assignment, we investigated the use of neural networks for supervised classification

CS6220: DATA MINING TECHNIQUES

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

Introduction AL Neuronale Netzwerke. VL Algorithmisches Lernen, Teil 2b. Norman Hendrich

Artificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu

COMPUTATIONAL INTELLIGENCE

A Systematic Overview of Data Mining Algorithms. Sargur Srihari University at Buffalo The State University of New York

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Multi Layer Perceptron with Back Propagation. User Manual

Introduction to Neural Networks

Week 3: Perceptron and Multi-layer Perceptron

A Systematic Overview of Data Mining Algorithms

Simple Model Selection Cross Validation Regularization Neural Networks

Artificial Neural Networks Lecture Notes Part 5. Stephen Lucci, PhD. Part 5

Ensemble methods in machine learning. Example. Neural networks. Neural networks

Instructor: Jessica Wu Harvey Mudd College

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Neural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /10/2017

Artificial Neural Networks MLP, RBF & GMDH

11/14/2010 Intelligent Systems and Soft Computing 1

Neural Networks (Overview) Prof. Richard Zanibbi

Neural Networks and Deep Learning

Introduction to Machine Learning

Image Compression: An Artificial Neural Network Approach

CS 8520: Artificial Intelligence

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Support Vector Machines

PARALLELIZED IMPLEMENTATION OF LOGISTIC REGRESSION USING MPI

Optimization. Industrial AI Lab.

CS 8520: Artificial Intelligence. Machine Learning 2. Paula Matuszek Fall, CSC 8520 Fall Paula Matuszek

Lecture 20: Neural Networks for NLP. Zubin Pahuja

ECG782: Multidimensional Digital Signal Processing

Neural Networks (pp )

Perceptron: This is convolution!

Neural Nets. CSCI 5582, Fall 2007

Machine Learning Week 3

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Support Vector Machines

Hybrid PSO-SA algorithm for training a Neural Network for Classification

Notes on Multilayer, Feedforward Neural Networks

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Introduction to Support Vector Machines

Optimization. 1. Optimization. by Prof. Seungchul Lee Industrial AI Lab POSTECH. Table of Contents

Machine Learning 13. week

Fall 09, Homework 5

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Motivation. Problem: With our linear methods, we can train the weights but not the basis functions: Activator Trainable weight. Fixed basis function

Alex Waibel

Lecture #11: The Perceptron

5 Learning hypothesis classes (16 points)

Yuki Osada Andrew Cannon

Technical University of Munich. Exercise 7: Neural Network Basics

The exam is closed book, closed notes except your one-page cheat sheet.

CS570: Introduction to Data Mining

Deep Reinforcement Learning

Optimization Methods for Machine Learning (OMML)

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

Deep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES

6. Backpropagation training 6.1 Background

Climate Precipitation Prediction by Neural Network

Model learning for robot control: a survey

Exercise: Training Simple MLP by Backpropagation. Using Netlab.

Data Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs)

arxiv: v1 [cs.lg] 25 Jan 2018

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Reservoir Computing for Neural Networks

NEURAL NETWORKS. Typeset by FoilTEX 1

CS 4510/9010 Applied Machine Learning

Transcription:

Learning Learning agents Inductive learning Different Learning Scenarios Evaluation Slides based on Slides by Russell/Norvig, Ronald Williams, and Torsten Reil Material from Russell & Norvig, chapters 18.1, 18., 0.5 and 1 Perceptrons Multilayer Perceptrons Reinforcement Learning Temporal Differences Q-Learning SARSA 1

Learning Learning is essential for unknown environments, Learning is useful as a system construction method, i.e., when designer lacks omniscience i.e., expose the agent to reality rather than trying to write it down Learning modifies the agent's decision mechanisms to improve performance

Learning Agents 3

Learning Element Design of a learning element is affected by Which components of the performance element are to be learned What feedback is available to learn these components What representation is used for the components Type of feedback: Supervised learning: correct answers for each example Unsupervised learning: correct answers not given Reinforcement learning: occasional rewards for good actions 4

Different Learning Scenarios Supervised Learning A teacher provides the value for the target function for all training examples (labeled examples) concept learning, classification, regression Semi-supervised Learning Reinforcement Learning The teacher only provides feedback but not example values Only a subset of the training examples are labeled Unsupervised Learning There is no information except the training examples clustering, subgroup discovery, association rule discovery 5

Inductive Learning Simplest form: learn a function from examples f is the target function An example is a pair (x, f(x)) Problem: find a hypothesis h given a training set of examples such that h f on all examples i.e. the hypothesis must generalize from the training examples This is a highly simplified model of real learning: Ignores prior knowledge Assumes examples are given 6

Inductive Learning Method Construct/adjust h to agree with f on training set h is consistent if it agrees with f on all examples Example: curve fitting Ockham's Razor The best explanation is the simplest explanation that fits the data Overfitting Avoidance maximize a combination of consistency and simplicity 7

Performance Measurement How do we know that h f? Use theorems of computational/statistical learning theory Or try h on a new test set of examples where f is known (use same distribution over example space as training set) Learning curve = % correct on test set over training set size 8

What are? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning Very simple principles Very complex behaviours Applications As powerful problem solvers As biological models 9

Pigeons as Art Experts Famous experiment (Watanabe et al. 1995, 001) Pigeon in Skinner box Present paintings of two different artists (e.g. Chagall / Van Gogh) Reward for pecking when presented a particular artist 10

Results Pigeons were able to discriminate between Van Gogh and Chagall with 95% accuracy when presented with pictures they had been trained on Discrimination still 85% successful for previously unseen paintings of the artists Pigeons do not simply memorise the pictures They can extract and recognise patterns (the style ) They generalise from the already seen to make predictions This is what neural networks (biological and artificial) are good at (unlike conventional computer) 11

A Biological Neuron Neurons are connected to each other via synapses If a neuron is activated, it spreads its activation to all connected neurons 1

An Artificial Neuron (McCulloch-Pitts,1943) n a i =g in i =g W ji a j j=0 Neurons correspond to nodes or units A link from unit i to unit j propagates activation aj from j to i The weight Wi,j of the link determines the strength and sign of the connection The total input activation is the sum of the input activations The output activation is determined by the activiation function g 13

Perceptron A single node connecting n input signals with one output signal typically signals are 1 or +1 Activation function (Rosenblatt 1957, 1960) A simple threshold function: ai = { n 1 if W ij a j 0 i=0 n 1 if W ij a j 0 i=0 Thus it implements a linear separator i.e., a hyperplane that divides n-dimensional space into a region with output 1 and a region with output 1 14

Perceptrons and Boolean Fucntions a Perceptron can implement all elementary logical functions W 0= 0.5 (McCullogh & Pitts, 1943) W 1 = 1 more complex functions like XOR cannot be modeled (Minsky & Papert, 1969) no linear separation possible 1 1 1 1 1 1 15

Perceptron Learning Perceptron Learning Rule for Supervised Learning W i W i f x h x xi learning rate Example: error Computation of output signal h(x) in x = 1 0. 1 0.5 1 0.8=1.1-1 1 1 h x =1 because in x 0 0. 0.5 0.8 1 Assume target value f (x) = 1 (and α = 0.5) W 0 0. 0.5 1 1 1=0. 1=1. W 1 0.5 0.5 1 1 1=0.5 1= 0.5 W 0.8 0.5 1 1 1=0.8 1= 0. 16

Measuring the Error of a Network The error for one training example x can be measured by the squared error the squared difference of the output value h(x) and the desired target value f (x) n 1 1 1 E x = Err = f x h x = f x g W j x j i=0 For evaluating the performance of a network, we can try the network on a set of datapoints and average the value (= sum of squared errors) N E Network = i=1 E x i 17

Error Landscape The error function for one training example may be considered as a function in a multi-dimensional weight space E W = n 1 f x g W j x j i =0 The best weight setting for one example is where the error measure for this example is minimal 18

Error Minimization via Gradient Descent In order to find the point with the minimal error: go downhill in the direction where it is steepest E W = n 1 f x g W j x j i =0 but make small steps, or you might step over the target 19

Error Minimization It is easy to derive a perceptron training algorithm that minimizes the squared error n 1 1 1 E = Err = f x h x = f x g W j x j i=0 Change weights into the direction of the steepest descent of the error function n E Err =Err =Err f x g W j x j = Err g ' in x j W j W j W j i=0 To compute this, we need a continuous and differentiable activation function g! Weight update with learning rate α: W j =W j Err g ' in x j positive error increase network output increase weights of nodes with positive input decrease weights of nodes with negative input 0

Threshold Activation Function The regular threshold activation function is problematic g'(x) = 0, therefore E = Err g ' in x j =0 W j no weight changes! { g x = 0 if x 0 1 if x 0 g ' x =0 1

Sigmoid Activation Function A commonly used activation function is the sigmoid function similar to the threshold function easy to differentiate non-linear { g x = 0 if x 0 1 if x 0 g x = g ' x =g x 1 g x g ' x =0 1 1 e x

Multilayer Perceptrons Perceptrons may have multiple output nodes The output nodes may be combined with another perceptron may be viewed as multiple parallel perceptrons which may also have multiple output nodes The size of this hidden layer is determined manually 3

Multilayer Perceptrons Information flow is unidirectional Data is presented to Input layer Passed on to Hidden Layer Passed on to Output layer Information is distributed Information processing is parallel 4

Expressiveness of MLPs Every continuous function can be modeled with three layers i.e., with one hidden layer Every function can be modeled with four layers i.e., with two hidden layers 5

Backpropagation Learning The output nodes are trained like a normal perceptron W ji W ji Err i g ' in i x j =W ji i x j Δi is the error term of output node i times the derivation of its inputs the error term Δi of the output layers is propagated back to the hidden layer j = W ji i g ' in j i W kj =W kj j x k the training signal of hidden layer node j is the weighted sum of the errors of the output nodes 6

Minimizing the Network Error The error landscape for the entire network may be thought of as the sum of the error functions of all examples Minimizing the error for one training example may destroy what has been learned for other examples a good location in weight space for one example may be a bad location for another examples Training procedure: will yield many local minima hard to find global minimum try all examples in turn make small adjustments for each example repeat until convergence One Epoch = One iteration through all examples 7

Overfitting Training Set Error continues to decrease with increasing number of training examples / number of epochs an epoch is a complete pass through all training examples Test Set Error will start to increase because of overfitting Simple training protocol: keep a separate validation set to watch the performance validation set is different from training and test sets! stop training if error on validation set gets down 8

Wide Variety of Applications Speech Recognition Autonomous Driving Handwritten Digit Recognition Credit Approval Backgammon etc. Good for problems where the final output depends on combinations of many input features rule learning is better when only a few features are relevant Bad if explicit representations of the learned concept are needed takes some effort to interpret the concepts that form in the hidden layers 9

Autonomous Land Vehicle In a Neural Network (ALVINN) (Dean Pomerleau) Training Data: Human Driver on real roads and simulator Sensory input is directly fed into the network 30