Deep Learning & Neural Networks

Similar documents
Neural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /10/2017

CS 2750: Machine Learning. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh April 13, 2016

Deep Learning for Computer Vision II

Simple Model Selection Cross Validation Regularization Neural Networks

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Computer Vision Lecture 16

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Deep Learning with Tensorflow AlexNet

Deep Neural Networks:

CS 1674: Intro to Computer Vision. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh November 16, 2016

ImageNet Classification with Deep Convolutional Neural Networks

Deep Convolutional Neural Networks. Nov. 20th, 2015 Bruce Draper

Machine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

Neural Networks and Deep Learning

Perceptron as a graph

Deep Learning for Computer Vision

Advanced Introduction to Machine Learning, CMU-10715

Deep Learning Workshop. Nov. 20, 2015 Andrew Fishberg, Rowan Zellers

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

Deep Learning With Noise

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Machine Learning. MGS Lecture 3: Deep Learning

Using neural nets to recognize hand-written digits. Srikumar Ramalingam School of Computing University of Utah

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic

Computer Vision Lecture 16

ECE 6504: Deep Learning for Perception

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

CSE 559A: Computer Vision

Computer Vision Lecture 16

ROB 537: Learning-Based Control

Perceptron: This is convolution!

Weighted Convolutional Neural Network. Ensemble.

Convolutional Neural Networks

Study of Residual Networks for Image Recognition

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

Global Optimality in Neural Network Training

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2016

Deep Learning. Volker Tresp Summer 2015

Deep (1) Matthieu Cord LIP6 / UPMC Paris 6

Deep Learning Basic Lecture - Complex Systems & Artificial Intelligence 2017/18 (VO) Asan Agibetov, PhD.

Apparel Classifier and Recommender using Deep Learning

Machine learning for vision. It s the features, stupid! cathedral. high-rise. Winter Roland Memisevic. Lecture 2, January 26, 2016

Deep Learning. Volker Tresp Summer 2014

EECS 496 Statistical Language Models. Winter 2018

Deep Learning for Vision: Tricks of the Trade

Real-time Object Detection CS 229 Course Project

Data Mining. Neural Networks

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Lecture 20: Neural Networks for NLP. Zubin Pahuja

Large-Scale Visual Recognition With Deep Learning

Classification of objects from Video Data (Group 30)

Machine Learning 13. week

Introduction to Neural Networks

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Deep neural networks II

Como funciona o Deep Learning

Restricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version

JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

Is Bigger CNN Better? Samer Hijazi on behalf of IPG CTO Group Embedded Neural Networks Summit (enns2016) San Jose Feb. 9th

CS6220: DATA MINING TECHNIQUES

CS 523: Multimedia Systems

Neural Nets & Deep Learning

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal

On the Effectiveness of Neural Networks Classifying the MNIST Dataset

Neural Networks. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn

Rotation Invariance Neural Network

DEEP LEARNING WITH GPUS Maxim Milakov, Senior HPC DevTech Engineer, NVIDIA

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

Neural Networks (pp )

Know your data - many types of networks

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Deep Learning for Computer Vision with MATLAB By Jon Cherrie

Practical Methodology. Lecture slides for Chapter 11 of Deep Learning Ian Goodfellow

arxiv: v1 [cs.lg] 16 Jan 2013

Multi-Glance Attention Models For Image Classification

Lecture 7: Neural network acoustic models in speech recognition

Locally Scale-Invariant Convolutional Neural Networks

For Monday. Read chapter 18, sections Homework:

A Deep Learning primer

A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images

Introduction. Neural Networks. Chapter , , 18.7 and Deep Learning paper. Recognizing Digits using a Neural Net

Su et al. Shape Descriptors - III

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Seminars in Artifiial Intelligenie and Robotiis

Part Localization by Exploiting Deep Convolutional Networks

Deep Learning and Its Applications

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

AttentionNet for Accurate Localization and Detection of Objects. (To appear in ICCV 2015)

A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images

Introduction to Neural Networks

Fast Semantic Segmentation of RGB-D Scenes with GPU-Accelerated Deep Neural Networks

Deconvolutions in Convolutional Neural Networks

Multi-Task Learning of Facial Landmarks and Expression

Training Convolutional Neural Networks for Translational Invariance on SAR ATR

Transcription:

Deep Learning & Neural Networks Machine Learning CSE4546 Sham Kakade University of Washington November 29, 2016 Sham Kakade 1 Announcements: HW4 posted Poster Session Thurs, Dec 8 Today: Review: EM Neural nets and deep learning Sham Kakade 2 1

Poster Session Thursday Dec 8, 9-11:30am Please arrive 20 mins early to set up Everyone is expected to attend Prepare a poster We provide poster board and pins Both one large poster (recommended) and several pinned pages are OK. Capture Problem you are solving Data you used ML methodology Results Prepare a 1-minute speech about your project Two instructors will visit your poster separately Project Grading: scope, depth, data Sham Kakade 3 Logistic regression 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 P(Y X) represented by: 0.2 0.1 0-6 -4-2 0 2 4 6 Learning rule MLE: Sham Kakade 4 2

Perceptron as a graph 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0-6 -4-2 0 2 4 6 Sham Kakade 5 Linear perceptron classification region 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0-6 -4-2 0 2 4 6 Sham Kakade 6 3

Percepton, linear classification, Boolean functions Can learn x 1 AND x 2 Can learn x 1 OR x 2 Can learn any conjunction or disjunction Sham Kakade 7 Percepton, linear classification, Boolean functions Can learn majority Can perceptrons do everything? Sham Kakade 8 4

Going beyond linear classification Solving the XOR problem Sham Kakade 9 Hidden layer Perceptron: 1-hidden layer: Sham Kakade 10 5

Example data for NN with hidden layer Sham Kakade 11 Learned weights for hidden layer Sham Kakade 12 6

NN for images Sham Kakade 13 Weights in NN for images Sham Kakade 14 7

Forward propagation for 1-hidden layer - Prediction 1-hidden layer: Sham Kakade 15 Gradient descent for 1-hidden layer Back-propagation: Computing Dropped w 0 to make derivation simpler Sham Kakade 16 8

Gradient descent for 1-hidden layer Back-propagation: Computing Dropped w 0 to make derivation simpler Sham Kakade 17 Multilayer neural networks Sham Kakade 18 9

Forward propagation prediction Recursive algorithm Start from input layer Output of node V k with parents U 1,U 2, : Sham Kakade 19 Back-propagation learning Just stochastic gradient descent!!! Recursive algorithm for computing gradient For each example Perform forward propagation Start from output layer Compute gradient of node V k with parents U 1,U 2, Update weight w i k Sham Kakade 20 10

Many possible response/link functions Sigmoid Linear Exponential Gaussian Hinge Max Sham Kakade 21 Convolutional Neural Networks & Application to Computer Vision Machine Learning CSE4546 Sham Kakade University of Washington November 29, 2016 Sham Kakade 22 11

Contains slides from LeCun & Ranzato Russ Salakhutdinov Honglak Lee Sham Kakade 23 Neural Networks in Computer Vision Neural nets have made an amazing come back Used to engineer high-level features of images Image features: Sham Kakade 24 12

Some hand- created image features Computer$vision$features$ SIFT$ Spin$image$ HoG$ RIFT$ Textons$ GLOH$ Slide$Credit:$Honglak$Lee$ Sham Kakade 25 Scanning an image with a detector n Detector = Classifier from image patches: n Typically scan image with detector: Sham Kakade 26 13

Using neural nets to learn non-linear features transformation Low-Level Feature Mid-Level Feature High-Level Feature Trainable Classifier Feature visualization of convolutional net trained on ImageNet from [Zeiler & Fergus 2013] Sham Kakade 27 But, many tricks needed to work well for$image$recogniaon$ Sham Kakade 28 14

Convolution Layer Example: 200x200 image Fully-connected, 400,000 hidden units = 16 billion parameters Locally-connected, 400,000 hidden units 10x10 fields = 40 million params Local connections capture local dependencies Sham Kakade 29 Parameter sharing Fundamental technique used throughout ML Neural net without parameter sharing: Sharing parameters: Sham Kakade 30 15

Pooling/Subsampling n Convolutions act like detectors: n But we don t expect true detections in every patch Pooling/subsampling nodes: n Sham Kakade 31 Example neural net architecture Sham Kakade 32 16

Sample results Traffic Sign Recognition (GTSRB) German Traffic Sign Reco Bench 99.2% accuracy #1: IDSIA; #2 NYU House Number Recognition (Google) Street View House Numbers 94.3 % accuracy Sham Kakade 33 Example from Krizhevsky, Sutskever, Hinton 2012 Won the 2012 ImageNet LSVRC. 60 Million parameters, 832M MAC ops 4M 16M 37M FULL CONNECT FULL 4096/ReLU FULL 4096/ReLU 4Mflop 16M 37M 442K 1.3M 884K 307K 35K MAX POOLING CONV 3x3/ReLU 256fm CONV 3x3ReLU 384fm CONV 3x3/ReLU 384fm MAX POOLING 2x2sub LOCAL CONTRAST NORM CONV 11x11/ReLU 256fm MAX POOL 2x2sub LOCAL CONTRAST NORM CONV 11x11/ReLU 96fm 74M 224M 149M 223M 105M Sham Kakade 34 17

Results by Krizhevsky, Sutskever, Hinton 2012 ImageNet Large Scale Visual Recognition Challenge 1000 categories, 1.5 Million labeled training samples Sham Kakade 35 Sham Kakade 36 18

TEST IMAGE RETRIEVED IMAGES Sham Kakade 37 Application to scene parsing [Farabet et al. ICML 2012, PAMI 2013] Sham Kakade 38 19

Learning challenges for neural nets Choosing architecture Slow per iteration and convergence Gradient diffusion across layers Many local optima Sham Kakade 39 Random dropouts Standard backprop: Random dropouts: randomly choose edges not to update: Functions as a type of regularization helps avoid diffusion of gradient Sham Kakade 40 20

Revival of neural networks Neural networks fell into disfavor in mid 90s -early 2000s Many methods have now been rediscovered J Exciting new results using modifications to optimization techniques and GPUs Challenges still remain: Architecture selection feels like a black art Optimization can be very sensitive to parameters Requires a significant amount of expertise to get good results Sham Kakade 41 21