Tutorial Deep Learning : Unsupervised Feature Learning

Similar documents
Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN.

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

Extracting and Composing Robust Features with Denoising Autoencoders

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Day 3 Lecture 1. Unsupervised Learning

Stacked Denoising Autoencoders for Face Pose Normalization

Neural Networks for Machine Learning. Lecture 15a From Principal Components Analysis to Autoencoders

Deep Learning Srihari. Autoencoders. Sargur Srihari

Variational Autoencoders. Sargur N. Srihari

Learning Two-Layer Contractive Encodings

Vulnerability of machine learning models to adversarial examples

Model Generalization and the Bias-Variance Trade-Off

Bilevel Sparse Coding

Machine Learning for Physicists Lecture 6. Summer 2017 University of Erlangen-Nuremberg Florian Marquardt

Novel Lossy Compression Algorithms with Stacked Autoencoders

Introduction to Deep Learning

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x

Autoencoders, denoising autoencoders, and learning deep networks

Neural Networks: promises of current research

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Energy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt

Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Relationship between Fourier Space and Image Space. Academic Resource Center

Image Restoration Using DNN

Deep Learning. Volker Tresp Summer 2014

Neural Networks for unsupervised learning From Principal Components Analysis to Autoencoders to semantic hashing

An Empirical Evaluation of Deep Architectures on Problems with Many Factors of Variation

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2018

Auto-Encoding Variational Bayes

Learning Discrete Representations via Information Maximizing Self-Augmented Training

Depth Image Dimension Reduction Using Deep Belief Networks

Deep Learning Applications

GAN Frontiers/Related Methods

Exploring the Structure of Data at Scale. Rudy Agovic, PhD CEO & Chief Data Scientist at Reliancy January 16, 2019

Bo#leneck Features from SNR- Adap9ve Denoising Deep Classifier for Speaker Iden9fica9on

arxiv: v1 [cs.lg] 24 Jan 2019

Outlier detection using autoencoders

Lab 8 CSC 5930/ Computer Vision

Unsupervised Learning

Deep Similarity Learning for Multimodal Medical Images

arxiv: v1 [cs.cv] 21 Nov 2017

Ruslan Salakhutdinov and Geoffrey Hinton. University of Toronto, Machine Learning Group IRGM Workshop July 2007

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

CHAPTER 3 DIFFERENT DOMAINS OF WATERMARKING. domain. In spatial domain the watermark bits directly added to the pixels of the cover

Transfer Learning Using Rotated Image Data to Improve Deep Neural Network Performance

Image Denoising and Inpainting with Deep Neural Networks

Grundlagen der Künstlichen Intelligenz

Deep Generative Models Variational Autoencoders

C. Poultney S. Cho pra (NYU Courant Institute) Y. LeCun

Hand Written Digit Recognition Using Tensorflow and Python

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images

A Sparse and Locally Shift Invariant Feature Extractor Applied to Document Images

When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration

Neural Network Neurons

Deep Learning for Vision: Tricks of the Trade

Clustering algorithms and autoencoders for anomaly detection

Machine Problem 8 - Mean Field Inference on Boltzman Machine

Tutorial on Machine Learning Tools

Modeling and Optimization of Thin-Film Optical Devices using a Variational Autoencoder

MODEL-based recovery of images from noisy and sparse. MoDL: Model Based Deep Learning Architecture for Inverse Problems

Robust image recovery via total-variation minimization

PRINCIPAL COMPONENT ANALYSIS IMAGE DENOISING USING LOCAL PIXEL GROUPING

Pouya Kousha Fall 2018 CSE 5194 Prof. DK Panda

Encoding RNNs, 48 End of sentence (EOS) token, 207 Exploding gradient, 131 Exponential function, 42 Exponential Linear Unit (ELU), 44

Dynamic Routing Between Capsules

CSC 411: Lecture 14: Principal Components Analysis & Autoencoders

Histograms of Sparse Codes for Object Detection

THE MNIST DATABASE of handwritten digits Yann LeCun, Courant Institute, NYU Corinna Cortes, Google Labs, New York

CSC 411: Lecture 14: Principal Components Analysis & Autoencoders

Neural Networks and Deep Learning

Alternatives to Direct Supervision

Challenges motivating deep learning. Sargur N. Srihari

arxiv: v2 [cs.lg] 23 Oct 2018

DEEP learning algorithms have been the object of much

Handwritten Hindi Numerals Recognition System

Preparation Meeting. Recent Advances in the Analysis of 3D Shapes. Emanuele Rodolà Matthias Vestner Thomas Windheuser Daniel Cremers

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Chapter 4 Face Recognition Using Orthogonal Transforms

Wavelet Applications. Texture analysis&synthesis. Gloria Menegaz 1

Gaussian Mixture Model Coupled with Independent Component Analysis for Palmprint Verification

Generalized Autoencoder: A Neural Network Framework for Dimensionality Reduction

Robust Image Watermarking based on Discrete Wavelet Transform, Discrete Cosine Transform & Singular Value Decomposition

M. Sc. (Artificial Intelligence and Machine Learning)

MACHINE LEARNING CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS

Dynamic Routing Between Capsules. Yiting Ethan Li, Haakon Hukkelaas, and Kaushik Ram Ramasamy

A Wavelet Tour of Signal Processing The Sparse Way

Dietrich Paulus Joachim Hornegger. Pattern Recognition of Images and Speech in C++

Shape Classification and Cell Movement in 3D Matrix Tutorial (Part I)

arxiv: v2 [cs.lg] 22 Mar 2014

Non-Linear Hyperspectral Subspace Mapping using Stacked Autoencoder

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Analyzing Flow-based Anomaly Intrusion Detection using Replicator Neural Networks. Carlos García Cordero Sascha Hauke Max Mühlhäuser Mathias Fischer

A supervised strategy for deep kernel machine

QR Code Watermarking Algorithm based on Wavelet Transform

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models

List of Accepted Papers for ICVGIP 2018

Image Processing. Filtering. Slide 1

A Robust Digital Watermarking Scheme using BTC-PF in Wavelet Domain

Deep Semi-Supervised Embeddings for Dynamic Targeted Anomaly Detection

Transcription:

Tutorial Deep Learning : Unsupervised Feature Learning Joana Frontera-Pons 4th September 2017 - Workshop Dictionary Learning on Manifolds

OUTLINE Introduction Representation Learning TensorFlow Examples

DEEP LEARNING Deep Learning : Unsupervised Learning methods that can learn invariant features hierarchies, Non-linear representations obtained with deep layer structures allow to bring out complex relationships and disentangle the variation factors of the inputs, How? - AutoEncoders, ConvNets, Deep Belief Networks, What kind of representations the model can extract?

DEEP LEARNING Hidden layer 1 Hidden layer 2 Hidden layer 3

Motivation Motivation Motivation MOTIVATION Linear representations are frequently used to model images: Linear representations are frequently used to model images: Linear representations are frequently used to model images: x=d x=d x=d x is the original signal or image, original signal or image, D xisisa the transformation matrix (forsignal e.g. Discrete cosine transform, x is the original or image, D is a transformation matrixcomponent (for e.g. Discrete cosine transform, Wavelets transform, Principal Analysis, K-SVD dictionary), D is a transformation matrix (for e.g. Discrete cosine transform Wavelets transform, Principal Component Analysis, K-SVD dictionary), is the feature Wavelets vector. is the feature vector. transform, Principal Component Analysis, K-SVD dict is the feature vector. In this context, linear representations are the most widely spread In this context, linear representations are the most widely spread In this context, linear representations are solving, the most...widely sprea approaches for denoising, compression, inverse problems approaches for denoising, compression, inverse problems solving,... approaches for denoising, compression, inverse problems solving They fail at capturing some common variations in the data such as They fail at capturing some common variations in the data such as They fail at capturing some common variations in the data such translation, rotation, zoom translation, rotation, zoom translation, rotation, zoom New representations Newadaptive adaptivenon-linear non-linear representations New adaptive non-linear representations Visit to the CMM - Research and Integration Project J.Frontera-Pons J.Frontera-Pons Visit to the CMM - Research and Integration Project J.Frontera-Pons Visit to the CMM - Research and Integration Project J.Frontera-Pons 3/ 3/12 3 3/ 3

AUTOENCODERS Autoencoders are artificial neural networks capable of learning efficient representations of the input data without supervision. They are powerful feature extractors, They work by learning a to copy their inputs to their outputs, We have to constrain the network to prevent the model to learn the identity : limit the size of the representation, add noise, They find their purpose in : dimensionality reductcion, feature extraction unsupervised pretraining, or as generative models.

AUTOENCODERS Basic Autoencoders : Parametric encoding function from inputs to their representations, and a decoding function that maps back to input space, Train the model in order to reconstruct as accurately as possible the input. Hidden representation : h 2 R nhid f (x) = (W T f x + b f ) Encoder Input : x 2 R d Reconstruction Error : L(x, ˆx) - MeanSquared Error - Binary Cross-Entropy g (h) = (W T g h + b g ) Decoder Reconstruction : ˆx 2 R d

WEIGHT FILTERS PCA DAE

STACKED AUTOENCODERS Greedy layer-wise initialization: Ø Pre-training all the layers using unsupervised feature learning, Ø Add one level at a time, Ø Build hierarchy of representations. Stacking layers of autoencoders : Ø Pairs of encoder/decoder combined to form a global encoder and global decoder, Ø Deep autoencoders jointly trained and optimized for an overall reconstruction error. Outputs Hidden 3 Hidden 2 Hidden 1 Inputs

REGULARIZED AUTOENCODERS Basic Autoencoders may learn the identity function to minimize the reconstruction error, Impose some constraint on the code to force the representation to be insensitive to local variations Regularization, Examples: Contractive AE, Sparse AE, DAE, etc. DENOISING AUTOENCODERS [Vincent 2008] Modify the training objective to retrieve a clean input from an artificially corrupted version of it, Make the transformation robust to small random perturbations in the input, Corruption noise: Gaussian, salt and pepper, masking, or adaptive corruption process. J. FRONTERA- PONS - DEEP LEARNING 10 DEDALE TUB MEETING 19/04/2016

DENOISING AUTOENCODERS x x ˆx x J. FRONTERA- PONS - DEEP LEARNING 11 DEDALE TUB MEETING 19/04/2016

BASICS TENSORFLOW TensorFlow is an open source software library for numerical computation developed by the Google Brain team, The user defines in Python a graph of computations and then TF takes that graph and runs it efficiently using optimised C++. + Operation x + f(x, y) =x 2 y + y +2 x y 2 x Variable Constant

BASICS TENSORFLOW TensorFlow programs are typically split into two parts: Construction phase: Builds a computation graph representing the machine learning model and the computations required to train it. Execution phase: Runs a loop that evaluates a training step repeatedly. TensorBoard example

EXPERIMENTAL RESULTS Mixed National Institute of Standards and Technology database (MNIST) is a large database of handwritten digits, Training set containing 60000 samples and test set 1000 samples, BUILD FEATURES THAT CAPTURE THE VARIATIONS ALONG THE DIFFERENT DIGITS

EXPERIMENTAL RESULTS Encode Decode x h ˆx

Thank you for your attention!