Amortised MAP Inference for Image Super-resolution. Casper Kaae Sønderby, Jose Caballero, Lucas Theis, Wenzhe Shi & Ferenc Huszár ICLR 2017

Similar documents
AMORTISED MAP INFERENCE FOR IMAGE SUPER-RESOLUTION

Deep generative models of natural images

Pixel-level Generative Model

Unsupervised Learning

Alternatives to Direct Supervision

DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE MACHINE LEARNING & DATA MINING - LECTURE 17

Progress on Generative Adversarial Networks

GAN Frontiers/Related Methods

Deep Generative Models Variational Autoencoders

PIXELCNN++: IMPROVING THE PIXELCNN WITH DISCRETIZED LOGISTIC MIXTURE LIKELIHOOD AND OTHER MODIFICATIONS

Implicit generative models: dual vs. primal approaches

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

GAN and Feature Representation. Hung-yi Lee

Recovering Realistic Texture in Image Super-resolution by Deep Spatial Feature Transform. Xintao Wang Ke Yu Chao Dong Chen Change Loy

arxiv: v2 [cs.lg] 17 Dec 2018

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x

Adversarial Symmetric Variational Autoencoder

IMPROVING SAMPLING FROM GENERATIVE AUTOENCODERS WITH MARKOV CHAINS

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17

Towards Principled Methods for Training Generative Adversarial Networks. Martin Arjovsky & Léon Bottou

arxiv: v2 [stat.ml] 21 Oct 2017

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Towards Conceptual Compression

arxiv: v1 [cs.lg] 28 Feb 2017

Variational Autoencoders. Sargur N. Srihari

Deep Learning for Visual Manipulation and Synthesis

When Variational Auto-encoders meet Generative Adversarial Networks

Bc. David Honzátko. Generative Neural Networks in Image Reconstruction

arxiv: v3 [cs.lg] 30 Dec 2016

IMPLICIT AUTOENCODERS

Introduction to Generative Adversarial Networks

arxiv: v4 [stat.ml] 4 Dec 2017

arxiv: v1 [cs.cv] 4 Apr 2018

IVE-GAN: INVARIANT ENCODING GENERATIVE AD-

Brain MRI super-resolution using 3D generative adversarial networks

Introduction to Generative Adversarial Networks

GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin

DCGANs for image super-resolution, denoising and debluring

Video Compression Using Recurrent Convolutional Neural Networks

Generative Adversarial Networks (GANs) Based on slides from Ian Goodfellow s NIPS 2016 tutorial

arxiv: v2 [cs.lg] 9 Jun 2017

19: Inference and learning in Deep Learning

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

Symmetric Variational Autoencoder and Connections to Adversarial Learning

DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION

Stochastic Simulation with Generative Adversarial Networks

Controllable Generative Adversarial Network

Single Image Super Resolution of Textures via CNNs. Andrew Palmer

Introduction to Generative Models (and GANs)

Auxiliary Guided Autoregressive Variational Autoencoders

arxiv: v2 [cs.lg] 7 Feb 2019

DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION

The Amortized Bootstrap

Semantic Segmentation. Zhongang Qi

Denoising Adversarial Autoencoders

Semi-Amortized Variational Autoencoders

Extracting and Composing Robust Features with Denoising Autoencoders

Bayesian model ensembling using meta-trained recurrent neural networks

arxiv: v1 [stat.ml] 1 Mar 2017

Generative Adversarial Networks (GANs)

Generative Models II. Phillip Isola, MIT, OpenAI DLSS 7/27/18

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

Improved Boundary Equilibrium Generative Adversarial Networks

Inverting The Generator Of A Generative Adversarial Network

CSC412: Stochastic Variational Inference. David Duvenaud

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

arxiv: v1 [eess.sp] 23 Oct 2018

Unrolling Inference: The Recurrent Inference Machine

Machine Learning Lecture 3

Supplementary Material: Unsupervised Domain Adaptation for Face Recognition in Unlabeled Videos

What is machine learning?

arxiv: v3 [cs.cv] 14 Nov 2017

Lab meeting (Paper review session) Stacked Generative Adversarial Networks

Deep Convolutional AutoEncoder-based Lossy Image Compression

arxiv: v1 [cs.lg] 8 Dec 2016 Abstract

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

RNNs as Directed Graphical Models

Machine Learning. MGS Lecture 3: Deep Learning

Generative Adversarial Networks (GANs) Ian Goodfellow, Research Scientist MLSLP Keynote, San Francisco

UCLA UCLA Electronic Theses and Dissertations

Adversarially Learned Inference

arxiv: v2 [cs.cv] 4 Jun 2017

An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks

arxiv: v1 [cs.cv] 5 Jul 2017

27: Hybrid Graphical Models and Neural Networks

Flow-GAN: Bridging implicit and prescribed learning in generative models

Dist-GAN: An Improved GAN using Distance Constraints

Lecture 19: Generative Adversarial Networks

arxiv: v1 [stat.ml] 15 Feb 2018

arxiv: v1 [cs.cv] 17 Apr 2017 Abstract

Unsupervised Image-to-Image Translation Networks

Generative Adversarial Network

arxiv: v3 [cs.cv] 3 Jul 2018

Gradient of the lower bound

Auto-Encoding Variational Bayes

Energy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt

An Inner-loop Free Solution to Inverse Problems using Deep Neural Networks

Deep Generative Models and a Probabilistic Programming Library

Transcription:

Amortised MAP Inference for Image Super-resolution Casper Kaae Sønderby, Jose Caballero, Lucas Theis, Wenzhe Shi & Ferenc Huszár ICLR 2017

Super Resolution Inverse problem: Given low resolution representation x reconstruct high resolution image y

Ranked Inference Choices 1. Empirical Risk Minimization: minimize a loss function measuring what we care about (We don t know the right loss function, in fact we can t even measure perceptual quality) 2. Maximum a Posteriori (MAP) inference using knowledge of image prior (We don t know the prior) 3. Approximate MAP

Motivation: Blurry Images MSE is the wrong objective for photo-realistic results MSE Super-Resolution (4x) Original HR

Motivation: The Divergence Perspective Data KL[P Q] Maximum Likelihood (overdispersed) JSD[PIQ] Original GAN criterion (in between) KL[QIP] Modified GAN criterion (mode seeking)

Motivation: The 2D Perspective Target: Input: 2D High resolution, y= [y1, y2] drawn from a Swiss-roll 1D Low resolution, x = (y1+y2)/2, average of high resolution Example Input: x = 0.5 Valid outputs fall on the line: y1= 1- y2

Approximate Amortized MAP Inference Maximize Log-posterior evaluated at the predicted output Decomposed via Bayes rule Likelihood data consistency Prior plausible output

Data Consistency: affine projections Input Linear downsampling (strided convolution) Affine Projected Network Moore-Penrose pseudo inverse of A ( subpixel (transposed) convolution) For Affine projected networks the likelihood is always maximally satisfied and can be ignored

Plausible output: Cross Entropy Cross Entropy objective We propose three methods to optimize this 1. AffGAN: GAN using modified update rule minimizing 2. AffDAE: Optimize using gradients from denoiser 3. AffLL: Fit differentiable parametric density model to the image distribution and get gradient estimates directly.

Fix GAN Instability: Instance Noise The data and model distributions will most likely not share support Discriminator can always perfectly separate the samples KL-divergence is undefined and in general the GAN convergence proof does not hold No gradient to train generator Perfectly separable distributions Instance Noise Instance Noise: Broaden support of model and data distributions by adding gaussian noise Minimises: (Arjovsky & and Bottou, ICLR 2017 arrives at the same solution)

2D Super resolution - Revisited Target: Input: 2D High resolution, y= [y1, y2] drawn from a Swiss-roll 1D Low resolution, x = (y1+y2)/2, average of high resolution

Results - Grass textures y x MSE AffGAN AffDG AffLL

Natural Images Mode seeking behavior Correct output is often ambiguous Stochastic extension Input: Low-res Output High-res Target High-res

AffGAN as Variational Inference Stochastic Affine Projected GAN (StAffGAN) (All two letter GAN acronyms were occupied) Add noise source z in addition to input x For super-resolution we can show that Consequence: StAffGAN performs Variational Inference for image super-resolution only requiring samples from px and py

Conditional GAN AffGAN BS-GAN EBGAN DCGAN InfoGAN f-gan Mode-seeking GAN GAN LSGAN JS-GAN Temporal GAN GRAN VEGAN StAffGAN SRGAN VGAN Wasserstein GAN Unrolled GAN

Conclusion Affine projections restricting model to the affine subspace of valid solutions We proposed three methods for amortized MAP inference in image super-resolution A modified GAN objective Denoiser guided optimization Direct parametric likelihood modelling In practice the GAN based objective produced the visually most appealing results Provide theoretical grounding for GANs in generative models We showed how GAN models be seen as performing Amortized Variational Inference

Further Slides

How to compute A+ In the language of Deep Learning: A is a strided convolution with, say, Gaussian kernel A+ is a subpixel (transposed) convolution For fixed A, we find A+ via numerical optimization:

Can Affine Projected Networks learn? Proof of Concept using MSE MSE, high-res SSIM, high-res MSE, low-res, (consistency) Legend tuple indicate: (F)ixed / (T)rainable affine transformation, (R)andom / (T)rained initialization of transformation

Approximate MAP inference - Method 1: Model p(y) Directly model log[p(y)] using maximum likelihood learning Fit differentiable parametric model to log[p(y)] maximize f(x) using gradients of log[p(y)] We use with PixelCNN + MCGSM PixelCNN: Convolutional Network satisfying the Chain Rule* MCGSM: Mixture of Conditional Gaussians Scale Mixture model** *Aaron van den Oord, Nal Kalchbrenner, and Koray Kavukcuoglu. Pixel recurrent neural networks. arxiv preprint arxiv:1601.06759, 2016. ** Theis, Lucas, and Matthias Bethge. Generative image modeling using spatial lstms. Advances in Neural Information Processing Systems. 2015.

Approximate MAP inference - Method 2: DAE To maximize the prior we need the gradients Luckily these can be approximated using a Denoising Autoencoder * *Guillaume Alain and Yoshua Bengio. What regularized auto-encoders learn from the data-generating distribution. Journal of Machine Learning Research, 15(1):3563 3593, 2014.

Approximate MAP inference - Method 3: GAN We propose a modified GAN update rule that minimizes KL[Q P] Minimizing KL[Q P] Maximize MAP + Entropy-term

Super Resolution Inverse problem: Given low resolution representation x reconstruct high resolution image y Ambiguous / Multimodal