Towards Principled Methods for Training Generative Adversarial Networks. Martin Arjovsky & Léon Bottou

Size: px
Start display at page:

Download "Towards Principled Methods for Training Generative Adversarial Networks. Martin Arjovsky & Léon Bottou"

Transcription

1 Towards Principled Methods for Training Generative Adversarial Networks Martin Arjovsky & Léon Bottou

2 Unsupervised learning - We have samples from an unknown distribution

3 Unsupervised learning - We have samples from an unknown distribution - We want to approximate it by a parametric distribution that s close to in some sense.

4 Unsupervised learning - We have samples from an unknown distribution - We want to approximate it by a parametric distribution that s close to in some sense. - Close how?

5 Maximum Likelihood - Maximum likelihood:

6 Maximum Likelihood - Maximum likelihood: - Assumptions: continuous with full support.

7 Maximum Likelihood - Maximum likelihood: - Assumptions: continuous with full support. - Problems: restricted capacity distributes mass. Modeling low dimensional distributions is impossible.

8 Kullback-Leibler Divergence - Closeness measured by KL divergence (equivalent to ML):

9 Kullback-Leibler Divergence - Closeness measured by KL divergence (equivalent to ML): - When integrand goes to infinity: high cost for mode dropping.

10 Kullback-Leibler Divergence - Closeness measured by KL divergence (equivalent to ML): - When integrand goes to infinity: high cost for mode dropping. - When integrand goes to 0: low cost for fake looking samples.

11 Generative Adversarial Networks (Goodfellow et al.) - Let be the dist of for some simple (e.g. Gaussian) r.v Z, passed through a complex function.

12 Generative Adversarial Networks (Goodfellow et al.) - Let be the dist of for some simple (e.g. Gaussian) r.v Z, passed through a complex function. - Discriminator maximizes and generator minimizes

13 Generative Adversarial Networks (Goodfellow et al.) - Let be the dist of for some simple (e.g. Gaussian) r.v Z, passed through a complex function. - Discriminator maximizes and generator minimizes

14 JSD seems maxed out..

15 Generative Adversarial Networks - Under optimal discriminator, minimizes - Problems: vanishing gradients very quickly when D s accuracy is high.

16 Discriminator is pretty good...

17 Vanishing gradients, original cost

18 Alternate update - Alternate update that has less vanishing gradients

19 Alternate update - Alternate update that has less vanishing gradients - Under optimality optimizes

20 Alternate update - Alternate update that has less vanishing gradients - Under optimality optimizes - Problems: JSD with the wrong sign, reverse KL has high mode dropping. Still unstable when D is good.

21 High variance updates

22 Problems of GANs (and divergences) - When and lie on low dimensional manifolds, there s always a perfect discriminator, that provides no usable gradients.

23 Manifold picture - Real - Generated

24 Problems of GANs (and divergences) - When and lie on low dimensional manifolds, there s always a perfect discriminator, that provides no usable gradients.

25 Problems of GANs (and divergences) - When and lie on low dimensional manifolds, there s always a perfect discriminator, that provides no usable gradients.

26 Problems of GANs (and divergences) - When and lie on low dimensional manifolds, there s always a perfect discriminator, that provides no usable gradients. - Under the same assumptions

27 A first step to a solution - Distributions are essentially disjoint

28 A first step to a solution - Distributions are essentially disjoint - Add noise during training to make them overlap!

29 A first step to a solution - Distributions are essentially disjoint - Add noise during training to make them overlap! - Matching noisy distributions amounts to matching the underlying ones.

30 Manifold picture - Real - Generated

31 Manifold picture with noise - Real - Generated

32 A first step to a solution We move our samples towards point in the data manifold, weighted by their probability and distance to our samples.

33 Theoretical guarantee

34 Theoretical guarantee - Wasserstein is well defined in the manifold setting.

35 Theoretical guarantee - Wasserstein is well defined in the manifold setting. - The noise method optimizes an upper bound of it.

36 Theoretical guarantee - Wasserstein is well defined in the manifold setting. - The noise method optimizes an upper bound of it. - We can reduce the first summand by annealing the noise, the second one by optimizing with noise.

37 Loads of work done since then! - Now we have more understanding of the relationship between Wasserstein, JSD and the rest: Weak vs strong.

38 Loads of work done since then! - Now we have more understanding of the relationship between Wasserstein, JSD and the rest: Weak vs strong. - Optimizing an approximation of Wasserstein directly is doable. (Arjovsky, Chintala & Bottou, 2017)

39 Loads of work done since then! - Now we have more understanding of the relationship between Wasserstein, JSD and the rest: Weak vs strong. - Optimizing an approximation of Wasserstein directly is doable. (Arjovsky, Chintala & Bottou, 2017) - Different ways to do this. (Gulrajani et al. 2017)

40 Loads of work done since then! - Now we have more understanding of the relationship between Wasserstein, JSD and the rest: Weak vs strong. - Optimizing an approximation of Wasserstein directly is doable. (Arjovsky, Chintala & Bottou, 2017) - Different ways to do this. (Gulrajani et al. 2017) - Time to scale up!

41

Implicit generative models: dual vs. primal approaches

Implicit generative models: dual vs. primal approaches Implicit generative models: dual vs. primal approaches Ilya Tolstikhin MPI for Intelligent Systems ilya@tue.mpg.de Machine Learning Summer School 2017 Tübingen, Germany Contents 1. Unsupervised generative

More information

Alternatives to Direct Supervision

Alternatives to Direct Supervision CreativeAI: Deep Learning for Graphics Alternatives to Direct Supervision Niloy Mitra Iasonas Kokkinos Paul Guerrero Nils Thuerey Tobias Ritschel UCL UCL UCL TUM UCL Timetable Theory and Basics State of

More information

Unsupervised Learning

Unsupervised Learning Deep Learning for Graphics Unsupervised Learning Niloy Mitra Iasonas Kokkinos Paul Guerrero Vladimir Kim Kostas Rematas Tobias Ritschel UCL UCL/Facebook UCL Adobe Research U Washington UCL Timetable Niloy

More information

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao Introduction to GAN Generative Adversarial Networks Junheng(Jeff) Hao Adversarial Training is the coolest thing since sliced bread. -- Yann LeCun Roadmap 1. Generative Modeling 2. GAN 101: What is GAN?

More information

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao Introduction to GAN Generative Adversarial Networks Junheng(Jeff) Hao Adversarial Training is the coolest thing since sliced bread. -- Yann LeCun Roadmap 1. Generative Modeling 2. GAN 101: What is GAN?

More information

Lecture 3 GANs and Their Applications in Image Generation

Lecture 3 GANs and Their Applications in Image Generation Lecture 3 GANs and Their Applications in Image Generation Lin ZHANG, PhD School of Software Engineering Tongji University Fall 2017 Outline Introduction Theoretical Part Application Part Existing Implementations

More information

arxiv: v2 [cs.lg] 17 Dec 2018

arxiv: v2 [cs.lg] 17 Dec 2018 Lu Mi 1 * Macheng Shen 2 * Jingzhao Zhang 2 * 1 MIT CSAIL, 2 MIT LIDS {lumi, macshen, jzhzhang}@mit.edu The authors equally contributed to this work. This report was a part of the class project for 6.867

More information

Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN.

Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN. Stacked Denoising Sparse Variational (Adapted from Paul Quint and Ian Goodfellow) Stacked Denoising Sparse Variational Autoencoding is training a network to replicate its input to its output Applications:

More information

Introduction to Generative Adversarial Networks

Introduction to Generative Adversarial Networks Introduction to Generative Adversarial Networks Luke de Oliveira Vai Technologies Lawrence Berkeley National Laboratory @lukede0 @lukedeo lukedeo@vaitech.io https://ldo.io 1 Outline Why Generative Modeling?

More information

Variational Autoencoders. Sargur N. Srihari

Variational Autoencoders. Sargur N. Srihari Variational Autoencoders Sargur N. srihari@cedar.buffalo.edu Topics 1. Generative Model 2. Standard Autoencoder 3. Variational autoencoders (VAE) 2 Generative Model A variational autoencoder (VAE) is a

More information

Deep Generative Models Variational Autoencoders

Deep Generative Models Variational Autoencoders Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative

More information

arxiv: v1 [eess.sp] 23 Oct 2018

arxiv: v1 [eess.sp] 23 Oct 2018 Reproducing AmbientGAN: Generative models from lossy measurements arxiv:1810.10108v1 [eess.sp] 23 Oct 2018 Mehdi Ahmadi Polytechnique Montreal mehdi.ahmadi@polymtl.ca Mostafa Abdelnaim University de Montreal

More information

An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks

An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks Report for Undergraduate Project - CS396A Vinayak Tantia (Roll No: 14805) Guide: Prof Gaurav Sharma CSE, IIT Kanpur, India

More information

Quantitative Evaluation of Generative Adversarial Networks and Improved Training Techniques

Quantitative Evaluation of Generative Adversarial Networks and Improved Training Techniques Quantitative Evaluation of Generative Adversarial Networks and Improved Training Techniques by Yadong Li to obtain the degree of Master of Science at the Delft University of Technology, to be defended

More information

Generative Adversarial Networks (GANs)

Generative Adversarial Networks (GANs) Generative Adversarial Networks (GANs) Hossein Azizpour Most of the slides are courtesy of Dr. Ian Goodfellow (Research Scientist at OpenAI) and from his presentation at NIPS 2016 tutorial Note. I am generally

More information

Variational Autoencoders

Variational Autoencoders red red red red red red red red red red red red red red red red red red red red Tutorial 03/10/2016 Generative modelling Assume that the original dataset is drawn from a distribution P(X ). Attempt to

More information

DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE MACHINE LEARNING & DATA MINING - LECTURE 17

DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE MACHINE LEARNING & DATA MINING - LECTURE 17 DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE 155 - MACHINE LEARNING & DATA MINING - LECTURE 17 GENERATIVE MODELS DATA 3 DATA 4 example 1 DATA 5 example 2 DATA 6 example 3 DATA 7 number of

More information

arxiv: v1 [cs.gr] 22 Jan 2019

arxiv: v1 [cs.gr] 22 Jan 2019 Generation High resolution 3D model from natural language by Generative Adversarial Network Kentaro Fukamizu, Masaaki Kondo, Ryuichi Sakamoto arxiv:1901.07165v1 [cs.gr] 22 Jan 2019 Abstract Since creating

More information

Amortised MAP Inference for Image Super-resolution. Casper Kaae Sønderby, Jose Caballero, Lucas Theis, Wenzhe Shi & Ferenc Huszár ICLR 2017

Amortised MAP Inference for Image Super-resolution. Casper Kaae Sønderby, Jose Caballero, Lucas Theis, Wenzhe Shi & Ferenc Huszár ICLR 2017 Amortised MAP Inference for Image Super-resolution Casper Kaae Sønderby, Jose Caballero, Lucas Theis, Wenzhe Shi & Ferenc Huszár ICLR 2017 Super Resolution Inverse problem: Given low resolution representation

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x Deep Learning 4 Autoencoder, Attention (spatial transformer), Multi-modal learning, Neural Turing Machine, Memory Networks, Generative Adversarial Net Jian Li IIIS, Tsinghua Autoencoder Autoencoder Unsupervised

More information

Progress on Generative Adversarial Networks

Progress on Generative Adversarial Networks Progress on Generative Adversarial Networks Wangmeng Zuo Vision Perception and Cognition Centre Harbin Institute of Technology Content Image generation: problem formulation Three issues about GAN Discriminate

More information

Introduction to Generative Models (and GANs)

Introduction to Generative Models (and GANs) Introduction to Generative Models (and GANs) Haoqiang Fan fhq@megvii.com Nov. 2017 Figures adapted from NIPS 2016 Tutorial Generative Adversarial Networks Generative Models: Learning the Distributions

More information

Earth Mover s Distance and The Applications

Earth Mover s Distance and The Applications Earth Mover s Distance and The Applications Hu Ding Computer Science and Engineering, Michigan State University The Motivations It is easy to compare two single objects: the pairwise distance. For example:

More information

Generative Adversarial Network: a Brief Introduction. Lili Mou

Generative Adversarial Network: a Brief Introduction. Lili Mou Generative Adversarial Network: a Brief Introduction Lili Mou doublepower.mou@gmail.com Outline Generative adversarial net Conditional generative adversarial net Deep generative image models using Laplacian

More information

Learning to generate with adversarial networks

Learning to generate with adversarial networks Learning to generate with adversarial networks Gilles Louppe June 27, 2016 Problem statement Assume training samples D = {x x p data, x X } ; We want a generative model p model that can draw new samples

More information

arxiv: v1 [cs.ne] 11 Jun 2018

arxiv: v1 [cs.ne] 11 Jun 2018 Generative Adversarial Network Architectures For Image Synthesis Using Capsule Networks arxiv:1806.03796v1 [cs.ne] 11 Jun 2018 Yash Upadhyay University of Minnesota, Twin Cities Minneapolis, MN, 55414

More information

Lecture 19: Generative Adversarial Networks

Lecture 19: Generative Adversarial Networks Lecture 19: Generative Adversarial Networks Roger Grosse 1 Introduction Generative modeling is a type of machine learning where the aim is to model the distribution that a given set of data (e.g. images,

More information

Generative Modeling with Convolutional Neural Networks. Denis Dus Data Scientist at InData Labs

Generative Modeling with Convolutional Neural Networks. Denis Dus Data Scientist at InData Labs Generative Modeling with Convolutional Neural Networks Denis Dus Data Scientist at InData Labs What we will discuss 1. 2. 3. 4. Discriminative vs Generative modeling Convolutional Neural Networks How to

More information

Gaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017

Gaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017 Gaussian Processes for Robotics McGill COMP 765 Oct 24 th, 2017 A robot must learn Modeling the environment is sometimes an end goal: Space exploration Disaster recovery Environmental monitoring Other

More information

19: Inference and learning in Deep Learning

19: Inference and learning in Deep Learning 10-708: Probabilistic Graphical Models 10-708, Spring 2017 19: Inference and learning in Deep Learning Lecturer: Zhiting Hu Scribes: Akash Umakantha, Ryan Williamson 1 Classes of Deep Generative Models

More information

arxiv: v1 [stat.ml] 15 Feb 2018

arxiv: v1 [stat.ml] 15 Feb 2018 Conditioning of three-dimensional generative adversarial networks for pore and reservoir-scale models arxiv:1802.05622v1 [stat.ml] 15 Feb 2018 Lukas J. Mosser lukas.mosser15@imperial.ac.uk Martin J. Blunt

More information

GAN Frontiers/Related Methods

GAN Frontiers/Related Methods GAN Frontiers/Related Methods Improving GAN Training Improved Techniques for Training GANs (Salimans, et. al 2016) CSC 2541 (07/10/2016) Robin Swanson (robin@cs.toronto.edu) Training GANs is Difficult

More information

Mode Regularized Generative Adversarial Networks

Mode Regularized Generative Adversarial Networks Mode Regularized Generative Adversarial Networks Tong Che 1 Yanran Li 2 Athul Paul Jacob 3 Yoshua Bengio 1 Wenjie Li 2 1 Montreal Institute for Learning Algorithms, Universite de Montreal, Montreal, Canada

More information

S+U Learning through ANs - Pranjit Kalita

S+U Learning through ANs - Pranjit Kalita S+U Learning through ANs - Pranjit Kalita - (from paper) Learning from Simulated and Unsupervised Images through Adversarial Training - Ashish Shrivastava, Tomas Pfister, Oncel Tuzel, Josh Susskind, Wenda

More information

Stochastic Simulation with Generative Adversarial Networks

Stochastic Simulation with Generative Adversarial Networks Stochastic Simulation with Generative Adversarial Networks Lukas Mosser, Olivier Dubrule, Martin J. Blunt lukas.mosser15@imperial.ac.uk, o.dubrule@imperial.ac.uk, m.blunt@imperial.ac.uk (Deep) Generative

More information

Symmetric Variational Autoencoder and Connections to Adversarial Learning

Symmetric Variational Autoencoder and Connections to Adversarial Learning Symmetric Variational Autoencoder and Connections to Adversarial Learning Liqun Chen 1 Shuyang Dai 1 Yunchen Pu 1 Erjin Zhou 4 Chunyuan Li 1 Qinliang Su 2 Changyou Chen 3 Lawrence Carin 1 1 Duke University,

More information

Adversarially Learned Inference

Adversarially Learned Inference Institut des algorithmes d apprentissage de Montréal Adversarially Learned Inference Aaron Courville CIFAR Fellow Université de Montréal Joint work with: Vincent Dumoulin, Ishmael Belghazi, Olivier Mastropietro,

More information

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17 Improving Generative Adversarial Networks with Denoising Feature Matching David Warde-Farley 1 Yoshua Bengio 1 1 University of Montreal, ICLR,2017 Presenter: Bargav Jayaraman Outline 1 Introduction 2 Background

More information

Model Generalization and the Bias-Variance Trade-Off

Model Generalization and the Bias-Variance Trade-Off Charu C. Aggarwal IBM T J Watson Research Center Yorktown Heights, NY Model Generalization and the Bias-Variance Trade-Off Neural Networks and Deep Learning, Springer, 2018 Chapter 4, Section 4.1-4.2 What

More information

Reconstructing Pore Networks Using Generative Adversarial Networks

Reconstructing Pore Networks Using Generative Adversarial Networks Reconstructing Pore Networks Using Generative Adversarial Networks Kelly Guan (kmguan@stanford.edu) I. INTRODUCTION Understanding fluid flow in porous media at the microscale is relevant to many fields,

More information

arxiv: v3 [cs.cv] 30 Oct 2017

arxiv: v3 [cs.cv] 30 Oct 2017 Improved Adversarial Systems for 3D Object Generation and Reconstruction Edward J. Smith Department of Computer Science McGill University Canada edward.smith@mail.mcgill.ca David Meger Department of Computer

More information

Lecture 7: Support Vector Machine

Lecture 7: Support Vector Machine Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each

More information

Class-Splitting Generative Adversarial Networks

Class-Splitting Generative Adversarial Networks Class-Splitting Generative Adversarial Networks Guillermo L. Grinblat 1, Lucas C. Uzal 1, and Pablo M. Granitto 1 arxiv:1709.07359v2 [stat.ml] 17 May 2018 1 CIFASIS, French Argentine International Center

More information

Dist-GAN: An Improved GAN using Distance Constraints

Dist-GAN: An Improved GAN using Distance Constraints Dist-GAN: An Improved GAN using Distance Constraints Ngoc-Trung Tran [0000 0002 1308 9142], Tuan-Anh Bui [0000 0003 4123 262], and Ngai-Man Cheung [0000 0003 0135 3791] ST Electronics - SUTD Cyber Security

More information

arxiv: v3 [cs.cv] 3 Jul 2018

arxiv: v3 [cs.cv] 3 Jul 2018 Improved Training of Generative Adversarial Networks using Representative Features Duhyeon Bang 1 Hyunjung Shim 1 arxiv:1801.09195v3 [cs.cv] 3 Jul 2018 Abstract Despite the success of generative adversarial

More information

GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin

GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin GENERATIVE MODEL Given a training dataset, x, try to estimate the distribution, Pdata(x) Explicitly or Implicitly (GAN) Explicitly

More information

An Introduction to Statistical Methods of Medical Image Registration

An Introduction to Statistical Methods of Medical Image Registration his is page 1 Printer: Opaque this An Introduction to Statistical Methods of Medical Image Registration Lilla Zöllei, John Fisher, William Wells ABSRAC After defining the medical image registration problem,

More information

Machine Learning Lecture 3

Machine Learning Lecture 3 Machine Learning Lecture 3 Probability Density Estimation II 19.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Exam dates We re in the process

More information

Autoencoders, denoising autoencoders, and learning deep networks

Autoencoders, denoising autoencoders, and learning deep networks 4 th CiFAR Summer School on Learning and Vision in Biology and Engineering Toronto, August 5-9 2008 Autoencoders, denoising autoencoders, and learning deep networks Part II joint work with Hugo Larochelle,

More information

maagma Modified-Adversarial-Autoencoding Generator with Multiple Adversaries: Optimizing Multiple Learning Objectives for Image Generation with GANs

maagma Modified-Adversarial-Autoencoding Generator with Multiple Adversaries: Optimizing Multiple Learning Objectives for Image Generation with GANs maagma Modified-Adversarial-Autoencoding Generator with Multiple Adversaries: Optimizing Multiple Learning Objectives for Image Generation with GANs Sahil Chopra Stanford University 353 Serra Mall, Stanford

More information

CSC412: Stochastic Variational Inference. David Duvenaud

CSC412: Stochastic Variational Inference. David Duvenaud CSC412: Stochastic Variational Inference David Duvenaud Admin A3 will be released this week and will be shorter Motivation for REINFORCE Class projects Class Project ideas Develop a generative model for

More information

Convex Clustering with Exemplar-Based Models

Convex Clustering with Exemplar-Based Models Convex Clustering with Exemplar-Based Models Danial Lashkari Polina Golland Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, MA 2139 {danial, polina}@csail.mit.edu

More information

Convex Clustering with Exemplar-Based Models

Convex Clustering with Exemplar-Based Models Convex Clustering with Exemplar-Based Models Danial Lashkari Polina Golland Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, MA 2139 {danial, polina}@csail.mit.edu

More information

Distribution Distance Functions

Distribution Distance Functions COMP 875 November 10, 2009 Matthew O Meara Question How similar are these? Outline Motivation Protein Score Function Object Retrieval Kernel Machines 1 Motivation Protein Score Function Object Retrieval

More information

arxiv: v1 [stat.ml] 11 Feb 2018

arxiv: v1 [stat.ml] 11 Feb 2018 Paul K. Rubenstein Bernhard Schölkopf Ilya Tolstikhin arxiv:80.0376v [stat.ml] Feb 08 Abstract We study the role of latent space dimensionality in Wasserstein auto-encoders (WAEs). Through experimentation

More information

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA

Modeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA Modeling and Reasoning with Bayesian Networks Adnan Darwiche University of California Los Angeles, CA darwiche@cs.ucla.edu June 24, 2008 Contents Preface 1 1 Introduction 1 1.1 Automated Reasoning........................

More information

MTTTS17 Dimensionality Reduction and Visualization. Spring 2018 Jaakko Peltonen. Lecture 11: Neighbor Embedding Methods continued

MTTTS17 Dimensionality Reduction and Visualization. Spring 2018 Jaakko Peltonen. Lecture 11: Neighbor Embedding Methods continued MTTTS17 Dimensionality Reduction and Visualization Spring 2018 Jaakko Peltonen Lecture 11: Neighbor Embedding Methods continued This Lecture Neighbor embedding by generative modeling Some supervised neighbor

More information

Parameter Continuation with Secant Approximation for Deep Neural Networks. Harsh Nilesh Pathak

Parameter Continuation with Secant Approximation for Deep Neural Networks. Harsh Nilesh Pathak Parameter Continuation with Secant Approximation for Deep Neural Networks by Harsh Nilesh Pathak A dissertation submitted in partial satisfaction of the requirements for the degree of Master of Science

More information

Supervised Learning for Image Segmentation

Supervised Learning for Image Segmentation Supervised Learning for Image Segmentation Raphael Meier 06.10.2016 Raphael Meier MIA 2016 06.10.2016 1 / 52 References A. Ng, Machine Learning lecture, Stanford University. A. Criminisi, J. Shotton, E.

More information

Generative Adversarial Network

Generative Adversarial Network Generative Adversarial Network Many slides from NIPS 2014 Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio Generative adversarial

More information

Motivation. Mechanisms of Physical-Layer Security. Securenets 2011 May 20th, Secrecy metrics and examples. Information-Theoretic Secrecy Metrics

Motivation. Mechanisms of Physical-Layer Security. Securenets 2011 May 20th, Secrecy metrics and examples. Information-Theoretic Secrecy Metrics Mechanisms of Physical-Layer Security Securenets 2011 May 20th, 2011 Motivation Matthieu Bloch Georgia Institute of Technology School of Electrical and Computer Engineering Secrecy metrics and examples

More information

Flow-GAN: Combining Maximum Likelihood and Adversarial Learning in Generative Models

Flow-GAN: Combining Maximum Likelihood and Adversarial Learning in Generative Models Flow-GAN: Combining Maximum Likelihood and Adversarial Learning in Generative Models Aditya Grover, Manik Dhar, Stefano Ermon Computer Science Department Stanford University {adityag, dmanik, ermon}@cs.stanford.edu

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

Conditional DCGAN For Anime Avatar Generation

Conditional DCGAN For Anime Avatar Generation Conditional DCGAN For Anime Avatar Generation Wang Hang School of Electronic Information and Electrical Engineering Shanghai Jiao Tong University Shanghai 200240, China Email: wang hang@sjtu.edu.cn Abstract

More information

What is machine learning?

What is machine learning? Machine learning, pattern recognition and statistical data modelling Lecture 12. The last lecture Coryn Bailer-Jones 1 What is machine learning? Data description and interpretation finding simpler relationship

More information

MTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen

MTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen MTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen Lecture 2: Feature selection Feature Selection feature selection (also called variable selection): choosing k < d important

More information

arxiv: v5 [cs.ai] 10 Dec 2017

arxiv: v5 [cs.ai] 10 Dec 2017 ON CONVERGENCE AND STABILITY OF GANS Naveen Kodali, Jacob Abernethy, James Hays & Zsolt Kira College of Computing Georgia Institute of Technology Atlanta, GA 30332, USA {nkodali3,prof,hays,zkira}@gatech.edu

More information

Stacking VAE and GAN for Context-aware Text-to-Image Generation

Stacking VAE and GAN for Context-aware Text-to-Image Generation 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM) Stacking VAE and GAN for Context-aware Text-to-Image Generation Chenrui Zhang and Yuxin Peng* Institute of Computer Science and

More information

Bayes Net Learning. EECS 474 Fall 2016

Bayes Net Learning. EECS 474 Fall 2016 Bayes Net Learning EECS 474 Fall 2016 Homework Remaining Homework #3 assigned Homework #4 will be about semi-supervised learning and expectation-maximization Homeworks #3-#4: the how of Graphical Models

More information

Neural Networks: promises of current research

Neural Networks: promises of current research April 2008 www.apstat.com Current research on deep architectures A few labs are currently researching deep neural network training: Geoffrey Hinton s lab at U.Toronto Yann LeCun s lab at NYU Our LISA lab

More information

Recap: Gaussian (or Normal) Distribution. Recap: Minimizing the Expected Loss. Topics of This Lecture. Recap: Maximum Likelihood Approach

Recap: Gaussian (or Normal) Distribution. Recap: Minimizing the Expected Loss. Topics of This Lecture. Recap: Maximum Likelihood Approach Truth Course Outline Machine Learning Lecture 3 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Probability Density Estimation II 2.04.205 Discriminative Approaches (5 weeks)

More information

Feature Selection for Image Retrieval and Object Recognition

Feature Selection for Image Retrieval and Object Recognition Feature Selection for Image Retrieval and Object Recognition Nuno Vasconcelos et al. Statistical Visual Computing Lab ECE, UCSD Presented by Dashan Gao Scalable Discriminant Feature Selection for Image

More information

Lab 9. Julia Janicki. Introduction

Lab 9. Julia Janicki. Introduction Lab 9 Julia Janicki Introduction My goal for this project is to map a general land cover in the area of Alexandria in Egypt using supervised classification, specifically the Maximum Likelihood and Support

More information

Optimal transport for machine learning

Optimal transport for machine learning Optimal transport for machine learning Practical sessions Rémi Flamary, Nicolas Courty, Marco Cuturi Data SCience Summer School (DS3) 2018, Paris, France 1 Course organization A day in Optimal Transport

More information

Machine Learning Lecture 3

Machine Learning Lecture 3 Many slides adapted from B. Schiele Machine Learning Lecture 3 Probability Density Estimation II 26.04.2016 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course

More information

UCLA UCLA Electronic Theses and Dissertations

UCLA UCLA Electronic Theses and Dissertations UCLA UCLA Electronic Theses and Dissertations Title Application of Generative Adversarial Network on Image Style Transformation and Image Processing Permalink https://escholarship.org/uc/item/66w654x7

More information

Machine Learning Lecture 3

Machine Learning Lecture 3 Course Outline Machine Learning Lecture 3 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Probability Density Estimation II 26.04.206 Discriminative Approaches (5 weeks) Linear

More information

Machine Learning. Unsupervised Learning. Manfred Huber

Machine Learning. Unsupervised Learning. Manfred Huber Machine Learning Unsupervised Learning Manfred Huber 2015 1 Unsupervised Learning In supervised learning the training data provides desired target output for learning In unsupervised learning the training

More information

arxiv: v1 [cs.cv] 4 Nov 2018

arxiv: v1 [cs.cv] 4 Nov 2018 Improving GAN with neighbors embedding and gradient matching Ngoc-Trung Tran, Tuan-Anh Bui, Ngai-Man Cheung ST Electronics - SUTD Cyber Security Laboratory Singapore University of Technology and Design

More information

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures: Homework Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression 3.0-3.2 Pod-cast lecture on-line Next lectures: I posted a rough plan. It is flexible though so please come with suggestions Bayes

More information

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Bryan Poling University of Minnesota Joint work with Gilad Lerman University of Minnesota The Problem of Subspace

More information

Rate-coded Restricted Boltzmann Machines for Face Recognition

Rate-coded Restricted Boltzmann Machines for Face Recognition Rate-coded Restricted Boltzmann Machines for Face Recognition Yee Whye Teh Department of Computer Science University of Toronto Toronto M5S 2Z9 Canada ywteh@cs.toronto.edu Geoffrey E. Hinton Gatsby Computational

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

Adversarial Symmetric Variational Autoencoder

Adversarial Symmetric Variational Autoencoder Adversarial Symmetric Variational Autoencoder Yunchen Pu, Weiyao Wang, Ricardo Henao, Liqun Chen, Zhe Gan, Chunyuan Li and Lawrence Carin Department of Electrical and Computer Engineering, Duke University

More information

On Information-Maximization Clustering: Tuning Parameter Selection and Analytic Solution

On Information-Maximization Clustering: Tuning Parameter Selection and Analytic Solution ICML2011 Jun. 28-Jul. 2, 2011 On Information-Maximization Clustering: Tuning Parameter Selection and Analytic Solution Masashi Sugiyama, Makoto Yamada, Manabu Kimura, and Hirotaka Hachiya Department of

More information

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes

More information

Face Transfer with Generative Adversarial Network

Face Transfer with Generative Adversarial Network Face Transfer with Generative Adversarial Network Runze Xu, Zhiming Zhou, Weinan Zhang, Yong Yu Shanghai Jiao Tong University We explore the impact of discriminators with different receptive field sizes

More information

A Fast Learning Algorithm for Deep Belief Nets

A Fast Learning Algorithm for Deep Belief Nets A Fast Learning Algorithm for Deep Belief Nets Geoffrey E. Hinton, Simon Osindero Department of Computer Science University of Toronto, Toronto, Canada Yee-Whye Teh Department of Computer Science National

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map

Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map Texture Classification by Combining Local Binary Pattern Features and a Self-Organizing Map Markus Turtinen, Topi Mäenpää, and Matti Pietikäinen Machine Vision Group, P.O.Box 4500, FIN-90014 University

More information

Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks Supplementary Material

Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks Supplementary Material Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks Supplementary Material Emily Denton Dept. of Computer Science Courant Institute New York University Soumith Chintala Arthur

More information

arxiv: v1 [cs.cv] 5 Jul 2017

arxiv: v1 [cs.cv] 5 Jul 2017 AlignGAN: Learning to Align Cross- Images with Conditional Generative Adversarial Networks Xudong Mao Department of Computer Science City University of Hong Kong xudonmao@gmail.com Qing Li Department of

More information

How to Certify the Leakage of a Chip?

How to Certify the Leakage of a Chip? How to Certify the Leakage of a Chip? F. Durvaux, F.-X. Standaert, N. Veyrat-Charvillon UCL Crypto Group, Belgium EUROCRYPT 2014, Copenhagen, Denmark Problem statement Evaluation / certification of leaking

More information

Text to Image Synthesis Using Generative Adversarial Networks

Text to Image Synthesis Using Generative Adversarial Networks Text to Image Synthesis Using Generative Adversarial Networks arxiv:1805.00676v1 [cs.cv] 2 May 2018 Cristian Bodnar Supervisor: Dr Jon Shapiro University of Manchester School of Computer Science A final

More information

arxiv: v1 [cs.cv] 7 Mar 2018

arxiv: v1 [cs.cv] 7 Mar 2018 Accepted as a conference paper at the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN) 2018 Inferencing Based on Unsupervised Learning of Disentangled

More information

Sub-GAN: An Unsupervised Generative Model via Subspaces

Sub-GAN: An Unsupervised Generative Model via Subspaces Sub-GAN: An Unsupervised Generative Model via Subspaces Jie Liang 1, Jufeng Yang 1*, Hsin-Ying Lee 2, Kai Wang 1, Ming-Hsuan Yang 2,3 1 Nankai University 2 University of California, Merced 3 Google Cloud

More information

10701 Machine Learning. Clustering

10701 Machine Learning. Clustering 171 Machine Learning Clustering What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally, finding natural groupings among

More information

Facial Expression Recognition Using Non-negative Matrix Factorization

Facial Expression Recognition Using Non-negative Matrix Factorization Facial Expression Recognition Using Non-negative Matrix Factorization Symeon Nikitidis, Anastasios Tefas and Ioannis Pitas Artificial Intelligence & Information Analysis Lab Department of Informatics Aristotle,

More information

Building Classifiers using Bayesian Networks

Building Classifiers using Bayesian Networks Building Classifiers using Bayesian Networks Nir Friedman and Moises Goldszmidt 1997 Presented by Brian Collins and Lukas Seitlinger Paper Summary The Naive Bayes classifier has reasonable performance

More information