Auto-encoder with Adversarially Regularized Latent Variables

Size: px
Start display at page:

Download "Auto-encoder with Adversarially Regularized Latent Variables"

Transcription

1 Information Engineering Express International Institute of Applied Informatics 2017, Vol.3, No.3, P Auto-encoder with Adversarially Regularized Latent Variables for Semi-Supervised Learning Ryosuke Tachibana, Takashi Matsubara and Kuniaki Uehara Abstract The amount of accessible raw data is ever-increasing in spite of the difficulty in obtaining a variety of labeled information; this makes semi-supervised learning a topic of practical importance. This paper proposes a novel regularization algorithm of an autoencoding deep neural network for semi-supervised learning. Given an input data, the deep neural network outputs the estimated label, and the remaining information called style. On the basis of the framework of a generative adversarial network, the proposed algorithm regularizes the label and the style according to a prior distribution, separating the label explicitly from the style. As a result, the deep neural network is trained to estimate correct labels by using a limitedly labeled dataset. The proposed algorithm achieved accuracy comparable with or superior to that of the existing state-of-the-art semi-supervised algorithms for the benchmark tasks, the MNIST database, and the SVHN dataset. Keywords: Auto-encoder, Deep Learning, Generative Adversarial Networks, Semi-Supervised Learning 1 Introduction While the amount of accessible data is ever-increasing, it often lacks auxiliary information such as class labels, and hand-labeling of the entire dataset is impossible or, at least, not economical. Therefore, there is considerable practical interest in semi-supervised learning. Semi-supervised learning deals with a classification task under the condition that only a small subset of a given dataset has corresponding class labels [1]. Semi-supervised algorithms utilize a large amount of unlabeled data and enable a more accurate classification than classifiers trained only with labeled data. Neural networks with deep architectures, also known as deep learning, performed impressively on a wide range of machine learning tasks, including semi-supervised learning [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]. Numerous approaches employed an unsupervised dimension reduction called autoencoder (AE) [4], consisting of two neural networks; encoder and decoder. The encoder is trained as a classifier in supervised learning, whereas the decoder is trained to reconstruct the given data, working as an additional penalty [5, 6, 7]. Some autoencoder-based approaches implemented unsupervised Bayesian inference and a generative procedure of a given dataset Graduate School of System Informatics, Kobe University, Hyogo, Japan

2 12 R. Tachibana, T. Matsubara and K. Uehara on an autoencoder, along with supervised classification [8, 9, 17]. They naturally require computation of the integral of the posterior distribution over latent variables or Monte Carlo sampling from it, thereby increasing their computational time. Other methods were based on the framework of the generative adversarial network (GAN) [10, 11, 12, 13, 14], obtaining latent representations of a given dataset by comparison with artificially generated datasets. The objective function of the GAN is an additional penalty for supervised classification. The learning procedure thereof is prone to destabilization, and thus requires careful adjustment of architectures and parameters [18, 19]. Therefore, finding a good penalty is the key to success in semi-supervised learning based on deep neural networks. Recently, a simple regularization approach for the autoencoder, called the adversarial autoencoder (AAE), was proposed for unsupervised dimension reduction [20]: It regularizes the latent representations using the framework of the GAN. This paper proposes a novel semi-supervised learning algorithm called adversarial regularization for autoencoder based on the principle of the adversarial autoencoder. Given a labeled data, the autoencoder is trained to estimate the correct label. Even without label information, the autoencoder outputs the estimated label, and a latent representation called style. The proposed algorithm regularizes the label and style according to a joint prior distribution on the basis of the framework of the GAN. As a result, a given data is decomposed explicitly into the label information contributing to the classification task, and the remaining style information. The proposed algorithm is evaluated using the semi-supervised tasks of the MNIST database and SVHN dataset as a benchmark. It achieves impressive accuracy, comparable or superior to the existing state-of-the-art semi-supervised algorithms based on deep neural networks. The preliminary and limited results are presented in a conference paper [21]. 2 Related Works While some semi-supervised learning algorithms focused on manifold learning [15, 16], many studies on deep neural networks employed a combination of unsupervised dimension reduction and supervised classification [1]. Previous studies employed autoencoder as an unsupervised dimension reduction. An autoencoder consists of two neural networks (encoder and decoder) and is trained by minimizing the objective function called the reconstruction error L rec [5, 6, 7, 8, 9]: L rec (θ q,θ p ) = E x pdata (x) [L(x, ˆx)], (1) where p data (x) denotes a given dataset; the encoder outputs the latent representation z = q(x;θ q ) given an input; the decoder reconstructs the input ˆx = p(z;θ p ); and L(, ) is a distance function, which is typically the mean squared error. Previous studies proposed a variational autoencoder (VAE), which is an implementation of unsupervised Bayesian inference and generative procedure of a given dataset on the encoder q and decoder p [17, 8, 9]. The objective function L vae is derived according to variational methods, as follows: [ ] L vae (θ q,θ p ) = E x pdata (x) D KL (q(z x;θ q ) p prior (z)) E z q(z x;θq )[log p(x z;θ p )], (2) where D KL is the Kullback-Leibler divergence and p prior (z) is the prior distribution of the latent representations z. The first and second terms can be considered as regularization terms and a general form of the reconstruction error L rec. The framework of the GAN

3 Auto-encoder with Adversarially Regularized Latent Variables for Semi-Supervised Learning 13 provides a further unsupervised dimension reduction [10]. A GAN consists of two neural networks, generator G and discriminator D. Given a latent representation z randomly chosen from a prior distribution p prior (z), the generator G outputs an artificial data ˆx. The discriminator D is trained to classify the real data x from the artificial data ˆx, whereas the generator G is trained to trick the discriminator D as follows: where minmaxl adv (θ D,θ G ), (3) θ G θ D L adv (θ D,θ G ) = E x pdata (x)[logd(x;θ D )] + E z pz (z)[log(1 D(G(z;θ G );θ D ))] (4) The output of the discriminator D denotes the probability that the input is sampled from the given dataset p data (x) rather than from the generator G. At the end of the training, the discriminator D is expected to output 0.5 for the input from both the given dataset p data (x) and the generator G, and the generator G is expected to output an artificial dataset that is distributed as the distribution of the given dataset x. When an alternative neural network is trained to estimate the latent representation z given an input artificial data ˆx, the neural network and generator G work as the encoder p and decoder q, respectively [22]. The learning procedure is known to be prone to be destabilized because the distribution shape of the given dataset x is complicated and it is difficult to model it using the generator G. The GAN requires careful adjustment of its architecture and parameters [18, 19]. The framework of the GAN can be used for regularization of the latent representation z of the autoencoder; this is called the adversarial autoencoder [20]. In this case, the encoder q of the autoencoder corresponds to the generator G of the GAN, and is expected to output latent representations q(z x;θ q ) that are distributed as in the prior distribution p prior (z). Since the prior distribution p(z) is far simpler than that of the given dataset x, its learning procedure is easier and more robust. For semi-supervised classification tasks, these unsupervised dimension reductions are combined with alternative supervised or semi-supervised learning algorithms. In some approaches, the latent representation z extracted from the input x by the encoder p is classified using another supervised or semi-supervised classification algorithm [8, 11, 14]. However, unsupervised dimension reduction is sometimes harmful to classification because it has a risk of extracting only information useless for classification, e.g., penmanship of handwritten strings in a classification of characters. Alternative approaches employ the objective function of the unsupervised dimension reduction as an additional penalty of supervised classification [5, 6, 7, 8, 9, 12, 13]. In this case, the encoder p of the autoencoder and the discriminator D of the GAN work as a classifier C and are trained with the following objective function L cls (θ c ) = E x,y plabeled (x) [ I(y = k)logc(y x;θ c ) k ], (5) where y denotes the label corresponding to the input x; p labeled (x) is the labeled subset of the given dataset p data (x); and I(cond) is the indicator function, which takes the value 1 when the condition cond is satisfied and 0 otherwise. In these approaches, the latent representation z other than the label y is separated from the label y and is sometimes called style.

4 14 R. Tachibana, T. Matsubara and K. Uehara ˆx Decoder ŷ, ẑ q(y, z x) Discriminator x Encoder d y, z p prior (y, z) Discriminator Figure 1: Diagram of the proposed autoencoder and adversarial regularization. 3 Adversarial Regularization for Semi-Supervised Learning In this section, we propose a novel adversarial regularization algorithm of semi-supervised learning based on deep neural networks. First, we propose the autoencoder depicted in Fig. 1. Given an input x, the encoder q, parameterized by θ q, outputs two latent representations denoted ŷy and ẑz. Using the softmax activation function, the summation of the elements ŷ k of the latent representation ŷy is clamped to one, while the latent representation ẑz has no activation function, and thus its elements ẑ j are distributed over the real number. The decoder p, parameterized by θ p, accepts the latent representations y and z and outputs an artificial data ˆx to reconstruct the input x. As per ordinary autoencoders, encoder q and decoder p are trained by minimizing the reconstruction error L rec (θ q,θ p ) = E x pdata (x) [ x p(ˆx q(ŷy,ẑz x;θ q );θ p ) 2]. (6) Note that, in spite of their formulations, the encoder q and the decoder p are deterministic in contrast to the variational autoencoder [17, 8, 9]. The encoder q works as a classifier and is trained by minimizing the classification error L cls (θ q ) = E x,y plabeled (x,y) [ I(y = k)logq(ŷ k x) k ]. (7) The latent representation ŷy denotes the posterior probability of the estimated label and the latent representation z represents the style of the input x. Here, we introduce the joint prior distribution p prior (y, z) of the label ŷy and the style ẑz. In contrast to the original study of the adversarial autoencoder, the label y and the style z are completely independent. The label y has only one element that takes one, with the remaining taking zero; this follows a uniform categorical distribution. Each element z k of the style z follows a normal distribution with zero-mean and a variance of 5 2. Therefore, p prior (y, z) = p prior (y)p prior (z) ( ) = 1 N y k (2 5 2 π) 1 2 exp z2 k 2 5 2, (8) where N y denotes the number of the class label. Using the framework of the GAN, the encoder q is also trained to match the joint distribution of the latent representations ŷy and ẑz to the joint prior distribution p prior (y,z). The discriminator D, parameterized by θ D,

5 Auto-encoder with Adversarially Regularized Latent Variables for Semi-Supervised Learning 15 Table 1: Results of semi-supervised classification. Test error (ave. (± std.) %) N l AE AAE [20] AR (ours) (±0.21) 1.90(±0.10) 0.98(±0.17) (±0.37) 0.97(±0.02) 1, (±0.22) 1.60(±0.08) 0.75(±0.05) 3, (±0.21) 0.70(±0.13) Table 2: Comparison of error rates (%) between the adversarial regularization (AR) and other published results on the MNIST database, with 100 labels. Test error (ave. (± std.) %) Methods N l = 100 DGM (2014) [8] 3.33(±0.14) VAT (2015) [16] 2.12 CatGAN (2016) [12] 1.39(±0.28) AAE (2015) [20] 1.90(±0.10) ADGM(10MC) (2016) [9] 0.96(±0.02) Improved-GAN (2016) [13] 0.96(±0.07) Ladder (2015) [7] 0.86(±0.89) AR (ours) 0.98(±0.17) accepts the latent representations ŷy and ẑz obtained from the encoder q or the samples y and z from the joint prior distribution p prior (y, z). The encoder q is trained by minimizing the adversarial regularization error L ar (θ D,θ q ) = E y,z p(y,z) [logd(y, z;θ D )] +E x pdata (x)[log(1 D(q(ŷy,ẑz x;θ q );θ D ))], (9) while the discriminator D is trained by maximizing the adversarial regularization error L ar Therefore, the objective function of the autoencoder with adversarial regularization is the weighted summation of all of the aforementioned errors: L = λ cls L cls + λ rec L rec + λ ar L ar, (10) where λ cls, λ rec, and λ ar are real-valued coefficients of the errors L cls, L rec, and L ar, respectively. 4 Results for Semi-Supervised Classification 4.1 Semi-Supervised Classification with Convolutional Neural Networks The proposed algorithm was evaluated on the MNIST handwritten digit database [23] and the cropped version of the Street View House Numbers (SVHN) Dataset [24]. The MNIST database is a dataset of 28-by-28 grayscale images of 70,000 handwritten digits, comprising 60,000 images for training and 10,000 images for testing. We divided the 60,000 training images into 50,000 training images and 10,000 validation images. To implement semi-supervised learning experiments, we chose N l images randomly as a labeled subset

6 16 R. Tachibana, T. Matsubara and K. Uehara Table 3: Comparison of error rates (%) between the adversarial regularization (AR) and other published results on the SVHN dataset, with 1000 labels. Test error (ave. (± std.) %) Methods N l = 1000 DGM (2014) [8] 36.02(±0.1) VAT (2015) [16] ADGM (10MC) (2016) [9] ALI (2016) [14] 19.14(±0.5) AAE (2015) [20] 17.70(±0.30) SDGM (2016) [9] 16.61(±0.24) Improved-GAN (2016) [13] 8.11(±1.3) AR (ours) 10.94(±0.27) Figure 2: Changing the label y with the clamped style z. The left digits are samples from the dataset p data (x) for the MNIST database, and the remaining digits are generated with the style z obtained from the leftmost images and the arbitrary label y. The left and right images use hyper-parameters for visualization and the best classification, respectively. p labeled (x), where each class has the same number of labeled images. We removed the label information from the remaining (50,000 N l ). We trained our deep neural networks using the training images and chose hyper-parameters according to the accuracy of the validation images. Then, the final classification error was evaluated on the test images, with the model configured by the chosen hyper-parameters. The SVHN dataset consists of 32-by-32 RGB images of digits in home numbers, comprising 604,388 images for training and 26,032 images for testing. We chose 6,000 validation images randomly from the training images. The remaining conditions were the same as those of the MNIST database. The architectures of the autoencoders were in accordance with preceding studies: [25] for the MNIST database and [11] for the SVHN dataset (see Tables 4 and 5 in Appendix for detail). We show the classification results of AE [4], AAE [20] and AR on the MNIST database and SVHN dataset (see Tables 1 and 2). The result of AAE is obtained from the original paper [20]. We searched the set of hyper-parameters that achieved the best performance on validation data, and used λ cl = 1, λ rec = 2000, and λ ar = 0 for AE; and λ cl = 1, λ rec = 1, and λ ar = 1 for AR (see Appendix in detail). We visualized the independence of the latent representations ŷy and ẑz by changing them [17, 8, 11]. We searched the best hyper-parameters for visualization and used λ cl = 1, λ rec = 500, and λ ar = 100. First, several images were sampled from the dataset p data (x) and their styles

7 Auto-encoder with Adversarially Regularized Latent Variables for Semi-Supervised Learning 17 ẑz were extracted by the encoder q. By using the extracted styles ẑz and the arbitrary label y, the decoder p generated artificial images (see Fig. 2). The generated images shared their styles, such as penmanship and font type, regardless of the digit types; this suggests that the latent representation y corresponded to digit type independently of style. The right image of Fig. 2 shows the generated images of the hyper-parameters when the model yields the best classification result. 5 Discussion The autoencoder has been studied as an unsupervised dimension reduction, contributing a penalty to the semi-supervised classification [4, 5, 6, 7]. Studies on VAE introduced a further penalty to the latent representation, based on Kullback-Leibler (KL) divergence, to match the posterior distribution q(z x) of the latent representation z obtained from each input x to the prior distribution p prior (z) [8, 9, 17]. However, the KL penalty was calculated image-by-image (see Eq. (2)), which does not guarantee a match between the distribution q(z x) over the latent representations and the prior distribution p prior (z). In contrast, attributed to the framework of the GAN, the adversarial regularization provides a stricter penalty: the adversarial regularization gives the distribution q(yŷ,ẑz x) a resemblance to the prior distribution p prior (y,z) [10, 11, 12, 13, 14]. It guarantees the independence of the label y and the style z, and thus obtains the label information y separated from the remaining style information z. This is one of the reasons why the adversarial regularization contributes to the semi-supervised classification. Recall that the hyper-parameters for visualization differ from those of classification. We consider that the separation of label and style and the reconstruction of images function well as regularization. However, they are simply penalty terms and are not always objectives compatible with classification. The discriminator D uses posterior distributions q(yŷ,zẑ x) itself, while the discriminator D draws a sample from the prior p prior (y,z). Thereby, each element ŷ k of the posterior distribution q(yŷ x) of the label y approaches 0 or 1 to mimic the sample from the prior distribution p prior (y,z), resulting in a decrease in the information H[q(yŷ x)]. This kind of penalty has been proposed in previous studies [12, 13] and is expected to increase the margin between classes and to promote the unambiguous classification. This is another reason for the success of the adversarial regularization. The discriminator D used in this study was a tiny three-layered perceptron, regardless of the size of dataset and the architecture of autoencoder, indicating that the computation time for the discriminator D and the adversarial regularization is almost negligible compared with that of the autoencoder consisting of two deep CNNs. Therefore, the total computation time is almost the same as or less than that of the existing semi-supervised learning algorithms. As shown in Tables 2 and 3, the autoencoder trained with our proposed regularization achieved accuracy comparable with or superior to that of the existing state-of-the-art semisupervised algorithms for the benchmark semi-supervised tasks. Unfortunately, Improved- GAN surpasses our proposed regularization in both tasks. However, approaches based on GAN, such as ALI and Improved-GAN, are well known to encounter the issue of mode collapse [28, 29]. A dataset used for classification problems has a multimodal distribu-tion in the data space because it is a collection of datasets, where each belongs to one of the classes. However, GAN is prone to fail to generate artificial data with a multimodal distribution; more specifically, the generator collapses, generating only a certain mode sample or a

8 18 R. Tachibana, T. Matsubara and K. Uehara Table 4: Architecture of Autoencoder for the MNIST database. encoder q decoder p grayscale image 60-D latent representation conv. (c32, k5, s1), BN, ReLU fc (n1024), BN, ReLU max-pool. (s2) fc (n2048), BN, ReLU conv. (c64, k5, s1), BN, ReLU up. (s2) max-pool. (s2) deconv. (c64, k5, s1), BN, ReLU conv. (c128, k5, s1), BN, ReLU up. (s2) max-pool. (s2) deconv. (c32, k5, s1), BN, ReLU fc (n1024), BN, ReLU, dropout(p0.5) up. (s2) fc (n60), BN, ReLU deconv. (c1, k5, s1), BN, ReLU small family of very similar samples. While GAN is required to find fine hyper-parameters to model a multimodal distribution, this is difficult in the case of semi-supervised learning because of the limited knowledge of the dataset. Therefore, Improved-GAN s high performance in the semi-supervised classification of MNIST does not guarantee its high performance for unknown datasets used in practical tasks. Conversely, our proposed regularization potentially overcomes this issue because it generates images in the manner of AE (which is free from mode collapse) and only uses GAN to regularize the latent variables, which follow a unimodal distribution (i.e., Gaussian distribution and uniform distribution.) Therefore, we consider our method to be more robust than other GAN-based approaches. A more detailed comparison will be conducted in future. In addition, the discriminator D used in this study was a tiny three-layered perceptron, regardless of the size of dataset and the architecture of the autoencoder. This indicates that the computation time for the discriminator D and our proposed regularization is almost negligible compared with that of the autoencoder consisting of two deep CNNs. Therefore, the total computation time is considerably less than that of the semi-supervised learning algorithms based on deep generative models (e.g., DGM, ADGM, and SDGM), which require a minimum of three deep CNNs. 6 Conclusion This paper proposed the adversarial regularization of the joint distribution of latent representations of an autoencoder for semi-supervised classification. The adversarial regularization provides a penalty that divides the latent representations into the label information and the remaining style information. Therefore, the autoencoder trained with the regularization achieved an accuracy comparable or superior to the existing state-of-the-art semi-supervised algorithms for the benchmark semi-supervised tasks. Acknowledgments This work was partially supported by the JSPS KAKENHI (16K12487) and the SEI Group CSR Foundation.

9 Auto-encoder with Adversarially Regularized Latent Variables for Semi-Supervised Learning 19 Table 5: Architecture of Autoencoder for the SVHN dataset. encoder q decoder p RGB image 60-D latent representation conv. (c64, k5, s1), BN, lrelu fc (n ), BN, ReLU conv. (c128, k4, s2), BN, lrelu deconv. (c256, k4, s1), BN, lrelu conv. (c256, k4, s1), BN, lrelu deconv. (c128, k4, s2), BN, lrelu conv. (c512, k4, s2), BN, lrelu deconv. (c64, k4, s1), BN, lrelu fc (n60) deconv. (c3, k4, s2), BN, lrelu A Details of Experimental Settings The appendix provides the details of the experimental settings of our results. The architectures of the autoencoder used for the MNIST database and SVHN dataset are shown in Table 4 and Table 5, respectively. conv. and deconv. denote convolution and fractionally strided convolution, respectively, outputing c feature maps with a k k kernel and stride of s; max-pool. and up. denote max-pooling and upscaling, respectively, with a stride of s; fc denotes matrix multiplication outputting an n-dimensional vector; BN denotes batch normalization; ReLU and sigmoid are activation functions; and lrelu is the leaky ReLU activation function with a slope of The autoencoder and discriminator were trained using the Adam optimization algorithm [27] with a weight decay of , where the parameter α was set to α = for the experiments detailed in Section 4.1 The MNIST database and SVHN dataset had batch sizes of 500 and 100, respectively. The coefficients λ rec and λ ar of the errors L rec and L ar, respectively, were grid-seared over the range of {..., 0.1, 0.2, 0.5, 1.0, 2.0, 5.0, 10.0,...}, where the coefficient λ cls was set to 1. References [1] J. Lasserre et al., IEEE, vol. 1, no. 6., pp , [2] Y. LeCun, et al., Nature, vol. 521, no. 7553, pp , [3] J. Schmidhuber, Neural Networks, vol. 61, pp , [4] G. E. Hinton and R. R. Salakhutdinov, Science, vol. 313, no. 5786, pp , [5] M. Ranzato and M. Szummer, ICML, pp , [6] X. Zhao and P. a. Robinson, Journal of Computational Neuroscience, pp , [7] A. Rasmus et al., NIPS, pp , [8] D. P. Kingma, et al., NIPS, pp , [9] L. Maaløe et al., ICML, vol. 48, pp. 1 5, [10] I. J. Goodfellow et al., NIPS, pp , [11] A. Radford, et al., ICLR, pp. 1 16, 2016.

10 20 R. Tachibana, T. Matsubara and K. Uehara [12] J. T. Springenberg, arxiv: , [13] T. Salimans et al., arxiv: , [14] V. Dumoulin et al., arxiv: , [15] S. Rifai et al., NIPS, pp , [16] T. Miyato et al., ICLR, pp. 1 18, [17] D. P. Kingma and M. Welling, ICLR, pp. 1 14, [18] S. Nowozin, et al., arxiv: , [19] M. Arjovsky and L. Bottou, NIPS, pp. 1 16, [20] A. Makhzani et al., arxiv: , [21] R. Tachibana, et al., ICIS, pp , 2016 [22] A. B. L. Larsen, et al., ICML, [23] THE MNIST DATABASE of handwritten digits. url: exdb/mnist/ [24] The Street View House Numbers (SVHN) Dataset. url: stanford.edu/housenumbers/ [25] Deep MNIST for Experts, in TensorFlow Turorials. url: tensorflow.org/tutorials/mnist/pros/ [26] B. Cheung et al., ICLR, [27] D. Kingma and J. Ba, ICLR, [28] X. Mao et al., arxiv: , [29] L.Metz et al., ICLR, 2017.

Adversarially Learned Inference

Adversarially Learned Inference Institut des algorithmes d apprentissage de Montréal Adversarially Learned Inference Aaron Courville CIFAR Fellow Université de Montréal Joint work with: Vincent Dumoulin, Ishmael Belghazi, Olivier Mastropietro,

More information

Alternatives to Direct Supervision

Alternatives to Direct Supervision CreativeAI: Deep Learning for Graphics Alternatives to Direct Supervision Niloy Mitra Iasonas Kokkinos Paul Guerrero Nils Thuerey Tobias Ritschel UCL UCL UCL TUM UCL Timetable Theory and Basics State of

More information

Deep Generative Models Variational Autoencoders

Deep Generative Models Variational Autoencoders Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative

More information

Unsupervised Learning

Unsupervised Learning Deep Learning for Graphics Unsupervised Learning Niloy Mitra Iasonas Kokkinos Paul Guerrero Vladimir Kim Kostas Rematas Tobias Ritschel UCL UCL/Facebook UCL Adobe Research U Washington UCL Timetable Niloy

More information

SEMI-SUPERVISED LEARNING WITH GANS:

SEMI-SUPERVISED LEARNING WITH GANS: SEMI-SUPERVISED LEARNING WITH GANS: REVISITING MANIFOLD REGULARIZATION Bruno Lecouat,1,2 Chuan-Sheng Foo,2, Houssam Zenati 2,3, Vijay R. Chandrasekhar 2,4 1 Télécom ParisTech, bruno.lecouat@gmail.com.

More information

Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN.

Autoencoders. Stephen Scott. Introduction. Basic Idea. Stacked AE. Denoising AE. Sparse AE. Contractive AE. Variational AE GAN. Stacked Denoising Sparse Variational (Adapted from Paul Quint and Ian Goodfellow) Stacked Denoising Sparse Variational Autoencoding is training a network to replicate its input to its output Applications:

More information

Deep Manga Colorization with Color Style Extraction by Conditional Adversarially Learned Inference

Deep Manga Colorization with Color Style Extraction by Conditional Adversarially Learned Inference Information Engineering Express International Institute of Applied Informatics 2017, Vol.3, No.4, P.55-66 Deep Manga Colorization with Color Style Extraction by Conditional Adversarially Learned Inference

More information

Implicit generative models: dual vs. primal approaches

Implicit generative models: dual vs. primal approaches Implicit generative models: dual vs. primal approaches Ilya Tolstikhin MPI for Intelligent Systems ilya@tue.mpg.de Machine Learning Summer School 2017 Tübingen, Germany Contents 1. Unsupervised generative

More information

Auto-Encoding Variational Bayes

Auto-Encoding Variational Bayes Auto-Encoding Variational Bayes Diederik P (Durk) Kingma, Max Welling University of Amsterdam Ph.D. Candidate, advised by Max Durk Kingma D.P. Kingma Max Welling Problem class Directed graphical model:

More information

When Variational Auto-encoders meet Generative Adversarial Networks

When Variational Auto-encoders meet Generative Adversarial Networks When Variational Auto-encoders meet Generative Adversarial Networks Jianbo Chen Billy Fang Cheng Ju 14 December 2016 Abstract Variational auto-encoders are a promising class of generative models. In this

More information

Denoising Adversarial Autoencoders

Denoising Adversarial Autoencoders Denoising Adversarial Autoencoders Antonia Creswell BICV Imperial College London Anil Anthony Bharath BICV Imperial College London Email: ac2211@ic.ac.uk arxiv:1703.01220v4 [cs.cv] 4 Jan 2018 Abstract

More information

19: Inference and learning in Deep Learning

19: Inference and learning in Deep Learning 10-708: Probabilistic Graphical Models 10-708, Spring 2017 19: Inference and learning in Deep Learning Lecturer: Zhiting Hu Scribes: Akash Umakantha, Ryan Williamson 1 Classes of Deep Generative Models

More information

arxiv: v1 [stat.ml] 10 Dec 2018

arxiv: v1 [stat.ml] 10 Dec 2018 1st Symposium on Advances in Approximate Bayesian Inference, 2018 1 7 Disentangled Dynamic Representations from Unordered Data arxiv:1812.03962v1 [stat.ml] 10 Dec 2018 Leonhard Helminger Abdelaziz Djelouah

More information

GAN Frontiers/Related Methods

GAN Frontiers/Related Methods GAN Frontiers/Related Methods Improving GAN Training Improved Techniques for Training GANs (Salimans, et. al 2016) CSC 2541 (07/10/2016) Robin Swanson (robin@cs.toronto.edu) Training GANs is Difficult

More information

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x Deep Learning 4 Autoencoder, Attention (spatial transformer), Multi-modal learning, Neural Turing Machine, Memory Networks, Generative Adversarial Net Jian Li IIIS, Tsinghua Autoencoder Autoencoder Unsupervised

More information

Clustering and Unsupervised Anomaly Detection with l 2 Normalized Deep Auto-Encoder Representations

Clustering and Unsupervised Anomaly Detection with l 2 Normalized Deep Auto-Encoder Representations Clustering and Unsupervised Anomaly Detection with l 2 Normalized Deep Auto-Encoder Representations Caglar Aytekin, Xingyang Ni, Francesco Cricri and Emre Aksu Nokia Technologies, Tampere, Finland Corresponding

More information

Variational Autoencoders. Sargur N. Srihari

Variational Autoencoders. Sargur N. Srihari Variational Autoencoders Sargur N. srihari@cedar.buffalo.edu Topics 1. Generative Model 2. Standard Autoencoder 3. Variational autoencoders (VAE) 2 Generative Model A variational autoencoder (VAE) is a

More information

Lab meeting (Paper review session) Stacked Generative Adversarial Networks

Lab meeting (Paper review session) Stacked Generative Adversarial Networks Lab meeting (Paper review session) Stacked Generative Adversarial Networks 2017. 02. 01. Saehoon Kim (Ph. D. candidate) Machine Learning Group Papers to be covered Stacked Generative Adversarial Networks

More information

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

Lecture 21 : A Hybrid: Deep Learning and Graphical Models 10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation

More information

Deep Generative Models and a Probabilistic Programming Library

Deep Generative Models and a Probabilistic Programming Library Deep Generative Models and a Probabilistic Programming Library Discriminative (Deep) Learning Learn a (differentiable) function mapping from input to output x f(x; θ) y Gradient back-propagation Generative

More information

arxiv: v1 [cs.ne] 11 Jun 2018

arxiv: v1 [cs.ne] 11 Jun 2018 Generative Adversarial Network Architectures For Image Synthesis Using Capsule Networks arxiv:1806.03796v1 [cs.ne] 11 Jun 2018 Yash Upadhyay University of Minnesota, Twin Cities Minneapolis, MN, 55414

More information

Deep Hybrid Discriminative-Generative Models for Semi-Supervised Learning

Deep Hybrid Discriminative-Generative Models for Semi-Supervised Learning Volodymyr Kuleshov 1 Stefano Ermon 1 Abstract We propose a framework for training deep probabilistic models that interpolate between discriminative and generative approaches. Unlike previously proposed

More information

arxiv: v1 [cs.cv] 17 Nov 2016

arxiv: v1 [cs.cv] 17 Nov 2016 Inverting The Generator Of A Generative Adversarial Network arxiv:1611.05644v1 [cs.cv] 17 Nov 2016 Antonia Creswell BICV Group Bioengineering Imperial College London ac2211@ic.ac.uk Abstract Anil Anthony

More information

Generating Images with Perceptual Similarity Metrics based on Deep Networks

Generating Images with Perceptual Similarity Metrics based on Deep Networks Generating Images with Perceptual Similarity Metrics based on Deep Networks Alexey Dosovitskiy and Thomas Brox University of Freiburg {dosovits, brox}@cs.uni-freiburg.de Abstract We propose a class of

More information

DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION

DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION Yen-Cheng Liu 1, Wei-Chen Chiu 2, Sheng-De Wang 1, and Yu-Chiang Frank Wang 1 1 Graduate Institute of Electrical Engineering,

More information

DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION

DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION 2017 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 25 28, 2017, TOKYO, JAPAN DOMAIN-ADAPTIVE GENERATIVE ADVERSARIAL NETWORKS FOR SKETCH-TO-PHOTO INVERSION Yen-Cheng Liu 1,

More information

GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin

GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin GENERATIVE ADVERSARIAL NETWORKS (GAN) Presented by Omer Stein and Moran Rubin GENERATIVE MODEL Given a training dataset, x, try to estimate the distribution, Pdata(x) Explicitly or Implicitly (GAN) Explicitly

More information

Autoencoders, denoising autoencoders, and learning deep networks

Autoencoders, denoising autoencoders, and learning deep networks 4 th CiFAR Summer School on Learning and Vision in Biology and Engineering Toronto, August 5-9 2008 Autoencoders, denoising autoencoders, and learning deep networks Part II joint work with Hugo Larochelle,

More information

Capsule Networks. Eric Mintun

Capsule Networks. Eric Mintun Capsule Networks Eric Mintun Motivation An improvement* to regular Convolutional Neural Networks. Two goals: Replace max-pooling operation with something more intuitive. Keep more info about an activated

More information

arxiv: v2 [cs.lg] 25 May 2016

arxiv: v2 [cs.lg] 25 May 2016 Adversarial Autoencoders Alireza Makhzani University of Toronto makhzani@psi.toronto.edu Jonathon Shlens & Navdeep Jaitly Google Brain {shlens,ndjaitly}@google.com arxiv:1511.05644v2 [cs.lg] 25 May 2016

More information

arxiv: v1 [cs.cv] 7 Mar 2018

arxiv: v1 [cs.cv] 7 Mar 2018 Accepted as a conference paper at the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN) 2018 Inferencing Based on Unsupervised Learning of Disentangled

More information

IMPLICIT AUTOENCODERS

IMPLICIT AUTOENCODERS IMPLICIT AUTOENCODERS Anonymous authors Paper under double-blind review ABSTRACT In this paper, we describe the implicit autoencoder (IAE), a generative autoencoder in which both the generative path and

More information

Deep Learning for Computer Vision II

Deep Learning for Computer Vision II IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L

More information

Bidirectional GAN. Adversarially Learned Inference (ICLR 2017) Adversarial Feature Learning (ICLR 2017)

Bidirectional GAN. Adversarially Learned Inference (ICLR 2017) Adversarial Feature Learning (ICLR 2017) Bidirectional GAN Adversarially Learned Inference (ICLR 2017) V. Dumoulin 1, I. Belghazi 1, B. Poole 2, O. Mastropietro 1, A. Lamb 1, M. Arjovsky 3 and A. Courville 1 1 Universite de Montreal & 2 Stanford

More information

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17 Improving Generative Adversarial Networks with Denoising Feature Matching David Warde-Farley 1 Yoshua Bengio 1 1 University of Montreal, ICLR,2017 Presenter: Bargav Jayaraman Outline 1 Introduction 2 Background

More information

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all

More information

Semantic Segmentation. Zhongang Qi

Semantic Segmentation. Zhongang Qi Semantic Segmentation Zhongang Qi qiz@oregonstate.edu Semantic Segmentation "Two men riding on a bike in front of a building on the road. And there is a car." Idea: recognizing, understanding what's in

More information

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models [Supplemental Materials] 1. Network Architecture b ref b ref +1 We now describe the architecture of the networks

More information

Vulnerability of machine learning models to adversarial examples

Vulnerability of machine learning models to adversarial examples Vulnerability of machine learning models to adversarial examples Petra Vidnerová Institute of Computer Science The Czech Academy of Sciences Hora Informaticae 1 Outline Introduction Works on adversarial

More information

MoonRiver: Deep Neural Network in C++

MoonRiver: Deep Neural Network in C++ MoonRiver: Deep Neural Network in C++ Chung-Yi Weng Computer Science & Engineering University of Washington chungyi@cs.washington.edu Abstract Artificial intelligence resurges with its dramatic improvement

More information

Auxiliary Deep Generative Models

Auxiliary Deep Generative Models Downloaded from orbit.dtu.dk on: Dec 12, 2018 Auxiliary Deep Generative Models Maaløe, Lars; Sønderby, Casper Kaae; Sønderby, Søren Kaae; Winther, Ole Published in: Proceedings of the 33rd International

More information

Smooth Neighbors on Teacher Graphs for Semi-supervised Learning

Smooth Neighbors on Teacher Graphs for Semi-supervised Learning Smooth Neighbors on Teacher Graphs for Semi-supervised Learning Yucen Luo, Jun Zhu, Mengxi Li, Yong Ren, Bo Zhang Department of Computer Science & Technology, Tsinghua University, Beijing, China Department

More information

Virtual Adversarial Ladder Networks for Semi-Supervised Learning

Virtual Adversarial Ladder Networks for Semi-Supervised Learning Virtual Adversarial Ladder Networks for Semi-Supervised Learning Saki Shinoda 1, Daniel E. Worrall 2 & Gabriel J. Brostow 2 Computer Science Department University College London United Kingdom 1 saki.shinoda.16@ucl.ac.uk

More information

Autoencoding Beyond Pixels Using a Learned Similarity Metric

Autoencoding Beyond Pixels Using a Learned Similarity Metric Autoencoding Beyond Pixels Using a Learned Similarity Metric International Conference on Machine Learning, 2016 Anders Boesen Lindbo Larsen, Hugo Larochelle, Søren Kaae Sønderby, Ole Winther Technical

More information

Learning Discrete Representations via Information Maximizing Self-Augmented Training

Learning Discrete Representations via Information Maximizing Self-Augmented Training A. Relation to Denoising and Contractive Auto-encoders Our method is related to denoising auto-encoders (Vincent et al., 2008). Auto-encoders maximize a lower bound of mutual information (Cover & Thomas,

More information

Energy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt

Energy Based Models, Restricted Boltzmann Machines and Deep Networks. Jesse Eickholt Energy Based Models, Restricted Boltzmann Machines and Deep Networks Jesse Eickholt ???? Who s heard of Energy Based Models (EBMs) Restricted Boltzmann Machines (RBMs) Deep Belief Networks Auto-encoders

More information

Inverting The Generator Of A Generative Adversarial Network

Inverting The Generator Of A Generative Adversarial Network 1 Inverting The Generator Of A Generative Adversarial Network Antonia Creswell and Anil A Bharath, Imperial College London arxiv:1802.05701v1 [cs.cv] 15 Feb 2018 Abstract Generative adversarial networks

More information

maagma Modified-Adversarial-Autoencoding Generator with Multiple Adversaries: Optimizing Multiple Learning Objectives for Image Generation with GANs

maagma Modified-Adversarial-Autoencoding Generator with Multiple Adversaries: Optimizing Multiple Learning Objectives for Image Generation with GANs maagma Modified-Adversarial-Autoencoding Generator with Multiple Adversaries: Optimizing Multiple Learning Objectives for Image Generation with GANs Sahil Chopra Stanford University 353 Serra Mall, Stanford

More information

arxiv: v1 [cs.lg] 12 Jun 2016

arxiv: v1 [cs.lg] 12 Jun 2016 InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets arxiv:1606.03657v1 [cs.lg] 12 Jun 2016 Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever,

More information

arxiv: v2 [cs.cv] 6 Dec 2017

arxiv: v2 [cs.cv] 6 Dec 2017 Arbitrary Facial Attribute Editing: Only Change What You Want arxiv:1711.10678v2 [cs.cv] 6 Dec 2017 Zhenliang He 1,2 Wangmeng Zuo 4 Meina Kan 1 Shiguang Shan 1,3 Xilin Chen 1 1 Key Lab of Intelligent Information

More information

arxiv: v2 [cs.lg] 17 Dec 2018

arxiv: v2 [cs.lg] 17 Dec 2018 Lu Mi 1 * Macheng Shen 2 * Jingzhao Zhang 2 * 1 MIT CSAIL, 2 MIT LIDS {lumi, macshen, jzhzhang}@mit.edu The authors equally contributed to this work. This report was a part of the class project for 6.867

More information

DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE MACHINE LEARNING & DATA MINING - LECTURE 17

DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE MACHINE LEARNING & DATA MINING - LECTURE 17 DEEP LEARNING PART THREE - DEEP GENERATIVE MODELS CS/CNS/EE 155 - MACHINE LEARNING & DATA MINING - LECTURE 17 GENERATIVE MODELS DATA 3 DATA 4 example 1 DATA 5 example 2 DATA 6 example 3 DATA 7 number of

More information

Generative Modeling with Convolutional Neural Networks. Denis Dus Data Scientist at InData Labs

Generative Modeling with Convolutional Neural Networks. Denis Dus Data Scientist at InData Labs Generative Modeling with Convolutional Neural Networks Denis Dus Data Scientist at InData Labs What we will discuss 1. 2. 3. 4. Discriminative vs Generative modeling Convolutional Neural Networks How to

More information

arxiv: v1 [cs.cv] 16 Jul 2017

arxiv: v1 [cs.cv] 16 Jul 2017 enerative adversarial network based on resnet for conditional image restoration Paper: jc*-**-**-****: enerative Adversarial Network based on Resnet for Conditional Image Restoration Meng Wang, Huafeng

More information

Dist-GAN: An Improved GAN using Distance Constraints

Dist-GAN: An Improved GAN using Distance Constraints Dist-GAN: An Improved GAN using Distance Constraints Ngoc-Trung Tran [0000 0002 1308 9142], Tuan-Anh Bui [0000 0003 4123 262], and Ngai-Man Cheung [0000 0003 0135 3791] ST Electronics - SUTD Cyber Security

More information

arxiv: v2 [cs.lg] 7 Feb 2019

arxiv: v2 [cs.lg] 7 Feb 2019 Implicit Autoencoders Alireza Makhzani Vector Institute for Artificial Intelligence University of Toronto makhzani@vectorinstitute.ai arxiv:1805.09804v2 [cs.lg 7 Feb 2019 Abstract In this paper, we describe

More information

arxiv: v1 [cs.lg] 24 Jan 2019

arxiv: v1 [cs.lg] 24 Jan 2019 Jaehoon Cha Kyeong Soo Kim Sanghuyk Lee arxiv:9.879v [cs.lg] Jan 9 Abstract Noting the importance of the latent variables in inference and learning, we propose a novel framework for autoencoders based

More information

arxiv: v2 [cs.lg] 2 Dec 2018

arxiv: v2 [cs.lg] 2 Dec 2018 Disentangled Variational Auto-Encoder for Semi-supervised Learning Yang Li a, Quan Pan a, Suhang Wang c, Haiyun Peng b, Tao Yang a, Erik Cambria b, a School of Automation, Northwestern Polytechnical University

More information

Neural Networks: promises of current research

Neural Networks: promises of current research April 2008 www.apstat.com Current research on deep architectures A few labs are currently researching deep neural network training: Geoffrey Hinton s lab at U.Toronto Yann LeCun s lab at NYU Our LISA lab

More information

Lecture 19: Generative Adversarial Networks

Lecture 19: Generative Adversarial Networks Lecture 19: Generative Adversarial Networks Roger Grosse 1 Introduction Generative modeling is a type of machine learning where the aim is to model the distribution that a given set of data (e.g. images,

More information

Generalized Loss-Sensitive Adversarial Learning with Manifold Margins

Generalized Loss-Sensitive Adversarial Learning with Manifold Margins Generalized Loss-Sensitive Adversarial Learning with Manifold Margins Marzieh Edraki and Guo-Jun Qi Laboratory for MAchine Perception and LEarning (MAPLE) http://maple.cs.ucf.edu/ University of Central

More information

Dynamic Routing Between Capsules

Dynamic Routing Between Capsules Report Explainable Machine Learning Dynamic Routing Between Capsules Author: Michael Dorkenwald Supervisor: Dr. Ullrich Köthe 28. Juni 2018 Inhaltsverzeichnis 1 Introduction 2 2 Motivation 2 3 CapusleNet

More information

An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks

An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks An Empirical Study of Generative Adversarial Networks for Computer Vision Tasks Report for Undergraduate Project - CS396A Vinayak Tantia (Roll No: 14805) Guide: Prof Gaurav Sharma CSE, IIT Kanpur, India

More information

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction Akarsh Pokkunuru EECS Department 03-16-2017 Contractive Auto-Encoders: Explicit Invariance During Feature Extraction 1 AGENDA Introduction to Auto-encoders Types of Auto-encoders Analysis of different

More information

Visual Recommender System with Adversarial Generator-Encoder Networks

Visual Recommender System with Adversarial Generator-Encoder Networks Visual Recommender System with Adversarial Generator-Encoder Networks Bowen Yao Stanford University 450 Serra Mall, Stanford, CA 94305 boweny@stanford.edu Yilin Chen Stanford University 450 Serra Mall

More information

Deep generative models of natural images

Deep generative models of natural images Spring 2016 1 Motivation 2 3 Variational autoencoders Generative adversarial networks Generative moment matching networks Evaluating generative models 4 Outline 1 Motivation 2 3 Variational autoencoders

More information

arxiv: v1 [cs.lg] 16 Jan 2013

arxiv: v1 [cs.lg] 16 Jan 2013 Stochastic Pooling for Regularization of Deep Convolutional Neural Networks arxiv:131.3557v1 [cs.lg] 16 Jan 213 Matthew D. Zeiler Department of Computer Science Courant Institute, New York University zeiler@cs.nyu.edu

More information

arxiv: v1 [stat.ml] 11 Feb 2018

arxiv: v1 [stat.ml] 11 Feb 2018 Paul K. Rubenstein Bernhard Schölkopf Ilya Tolstikhin arxiv:80.0376v [stat.ml] Feb 08 Abstract We study the role of latent space dimensionality in Wasserstein auto-encoders (WAEs). Through experimentation

More information

Model Generalization and the Bias-Variance Trade-Off

Model Generalization and the Bias-Variance Trade-Off Charu C. Aggarwal IBM T J Watson Research Center Yorktown Heights, NY Model Generalization and the Bias-Variance Trade-Off Neural Networks and Deep Learning, Springer, 2018 Chapter 4, Section 4.1-4.2 What

More information

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao Introduction to GAN Generative Adversarial Networks Junheng(Jeff) Hao Adversarial Training is the coolest thing since sliced bread. -- Yann LeCun Roadmap 1. Generative Modeling 2. GAN 101: What is GAN?

More information

arxiv: v4 [cs.lg] 5 Nov 2017

arxiv: v4 [cs.lg] 5 Nov 2017 Triple Generative Adversarial Nets Chongxuan Li, Kun Xu, Jun Zhu, Bo Zhang Dept. of Comp. Sci. & Tech., TNList Lab, State Key Lab of Intell. Tech. & Sys., Center for Bio-Inspired Computing Research, Tsinghua

More information

Learning Two-Layer Contractive Encodings

Learning Two-Layer Contractive Encodings In Proceedings of International Conference on Artificial Neural Networks (ICANN), pp. 620-628, September 202. Learning Two-Layer Contractive Encodings Hannes Schulz and Sven Behnke Rheinische Friedrich-Wilhelms-Universität

More information

InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets

InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, Pieter Abbeel UC Berkeley, Department

More information

Stacked Denoising Autoencoders for Face Pose Normalization

Stacked Denoising Autoencoders for Face Pose Normalization Stacked Denoising Autoencoders for Face Pose Normalization Yoonseop Kang 1, Kang-Tae Lee 2,JihyunEun 2, Sung Eun Park 2 and Seungjin Choi 1 1 Department of Computer Science and Engineering Pohang University

More information

IN practical regression and classification problems, the

IN practical regression and classification problems, the 1 Virtual Adversarial Training: a Regularization Method for Supervised and Semi-supervised Learning Takeru Miyato,,, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii, arxiv:1704.03976v1 [stat.ml] 13 Apr

More information

Implicit Mixtures of Restricted Boltzmann Machines

Implicit Mixtures of Restricted Boltzmann Machines Implicit Mixtures of Restricted Boltzmann Machines Vinod Nair and Geoffrey Hinton Department of Computer Science, University of Toronto 10 King s College Road, Toronto, M5S 3G5 Canada {vnair,hinton}@cs.toronto.edu

More information

Restricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version

Restricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version Shallow vs. deep networks Restricted Boltzmann Machines Shallow: one hidden layer Features can be learned more-or-less independently Arbitrary function approximator (with enough hidden units) Deep: two

More information

Generative Adversarial Networks (GANs) Ian Goodfellow, Research Scientist MLSLP Keynote, San Francisco

Generative Adversarial Networks (GANs) Ian Goodfellow, Research Scientist MLSLP Keynote, San Francisco Generative Adversarial Networks (GANs) Ian Goodfellow, Research Scientist MLSLP Keynote, San Francisco 2016-09-13 Generative Modeling Density estimation Sample generation Training examples Model samples

More information

To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine

To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine 2014 22nd International Conference on Pattern Recognition To be Bernoulli or to be Gaussian, for a Restricted Boltzmann Machine Takayoshi Yamashita, Masayuki Tanaka, Eiji Yoshida, Yuji Yamauchi and Hironobu

More information

A Fast Learning Algorithm for Deep Belief Nets

A Fast Learning Algorithm for Deep Belief Nets A Fast Learning Algorithm for Deep Belief Nets Geoffrey E. Hinton, Simon Osindero Department of Computer Science University of Toronto, Toronto, Canada Yee-Whye Teh Department of Computer Science National

More information

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin

More information

Classification of 1D-Signal Types Using Semi-Supervised Deep Learning

Classification of 1D-Signal Types Using Semi-Supervised Deep Learning UNIVERSITY OF ZAGREB FACULTY OF ELECTRICAL ENGINEERING AND COMPUTING MASTER THESIS No. 1414 Classification of 1D-Signal Types Using Semi-Supervised Deep Learning Tomislav Šebrek Zagreb, June 2017. I

More information

Data Set Extension with Generative Adversarial Nets

Data Set Extension with Generative Adversarial Nets Department of Artificial Intelligence University of Groningen, The Netherlands Data Set Extension with Generative Adversarial Nets Master s Thesis Luuk Boulogne S2366681 Primary supervisor: Secondary supervisor:

More information

Overall Description. Goal: to improve spatial invariance to the input data. Translation, Rotation, Scale, Clutter, Elastic

Overall Description. Goal: to improve spatial invariance to the input data. Translation, Rotation, Scale, Clutter, Elastic Philippe Giguère Overall Description Goal: to improve spatial invariance to the input data Translation, Rotation, Scale, Clutter, Elastic How: add a learnable module which explicitly manipulate spatially

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

GRADIENT-BASED OPTIMIZATION OF NEURAL

GRADIENT-BASED OPTIMIZATION OF NEURAL Workshop track - ICLR 28 GRADIENT-BASED OPTIMIZATION OF NEURAL NETWORK ARCHITECTURE Will Grathwohl, Elliot Creager, Seyed Kamyar Seyed Ghasemipour, Richard Zemel Department of Computer Science University

More information

AdaDepth: Unsupervised Content Congruent Adaptation for Depth Estimation

AdaDepth: Unsupervised Content Congruent Adaptation for Depth Estimation AdaDepth: Unsupervised Content Congruent Adaptation for Depth Estimation Introduction Supplementary material In the supplementary material, we present additional qualitative results of the proposed AdaDepth

More information

Image Restoration with Deep Generative Models

Image Restoration with Deep Generative Models Image Restoration with Deep Generative Models Raymond A. Yeh *, Teck-Yian Lim *, Chen Chen, Alexander G. Schwing, Mark Hasegawa-Johnson, Minh N. Do Department of Electrical and Computer Engineering, University

More information

Progress on Generative Adversarial Networks

Progress on Generative Adversarial Networks Progress on Generative Adversarial Networks Wangmeng Zuo Vision Perception and Cognition Centre Harbin Institute of Technology Content Image generation: problem formulation Three issues about GAN Discriminate

More information

Day 3 Lecture 1. Unsupervised Learning

Day 3 Lecture 1. Unsupervised Learning Day 3 Lecture 1 Unsupervised Learning Semi-supervised and transfer learning Myth: you can t do deep learning unless you have a million labelled examples for your problem. Reality You can learn useful representations

More information

Deep Learning on Graphs

Deep Learning on Graphs Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016

More information

Introduction to Deep Learning

Introduction to Deep Learning ENEE698A : Machine Learning Seminar Introduction to Deep Learning Raviteja Vemulapalli Image credit: [LeCun 1998] Resources Unsupervised feature learning and deep learning (UFLDL) tutorial (http://ufldl.stanford.edu/wiki/index.php/ufldl_tutorial)

More information

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Tianyu Wang Australia National University, Colledge of Engineering and Computer Science u@anu.edu.au Abstract. Some tasks,

More information

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao

Introduction to GAN. Generative Adversarial Networks. Junheng(Jeff) Hao Introduction to GAN Generative Adversarial Networks Junheng(Jeff) Hao Adversarial Training is the coolest thing since sliced bread. -- Yann LeCun Roadmap 1. Generative Modeling 2. GAN 101: What is GAN?

More information

RECENTLY, deep neural network models have seen

RECENTLY, deep neural network models have seen IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 1 Semisupervised Text Classification by Variational Autoencoder Weidi Xu and Ying Tan, Senior Member, IEEE Abstract Semisupervised text classification

More information

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal Object Detection Lecture 10.3 - Introduction to deep learning (CNN) Idar Dyrdal Deep Learning Labels Computational models composed of multiple processing layers (non-linear transformations) Used to learn

More information

Study of Residual Networks for Image Recognition

Study of Residual Networks for Image Recognition Study of Residual Networks for Image Recognition Mohammad Sadegh Ebrahimi Stanford University sadegh@stanford.edu Hossein Karkeh Abadi Stanford University hosseink@stanford.edu Abstract Deep neural networks

More information

The Multi-Entity Variational Autoencoder

The Multi-Entity Variational Autoencoder The Multi-Entity Variational Autoencoder Charlie Nash 1,2, S. M. Ali Eslami 2, Chris Burgess 2, Irina Higgins 2, Daniel Zoran 2, Theophane Weber 2, Peter Battaglia 2 1 Edinburgh University 2 DeepMind Abstract

More information

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Tutorial on Keras CAP 6412 - ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Deep learning packages TensorFlow Google PyTorch Facebook AI research Keras Francois Chollet (now at Google) Chainer Company

More information

Variational Autoencoders

Variational Autoencoders red red red red red red red red red red red red red red red red red red red red Tutorial 03/10/2016 Generative modelling Assume that the original dataset is drawn from a distribution P(X ). Attempt to

More information