Graphite: Iterative Generative Modeling of Graphs

Size: px
Start display at page:

Download "Graphite: Iterative Generative Modeling of Graphs"

Transcription

1 Graphite: Iterative Generative Modeling of Graphs Aditya Grover* Aaron Zweig* Stefano Ermon Abstract Graphs are a fundamental abstraction for modeling relational data. However, graphs are discrete and combinatiorial in nature, and learning representations suitable for machine learning tasks poses statistical and computational challenges. In this work, we propose Graphite an algorithmic framework for unsupervised learning of representations over nodes in a graph using deep latent variable generative models. Our model is based on variational autoencoders (VAE), and differs from existing VAE frameworks for data modalities such as images, speech, and text in the use of spectral graph convolutions for parameterizing both the generative model (a.k.a. decoder) and inference model (a.k.a. encoder). The use of graph convolutions directly incorporates inductive biases specific to graphs, such as permutation invariance to node orderings and locality preference for clustering node representations, in the generative model. We demonstrate empirically that Graphite outperforms state-of-the-art approaches for representation learning over graphs for the task of link prediction on benchmark datasets. 1 Introduction Latent variable generative modeling is an effective approach for unsupervised representation learning of high-dimensional data [11]. In recent years, representations learned by latent variable models parameterized by deep neural networks have shown impressive performance on many tasks such as semi-supervised learning and structured prediction [6, 16]. However, these successes have been restricted to specific data modalities such as images and speech. In particular, current deep generative models cannot be directly applied to graph-structured data. Graphs arise in a wide variety of domains in physical sciences, information sciences, and social sciences, and consequently algorithms for unsupervised representation learning of such data have several downstream applications, including node classification, link prediction, and community detection [3, 10, 4]. In recent works, the notion of spectral graph convolutions has been the guiding principle for designing neural network architectures that can directly operate on graphs [1, 2]. A spectral graph convolution is defined as the multiplication of a signal (i.e., feature matrix associated with the nodes) with a parameterized filter, in the Fourier space of a graph. Graph convolution networks (GCN), in particular, efficiently compute local first-order approximations to spectral graph convolutions, and have been successfully applied across several graph mining tasks such as semi-supervised learning and relational learning [9]. Such tasks involve encoding an input graph to a final output representation (such as the labels associated with the nodes) with potentially more intermediate layers. The inverse problem of learning to decode a hidden representation into a graph, as in the case of a latent variable generative model, is to the best of our knowledge largely an open question that we address in this work. We propose Graphite, a framework for iteratively learning latent variable generative models of graphs. Specifically, we learn a directed model expressing a joint distribution P θ (A, Z) over the adjacency matrix of a graph A {0, 1} n n and a latent feature matrix Z R n k such that every row corresponds to a feature vector for a node in the graph. The multi-layer iterative decoding *Equal Contribution. Second workshop on Bayesian Deep Learning (NIPS 2017), Long Beach, CA, USA.

2 process for specifying the conditional distribution P θ (A Z) first constructs an intermediate graph using an inner-product operation followed by a sequence of graph convolutional layers on this intermediate graph. We learn the model parameters θ by maximizing a variational lower bound to the log-likelihood assigned by the model to the observed graph. Our empirical evaluations show that Graphite outperforms competing algorithms for the task of link prediction on benchmark datasets. 2 Learning framework We first define some notation and give a brief background on spectral graph convolutions before presenting Graphite. Consider an undirected graph G = (V, E) where V and E denote the set of nodes and edges respectively. We represent the graph using an adjacency matrix A {0, 1} n n where n = V and diagonal entries A ii = 1. Additionally, we denote the feature matrix associated with the graph as X R n m for an m-dimensional signal associated with each node in the graph. If there are no explicit node features, we set X = I n (identity). Let D be the diagonal degree matrix such that D ii = (i,j) E A ij and hence the graph Laplacian L = D A. We define à to be the symmetric normalization of A given by à = D 1/2 AD 1/2. The Fourier basis of a graph is given by the eigenvector matrix of the graph Laplacian. A spectral graph convolution convolves the feature matrix X with a parameterized filter F θ = diag(θ) (where θ R n are the learned parameters), defined as a matrix multiplication in the Fourier basis of the graph. Formally, we define the graph convolution operation between F θ and X for the graph A as: F θ X = UF θ U T X where U is the eigenvector matrix for the graph Laplacian, i.e., L = UΛU T where Λ is the diagonal matrix of eigenvalues. Since eigendecomposition of the graph Laplacian is computationally prohibitive, [9] propose a local first-order approximation: F θ X D 1/2 AD 1/2 X. Extending the above approximation to a neural network design with several filters applied in each layer, we get the layerwise propagation rule for l 1 a graph convolutional network (GCN): H (l) = η(d 1/2 AD 1/2 H (l 1) Θ (l) ) where H (l) and H (l 1) are the l-th and l 1-th layers of the GCN with the base case H (0) = X, Θ (l) is a matrix of learnable parameters for the current layer, and η is a suitable activation function. For brevity throughout the paper, we will denote the propagation rule as: H (l) = GCN η,l (A, H (l 1) ). An alternate interpretation of a GCN layer is performing message passing over a graph using the Weisfeiler-Lehman (WL) algorithm [18]. Intuitively, every node in the graph passes messages or information about its local structure to its neighbors. For a single GCN layer, the messages are passed to neighbors at a 1-hop distance from the node. In a multi-layer GCN with L-layers, information is propagated to neighbors up till an L-hop distance from the source. Refer to [9] for more details. 2.1 Graphite Variational Autoencoder We use upper-case symbols to denote probability distributions and assume they all admit absolutely continuous densities (denoted by the corresponding lower-case notation) on a suitable reference measure. We are interested in learning a latent variable model that specifies a joint distribution P θ (A, Z X) conditioned on the feature matrix X of a graph A and latent variable matrix Z R n k where every row corresponds to a latent vector for a node in the graph. Similar to a variational autoencoder [7], our learning objective maximizes an evidence lower bound (ELBO) to the loglikelihood of the observed graph A: [ log p θ (A X) E qφ (Z A,X) log p ] θ(a, Z X) q φ (Z A, X) where q φ (Z A, X) is a variational approximation to the true posterior p θ (Z A, X), parameterized by φ. We specify an isotropic standard Gaussian prior over Z such that the joint distribution factorizes as P θ (A, Z X) = P θ (A Z, X)P (Z). The observation model P θ (A Z, X) and the variational posterior Q φ (Z A, X), also referred to as the encoders and decoders respectively, are specified using deep neural networks. In particular, we use graph convolutional networks as described below. 2

3 Table 1: Area Under the ROC Curve (AUC) scores for link prediction (* denotes dataset with features) PPI Cora Citeseer Pubmed Cora* Citeseer* Pubmed* SC 84.2 ± ± ± ± DW 68.2 ± ± ± ± GAE 88.8 ± ± ± ± ± ± ± 0.04 VGAE 89.5 ± ± ± ± ± ± ± 0.13 Graphite-AE 91.1 ± ± ± ± ± ± ± 0.03 Graphite-VAE 91.2 ± ± ± ± ± ± ± 0.04 Table 2: Average Precision (AP) scores for link prediction (* denotes dataset with features) PPI Cora Citeseer Pubmed Cora* Citeseer* Pubmed* SC 88.9 ± ± ± ± DW 69.0 ± ± ± ± GAE 89.4 ± ± ± ± ± ± ± 0.04 VGAE 89.6 ± ± ± ± ± ± ± 0.12 Graphite-AE 92.1 ± ± ± ± ± ± ± 0.03 Graphite-VAE 92.2 ± ± ± ± ± ± ± 0.04 Encoder. The encoder Q φ (Z A, X) is a factorized multivariate Gaussian distribution specified using a multi-layer GCN, referred as EGCN (Encoder-GCN). In our experiments, we use two layers for the encoder and a multivariate Gaussian posterior with diagonal covariance. Hence, the forward pass is given by: H (1) e = EGCN η1,1(a, X) µ q, log σ q = EGCN η2,2(a, H (1) e ) where η 1 is any activation function and η 2 is set to identity (no activation) to search the full range of possible parameters for the Gaussian posterior. Decoder. The decoder P θ (A Z, X) is a factorized Bernoulli distribution also specified using a multi-layer GCN, referred as a DGCN (Decoder-GCN). However, we do not have access to an explicit graph to perform graph convolutions unlike an EGCN that could directly access A. We resolve this shortcoming by generating intermediate graphs using a similarity matrix that is a function of the latent feature matrix Z Q φ (Z A, X). The forward pass for the two-layer decoder is specified as: H (1) d  = σ(zz T ) = DGCN η1,1(â, [Z X]) Z = DGCN η2,2(â, H(1) d ) where η 1 is any suitable activation function and η 2 is the identity (such that Z and Z can take the same range of values) and the vertical bars denote concatenation of the original feature matrix X with Z. For the final step of the decoding, we obtain the desired conditional distribution p θ (A Z, X) with an inner product over a feature matrix specified as the convex combination of Z and Z. Z = λz + (1 λ)z p θ (A Z, X) = Π n i=1π n j=1p θ (A ij Z, X) where p θ (A ij Z, X) = σ(z iz j). Here, Z i denotes the i-th row of Z and λ [0, 1] is a tunable hyperparameter. 2.2 Graphite Autoencoder Similar to the differences between a standard autoencoder (AE) and a variational autoencoder, the EGCN here directly represents Z (instead of a variational posterior). Additionally, the learning objective minimizes the cross-entropy between the reconstructed graph probabilities and true graphs. 3

4 (a) Graphite-AE (b) Graphite-VAE Figure 1: t-sne embeddings of the latent feature vectors for the Cora dataset. Colors denote labels. 3 Experimental Evaluation We consider the task of link prediction [11]. Even though Graphite learns a distribution over graphs, it can be used for predictive tasks within a single graph. In the case of link prediction, we learn a model for a random, connected training subgraph of the true graph. For validation and testing, we add positive and negative (false) edges to the original graph and evaluate the performance of the model based on the reconstruction probabilities assigned to the validation and test edges (similar to denoising of the input graph). In our experiments, we held out a set of 5% edges for validation, 10% edges for testing, and train all models on the remaining subgraph. We evaluate performance based on the Area Under the ROC Curve (AUC) and Average Precision (AP) metrics. Additionally, the validation and testing sets also each contain an equal number of non-edges. We compared across four standard benchmark datasets: a Protein-Protein Interaction (PPI) network for Homo Sapiens [17, 5] with proteins as nodes and interactions as edges, and three citation networks Cora, Citeseer, and Pubmed with papers as nodes and citations as edges [15]. For the citation networks, the text in the papers can be synthesized into optional node features. For Graphite-AE and Graphite-VAE, we used an architecture of units for the encoder and units for the decoder trained using the Adam optimizer [7] with a learning rate of The dropout rate (for edges) and λ were tuned as hyperparameters on the validation set to optimize the AUC. Additionally, we trained every model for 500 iterations and used the model checkpoint with the best validation loss for testing. We evaluated Graphite-VAE and Graphite-AE against competing methods for representation learning on graphs. The baselines we consider are Spectral Clustering (SC) [1], DeepWalk (DW) [14], Graph Autoencoder [8], and Variational Graph Autoencoder [8]. We used the SC implementation from [13] and public implementations for others made available by the authors. For SC, we used a dimension size of 128. For DW which uses a skipgram like objective on random walks from the graph, we used the same dimension size and default settings used in the paper of 10 random walks of length 80 per node and a context size of 10. For GAE and VGAE, we used the same architecture as VGAE and Adam optimizer with learning rate of Note that SC and DW do not provide the ability to incorporate node features while learning embeddings, and hence we evaluate them only on the featureless datasets. GAE and VGAE are a special case of Graphite, using only a single inner-product decoder (i.e., λ = 1). The AUC and AP results (along with standard errors) are shown in Table 1 and Table 2 respectively averaged over 50 random train/validation/test splits. On both metrics, Graphite-VAE gives the best performance overall. Graphite-AE also gives good results, generally outperforming its closest competitor GAE. We further visualize the embeddings generated by a 2D t-sne projection [12] of the latent feature vectors (given as rows for Z with λ = 0.5) on the Cora dataset in Figure 1. Even without any access to label information for the nodes during training, the Graphite models are able to cluster the nodes (papers) as per their labels (paper categories). 4

5 References [1] J. Bruna, W. Zaremba, A. Szlam, and Y. LeCun. Spectral networks and locally connected networks on graphs. In ICLR, [2] M. Defferrard, X. Bresson, and P. Vandergheynst. Convolutional neural networks on graphs with fast localized spectral filtering. In NIPS, [3] D. Easley and J. Kleinberg. Networks, crowds, and markets: Reasoning about a highly connected world. Cambridge University Press, [4] S. Fortunato. Community detection in graphs. Physics reports, 486(3):75 174, [5] A. Grover and J. Leskovec. node2vec: Scalable feature learning for networks. In KDD, [6] D. P. Kingma, S. Mohamed, D. J. Rezende, and M. Welling. Semi-supervised learning with deep generative models. In NIPS, [7] D. P. Kingma and M. Welling. Auto-encoding variational bayes. In ICLR, [8] T. N. Kipf and M. Welling. Variational graph auto-encoders. arxiv preprint arxiv: , [9] T. N. Kipf and M. Welling. Semi-supervised classification with graph convolutional networks. In ICLR, [10] D. Liben-Nowell and J. Kleinberg. The link-prediction problem for social networks. journal of the Association for Information Science and Technology, 58(7): , [11] J. C. Loehlin. Latent variable models: An introduction to factor, path, and structural analysis. Lawrence Erlbaum Associates Publishers, [12] L. v. d. Maaten and G. Hinton. Visualizing data using t-sne. Journal of Machine Learning Research, 9(Nov): , [13] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, et al. Scikit-learn: Machine learning in python. Journal of Machine Learning Research, 12(Oct): , [14] B. Perozzi, R. Al-Rfou, and S. Skiena. Deepwalk: Online learning of social representations. In KDD, [15] P. Sen, G. M. Namata, M. Bilgic, L. Getoor, B. Gallagher, and T. Eliassi-Rad. Collective classification in network data. AI Magazine, 29(3):93 106, [16] K. Sohn, H. Lee, and X. Yan. Learning structured output representation using deep conditional generative models. In NIPS, [17] C. Stark, B. J. Breitkreutz, T. Reguly, L. Boucher, A. Breitkreutz, and M. Tyers. Biogrid: A general repository for interaction datasets. Nucleic Acids Research, 34: , [18] B. Weisfeiler and A. Lehman. A reduction of a graph to a canonical form and an algebra arising during this reduction. Nauchno-Technicheskaya Informatsia, 2(9):12 16,

Deep Learning on Graphs

Deep Learning on Graphs Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 6 April 2017 joint work with Max Welling (University of Amsterdam) The success story of deep

More information

Deep Learning on Graphs

Deep Learning on Graphs Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016

More information

Deep Generative Models Variational Autoencoders

Deep Generative Models Variational Autoencoders Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative

More information

Network embedding. Cheng Zheng

Network embedding. Cheng Zheng Network embedding Cheng Zheng Outline Problem definition Factorization based algorithms --- Laplacian Eigenmaps(NIPS, 2001) Random walk based algorithms ---DeepWalk(KDD, 2014), node2vec(kdd, 2016) Deep

More information

A Higher-Order Graph Convolutional Layer

A Higher-Order Graph Convolutional Layer A Higher-Order Graph Convolutional Layer Sami Abu-El-Haija 1, Nazanin Alipourfard 1, Hrayr Harutyunyan 1, Amol Kapoor 2, Bryan Perozzi 2 1 Information Sciences Institute University of Southern California

More information

Deep Attributed Network Embedding

Deep Attributed Network Embedding Deep Attributed Network Embedding Hongchang Gao, Heng Huang Department of Electrical and Computer Engineering University of Pittsburgh, USA hongchanggao@gmail.com, heng.huang@pitt.edu Abstract Network

More information

node2vec: Scalable Feature Learning for Networks

node2vec: Scalable Feature Learning for Networks node2vec: Scalable Feature Learning for Networks A paper by Aditya Grover and Jure Leskovec, presented at Knowledge Discovery and Data Mining 16. 11/27/2018 Presented by: Dharvi Verma CS 848: Graph Database

More information

arxiv: v4 [cs.lg] 22 Feb 2017

arxiv: v4 [cs.lg] 22 Feb 2017 SEMI-SUPERVISED CLASSIFICATION WITH GRAPH CONVOLUTIONAL NETWORKS Thomas N. Kipf University of Amsterdam T.N.Kipf@uva.nl Max Welling University of Amsterdam Canadian Institute for Advanced Research (CIFAR)

More information

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

Lecture 21 : A Hybrid: Deep Learning and Graphical Models 10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation

More information

Adversarial Network Embedding

Adversarial Network Embedding Adversarial Network Embedding Quanyu Dai 1, Qiang Li 1,2, Jian Tang 3,4, Dan Wang 1 1 Department of Computing, The Hong Kong Polytechnic University, Hong Kong 2 School of Software, FEIT, The University

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Unsupervised learning Daniel Hennes 29.01.2018 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Supervised learning Regression (linear

More information

arxiv: v1 [cs.si] 20 Nov 2018

arxiv: v1 [cs.si] 20 Nov 2018 Temporal Graph Offset Reconstruction: Towards Temporally Robust Graph Representation Learning arxiv:1811.08366v1 [cs.si] 20 Nov 2018 Stephen Bonner, John Brennan, Ibad Kureshi, Georgios Theodoropoulos,

More information

GraphNet: Recommendation system based on language and network structure

GraphNet: Recommendation system based on language and network structure GraphNet: Recommendation system based on language and network structure Rex Ying Stanford University rexying@stanford.edu Yuanfang Li Stanford University yli03@stanford.edu Xin Li Stanford University xinli16@stanford.edu

More information

GeniePath: Graph Neural Networks with Adaptive Receptive Paths

GeniePath: Graph Neural Networks with Adaptive Receptive Paths GeniePath: Graph Neural Networks with Adaptive Receptive Paths Ziqi Liu Hangzhou, China ziqilau@gmail.com Jun Zhou Beijing, China ABSTRACT We present, GeniePath, a scalable approach for learning adaptive

More information

Supervised Hyperbolic Representation Learning for Real-World Networks

Supervised Hyperbolic Representation Learning for Real-World Networks Supervised Hyperbolic Representation Learning for Real-World Networks Manan Shah manans@stanford.edu Sagar Maheshwari msagar@stanford.edu Abstract Graphs are inherently complex structures, and learning

More information

Variational Autoencoders. Sargur N. Srihari

Variational Autoencoders. Sargur N. Srihari Variational Autoencoders Sargur N. srihari@cedar.buffalo.edu Topics 1. Generative Model 2. Standard Autoencoder 3. Variational autoencoders (VAE) 2 Generative Model A variational autoencoder (VAE) is a

More information

FeaStNet: Feature-Steered Graph Convolutions for 3D Shape Analysis Supplementary Material

FeaStNet: Feature-Steered Graph Convolutions for 3D Shape Analysis Supplementary Material FeaStNet: Feature-Steered Graph Convolutions for 3D Shape Analysis Supplementary Material Nitika Verma Edmond Boyer Jakob Verbeek Univ. Grenoble Alpes, Inria, CNRS, Grenoble INP, LJK, 38000 Grenoble, France

More information

Adversarial Network Embedding

Adversarial Network Embedding Adversarial Network Embedding Quanyu Dai 1, Qiang Li 1,2, Jian Tang 3,4, Dan Wang 1 1 Department of Computing, The Hong Kong Polytechnic University, Hong Kong 2 School of Software, FEIT, The University

More information

Inductive Representation Learning on Large Graphs

Inductive Representation Learning on Large Graphs Inductive Representation Learning on Large Graphs William L. Hamilton wleif@stanford.edu Rex Ying rexying@stanford.edu Jure Leskovec jure@cs.stanford.edu Department of Computer Science Stanford University

More information

Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning

Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning Qimai Li, Zhichao Han 2, Xiao-Ming Wu The Hong Kong Polytechnic University 2 ETH Zurich csqmli@comp.polyu.edu.hk, zhhan@student.ethz.ch,

More information

Alternatives to Direct Supervision

Alternatives to Direct Supervision CreativeAI: Deep Learning for Graphics Alternatives to Direct Supervision Niloy Mitra Iasonas Kokkinos Paul Guerrero Nils Thuerey Tobias Ritschel UCL UCL UCL TUM UCL Timetable Theory and Basics State of

More information

Autoencoders, denoising autoencoders, and learning deep networks

Autoencoders, denoising autoencoders, and learning deep networks 4 th CiFAR Summer School on Learning and Vision in Biology and Engineering Toronto, August 5-9 2008 Autoencoders, denoising autoencoders, and learning deep networks Part II joint work with Hugo Larochelle,

More information

Semi-supervised Methods for Graph Representation

Semi-supervised Methods for Graph Representation Semi-supervised Methods for Graph Representation Modeling Data With Networks + Network Embedding: Problems, Methodologies and Frontiers Ivan Brugere (University of Illinois at Chicago) Peng Cui (Tsinghua

More information

Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning

Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18) Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning Qimai Li, 1 Zhichao Han, 1,2 Xiao-Ming Wu 1 1 The Hong

More information

Modeling and Optimization of Thin-Film Optical Devices using a Variational Autoencoder

Modeling and Optimization of Thin-Film Optical Devices using a Variational Autoencoder Modeling and Optimization of Thin-Film Optical Devices using a Variational Autoencoder Introduction John Roberts and Evan Wang, {johnr3, wangevan}@stanford.edu Optical thin film systems are structures

More information

Deep Learning for Computer Vision

Deep Learning for Computer Vision Deep Learning for Computer Vision Spring 2018 http://vllab.ee.ntu.edu.tw/dlcv.html (primary) https://ceiba.ntu.edu.tw/1062dlcv (grade, etc.) FB: DLCV Spring 2018 Yu Chiang Frank Wang 王鈺強, Associate Professor

More information

Supplementary Material: The Emergence of. Organizing Structure in Conceptual Representation

Supplementary Material: The Emergence of. Organizing Structure in Conceptual Representation Supplementary Material: The Emergence of Organizing Structure in Conceptual Representation Brenden M. Lake, 1,2 Neil D. Lawrence, 3 Joshua B. Tenenbaum, 4,5 1 Center for Data Science, New York University

More information

LINK WEIGHT PREDICTION WITH NODE EMBEDDINGS

LINK WEIGHT PREDICTION WITH NODE EMBEDDINGS LINK WEIGHT PREDICTION WITH NODE EMBEDDINGS Anonymous authors Paper under double-blind review ABSTRACT Application of deep learning has been successful in various domains such as image recognition, speech

More information

Auto-Encoding Variational Bayes

Auto-Encoding Variational Bayes Auto-Encoding Variational Bayes Diederik P (Durk) Kingma, Max Welling University of Amsterdam Ph.D. Candidate, advised by Max Durk Kingma D.P. Kingma Max Welling Problem class Directed graphical model:

More information

Adversarially Regularized Graph Autoencoder for Graph Embedding

Adversarially Regularized Graph Autoencoder for Graph Embedding Adversarially Regularized Graph Autoencoder for Graph Embedding Shirui Pan 1, Ruiqi Hu 1, Guodong Long 1, Jing Jiang 1, Lina Yao 2, Chengqi Zhang 1 1 Centre for Artificial Intelligence, FEIT, University

More information

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all

More information

Unsupervised Learning

Unsupervised Learning Deep Learning for Graphics Unsupervised Learning Niloy Mitra Iasonas Kokkinos Paul Guerrero Vladimir Kim Kostas Rematas Tobias Ritschel UCL UCL/Facebook UCL Adobe Research U Washington UCL Timetable Niloy

More information

Probabilistic Programming with Pyro

Probabilistic Programming with Pyro Probabilistic Programming with Pyro the pyro team Eli Bingham Theo Karaletsos Rohit Singh JP Chen Martin Jankowiak Fritz Obermeyer Neeraj Pradhan Paul Szerlip Noah Goodman Why Pyro? Why probabilistic modeling?

More information

Representation Learning on Graphs: Methods and Applications

Representation Learning on Graphs: Methods and Applications Representation Learning on Graphs: Methods and Applications William L. Hamilton wleif@stanford.edu Rex Ying rexying@stanford.edu Department of Computer Science Stanford University Stanford, CA, 94305 Jure

More information

arxiv: v1 [cs.si] 14 Mar 2017

arxiv: v1 [cs.si] 14 Mar 2017 SNE: Signed Network Embedding Shuhan Yuan 1, Xintao Wu 2, and Yang Xiang 1 1 Tongji University, Shanghai, China, Email:{4e66,shxiangyang}@tongji.edu.cn 2 University of Arkansas, Fayetteville, AR, USA,

More information

arxiv: v1 [cs.lg] 20 Apr 2017

arxiv: v1 [cs.lg] 20 Apr 2017 Dynamic Graph Convolutional Networks Franco Manessi 1, Alessandro Rozza 1, and Mario Manzo 2 arxiv:1704.06199v1 [cs.lg] 20 Apr 2017 1 Research Team - Waynaut {name.surname}@waynaut.com 2 Servizi IT - Università

More information

arxiv: v4 [cs.lg] 13 Sep 2017

arxiv: v4 [cs.lg] 13 Sep 2017 Learning Edge Representations via Low-Rank Asymmetric Projections Sami Abu-El-Haija Google Research Mountain View, California haija@google.com Bryan Perozzi Google Research New York City, New York bperozzi@acm.org

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17

(University Improving of Montreal) Generative Adversarial Networks with Denoising Feature Matching / 17 Improving Generative Adversarial Networks with Denoising Feature Matching David Warde-Farley 1 Yoshua Bengio 1 1 University of Montreal, ICLR,2017 Presenter: Bargav Jayaraman Outline 1 Introduction 2 Background

More information

arxiv: v1 [cs.lg] 24 Jan 2019

arxiv: v1 [cs.lg] 24 Jan 2019 Jaehoon Cha Kyeong Soo Kim Sanghuyk Lee arxiv:9.879v [cs.lg] Jan 9 Abstract Noting the importance of the latent variables in inference and learning, we propose a novel framework for autoencoders based

More information

Day 3 Lecture 1. Unsupervised Learning

Day 3 Lecture 1. Unsupervised Learning Day 3 Lecture 1 Unsupervised Learning Semi-supervised and transfer learning Myth: you can t do deep learning unless you have a million labelled examples for your problem. Reality You can learn useful representations

More information

arxiv: v1 [cs.lg] 10 Dec 2017

arxiv: v1 [cs.lg] 10 Dec 2017 DGCNN: Disordered Graph Convolutional Neural Network Based on the Gaussian Mixture Model arxiv:1712.03563v1 [cs.lg] 10 Dec 2017 Bo Wu, Yang Liu, Bo Lang, Lei Huang State Key Laboratory of Software Development

More information

arxiv: v1 [cs.si] 7 Sep 2018

arxiv: v1 [cs.si] 7 Sep 2018 dyngraph2vec: Capturing Network Dynamics using Dynamic Graph Representation Learning Palash Goyal USC Information Sciences Institute palashgo@usc.edu Sujit Rokka Chhetri * University of California-Irvine

More information

arxiv: v1 [stat.ml] 10 Dec 2018

arxiv: v1 [stat.ml] 10 Dec 2018 1st Symposium on Advances in Approximate Bayesian Inference, 2018 1 7 Disentangled Dynamic Representations from Unordered Data arxiv:1812.03962v1 [stat.ml] 10 Dec 2018 Leonhard Helminger Abdelaziz Djelouah

More information

Deep Learning Applications

Deep Learning Applications October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning

More information

arxiv: v1 [cs.si] 12 Jan 2019

arxiv: v1 [cs.si] 12 Jan 2019 Predicting Diffusion Reach Probabilities via Representation Learning on Social Networks Furkan Gursoy furkan.gursoy@boun.edu.tr Ahmet Onur Durahim onur.durahim@boun.edu.tr arxiv:1901.03829v1 [cs.si] 12

More information

arxiv: v4 [stat.ml] 27 Feb 2018

arxiv: v4 [stat.ml] 27 Feb 2018 DEEP GAUSSIAN EMBEDDING OF GRAPHS: UNSUPERVISED INDUCTIVE LEARNING VIA RANKING Aleksandar Bojchevski, Stephan Günnemann Technical University of Munich, Germany {a.bojchevski,guennemann}@in.tum.de arxiv:1707.03815v4

More information

Feature Hashing for Network Representation Learning

Feature Hashing for Network Representation Learning Feature Hashing for Network Representation Learning Qixiang Wang 1, Shanfeng Wang 2, Maoguo Gong 1, Yue Wu 3 1 Key Laboratory of Intelligent Perception and Image Understanding, Xidian University, Xi an

More information

Unsupervised Learning

Unsupervised Learning Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised

More information

Clustering and Unsupervised Anomaly Detection with l 2 Normalized Deep Auto-Encoder Representations

Clustering and Unsupervised Anomaly Detection with l 2 Normalized Deep Auto-Encoder Representations Clustering and Unsupervised Anomaly Detection with l 2 Normalized Deep Auto-Encoder Representations Caglar Aytekin, Xingyang Ni, Francesco Cricri and Emre Aksu Nokia Technologies, Tampere, Finland Corresponding

More information

Online Social Networks and Media

Online Social Networks and Media Online Social Networks and Media Absorbing Random Walks Link Prediction Why does the Power Method work? If a matrix R is real and symmetric, it has real eigenvalues and eigenvectors: λ, w, λ 2, w 2,, (λ

More information

Fast Network Embedding Enhancement via High Order Proximity Approximation

Fast Network Embedding Enhancement via High Order Proximity Approximation Fast Network Embedding Enhancement via High Order Proximity Approximation Cheng Yang 1, Maosong Sun 1,2, Zhiyuan Liu 1,2, Cunchao Tu 1,2 1 Department of Computer Science and Technology, Tsinghua University,

More information

HARP: Hierarchical Representation Learning for Networks

HARP: Hierarchical Representation Learning for Networks HARP: Hierarchical Representation Learning for Networks Haochen Chen Stony Brook University haocchen@cs.stonybrook.edu Yifan Hu Yahoo! Research yifanhu@oath.com Bryan Perozzi Google Research bperozzi@acm.org

More information

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin Clustering K-means Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, 2014 Carlos Guestrin 2005-2014 1 Clustering images Set of Images [Goldberger et al.] Carlos Guestrin 2005-2014

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Scalable Clustering of Signed Networks Using Balance Normalized Cut

Scalable Clustering of Signed Networks Using Balance Normalized Cut Scalable Clustering of Signed Networks Using Balance Normalized Cut Kai-Yang Chiang,, Inderjit S. Dhillon The 21st ACM International Conference on Information and Knowledge Management (CIKM 2012) Oct.

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover

More information

Bayesian model ensembling using meta-trained recurrent neural networks

Bayesian model ensembling using meta-trained recurrent neural networks Bayesian model ensembling using meta-trained recurrent neural networks Luca Ambrogioni l.ambrogioni@donders.ru.nl Umut Güçlü u.guclu@donders.ru.nl Yağmur Güçlütürk y.gucluturk@donders.ru.nl Julia Berezutskaya

More information

Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting

Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting Yaguang Li Joint work with Rose Yu, Cyrus Shahabi, Yan Liu Page 1 Introduction Traffic congesting is wasteful of time,

More information

Unsupervised Multi-view Nonlinear Graph Embedding

Unsupervised Multi-view Nonlinear Graph Embedding Unsupervised Multi-view Nonlinear Graph Embedding Jiaming Huang,2, Zhao Li, Vincent W. Zheng 3, Wen Wen 2, Yifan Yang, Yuanmi Chen Alibaba Group, Hangzhou, China 2 School of Computers, Guangdong University

More information

Deep Generative Models and a Probabilistic Programming Library

Deep Generative Models and a Probabilistic Programming Library Deep Generative Models and a Probabilistic Programming Library Discriminative (Deep) Learning Learn a (differentiable) function mapping from input to output x f(x; θ) y Gradient back-propagation Generative

More information

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Tianyu Wang Australia National University, Colledge of Engineering and Computer Science u@anu.edu.au Abstract. Some tasks,

More information

A Taxonomy of Semi-Supervised Learning Algorithms

A Taxonomy of Semi-Supervised Learning Algorithms A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph

More information

Effectiveness of Sparse Features: An Application of Sparse PCA

Effectiveness of Sparse Features: An Application of Sparse PCA 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Virtual Adversarial Ladder Networks for Semi-Supervised Learning

Virtual Adversarial Ladder Networks for Semi-Supervised Learning Virtual Adversarial Ladder Networks for Semi-Supervised Learning Saki Shinoda 1, Daniel E. Worrall 2 & Gabriel J. Brostow 2 Computer Science Department University College London United Kingdom 1 saki.shinoda.16@ucl.ac.uk

More information

Visual Representations for Machine Learning

Visual Representations for Machine Learning Visual Representations for Machine Learning Spectral Clustering and Channel Representations Lecture 1 Spectral Clustering: introduction and confusion Michael Felsberg Klas Nordberg The Spectral Clustering

More information

arxiv: v2 [cs.si] 4 Jul 2018

arxiv: v2 [cs.si] 4 Jul 2018 FSCNMF: Fusing Structure and Content via Non-negative Matrix Factorization for Embedding Information Networks arxiv:1804.05313v2 [cs.si] 4 Jul 2018 ABSTRACT Sambaran Bandyopadhyay IBM Research sambband@in.ibm.com

More information

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x

Autoencoder. Representation learning (related to dictionary learning) Both the input and the output are x Deep Learning 4 Autoencoder, Attention (spatial transformer), Multi-modal learning, Neural Turing Machine, Memory Networks, Generative Adversarial Net Jian Li IIIS, Tsinghua Autoencoder Autoencoder Unsupervised

More information

GAN Frontiers/Related Methods

GAN Frontiers/Related Methods GAN Frontiers/Related Methods Improving GAN Training Improved Techniques for Training GANs (Salimans, et. al 2016) CSC 2541 (07/10/2016) Robin Swanson (robin@cs.toronto.edu) Training GANs is Difficult

More information

Deep Predictive Coverage Collection

Deep Predictive Coverage Collection Deep Predictive Coverage Collection Rajarshi Roy Chinmay Duvedi Saad Godil Mark Williams rajarshir@nvidia.com cduvedi@nvidia.com sgodil@nvidia.com mwilliams@nvidia.com NVIDIA Corporation 2701 San Tomas

More information

COMS 4771 Clustering. Nakul Verma

COMS 4771 Clustering. Nakul Verma COMS 4771 Clustering Nakul Verma Supervised Learning Data: Supervised learning Assumption: there is a (relatively simple) function such that for most i Learning task: given n examples from the data, find

More information

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation Learning 4 Supervised Learning 4 Unsupervised Learning 4

More information

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem

The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Int. J. Advance Soft Compu. Appl, Vol. 9, No. 1, March 2017 ISSN 2074-8523 The Un-normalized Graph p-laplacian based Semi-supervised Learning Method and Speech Recognition Problem Loc Tran 1 and Linh Tran

More information

Deep Graph Infomax. Petar Velic kovic 1,2, William Fedus2,3,5, William L. Hamilton2,4, Pietro Lio 1, Yoshua Bengio2,3 and R Devon Hjelm6,2,3

Deep Graph Infomax. Petar Velic kovic 1,2, William Fedus2,3,5, William L. Hamilton2,4, Pietro Lio 1, Yoshua Bengio2,3 and R Devon Hjelm6,2,3 Deep Graph Infomax Petar Velic kovic 1,2, William Fedus2,3,5, William L. Hamilton2,4, Pietro Lio 1, Yoshua Bengio2,3 and R Devon Hjelm6,2,3 1 University 4 McGill of Cambridge 2 Mila University 5 Google

More information

PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D.

PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D. PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D. Rhodes 5/10/17 What is Machine Learning? Machine learning

More information

CS224W: Analysis of Networks Jure Leskovec, Stanford University

CS224W: Analysis of Networks Jure Leskovec, Stanford University CS224W: Analysis of Networks Jure Leskovec, Stanford University http://cs224w.stanford.edu Jure Leskovec, Stanford CS224W: Analysis of Networks, http://cs224w.stanford.edu 2????? Machine Learning Node

More information

Semi-supervised protein classification using cluster kernels

Semi-supervised protein classification using cluster kernels Semi-supervised protein classification using cluster kernels Jason Weston Max Planck Institute for Biological Cybernetics, 72076 Tübingen, Germany weston@tuebingen.mpg.de Dengyong Zhou, Andre Elisseeff

More information

Semi-Amortized Variational Autoencoders

Semi-Amortized Variational Autoencoders Semi-Amortized Variational Autoencoders Yoon Kim Sam Wiseman Andrew Miller David Sontag Alexander Rush Code: https://github.com/harvardnlp/sa-vae Background: Variational Autoencoders (VAE) (Kingma et al.

More information

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points] CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.

More information

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos Machine Learning for Computer Vision 1 22 October, 2012 MVA ENS Cachan Lecture 5: Introduction to generative models Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Center for Visual Computing Ecole Centrale Paris

More information

Graph Convolutional Networks: Algorithms, Applications and Open Challenges

Graph Convolutional Networks: Algorithms, Applications and Open Challenges Graph Convolutional Networks: Algorithms, Applications and Open Challenges Si Zhang, Hanghang Tong, Jiejun Xu, and Ross Maciejewski Arizona State University, {szhan172, hanghang.tong, rmacieje}@asu.edu;

More information

Deep Learning Approach to Link Weight Prediction

Deep Learning Approach to Link Weight Prediction Deep Learning Approach to Link Weight Prediction Yuchen Hou School of Electrical Engineering and Computer Science Washington State University, Pullman, WA 99164 yuchen.hou@wsu.edu Lawrence B. Holder School

More information

JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation

JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS Puyang Xu, Ruhi Sarikaya Microsoft Corporation ABSTRACT We describe a joint model for intent detection and slot filling based

More information

GRADIENT-BASED OPTIMIZATION OF NEURAL

GRADIENT-BASED OPTIMIZATION OF NEURAL Workshop track - ICLR 28 GRADIENT-BASED OPTIMIZATION OF NEURAL NETWORK ARCHITECTURE Will Grathwohl, Elliot Creager, Seyed Kamyar Seyed Ghasemipour, Richard Zemel Department of Computer Science University

More information

Neural Networks: promises of current research

Neural Networks: promises of current research April 2008 www.apstat.com Current research on deep architectures A few labs are currently researching deep neural network training: Geoffrey Hinton s lab at U.Toronto Yann LeCun s lab at NYU Our LISA lab

More information

Don t Walk, Skip! Online Learning of Multi-scale Network Embeddings

Don t Walk, Skip! Online Learning of Multi-scale Network Embeddings Don t Walk, Skip! Online Learning of Multi-scale Network Embeddings Bryan Perozzi, Vivek Kulkarni, Haochen Chen, and Steven Skiena Department of Computer Science, Stony Brook University {bperozzi, vvkulkarni,

More information

Paper2vec: Combining Graph and Text Information for Scientific Paper Representation

Paper2vec: Combining Graph and Text Information for Scientific Paper Representation Paper2vec: Combining Graph and Text Information for Scientific Paper Representation Soumyajit Ganguly and Vikram Pudi International Institute of Information Technology Hyderabad, India. soumyajit.ganguly@research.iiit.ac.in,

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

This Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured

This Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 1 This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures

More information

Transductive Phoneme Classification Using Local Scaling And Confidence

Transductive Phoneme Classification Using Local Scaling And Confidence 202 IEEE 27-th Convention of Electrical and Electronics Engineers in Israel Transductive Phoneme Classification Using Local Scaling And Confidence Matan Orbach Dept. of Electrical Engineering Technion

More information

27: Hybrid Graphical Models and Neural Networks

27: Hybrid Graphical Models and Neural Networks 10-708: Probabilistic Graphical Models 10-708 Spring 2016 27: Hybrid Graphical Models and Neural Networks Lecturer: Matt Gormley Scribes: Jakob Bauer Otilia Stretcu Rohan Varma 1 Motivation We first look

More information

Link Mining & Entity Resolution. Lise Getoor University of Maryland, College Park

Link Mining & Entity Resolution. Lise Getoor University of Maryland, College Park Link Mining & Entity Resolution Lise Getoor University of Maryland, College Park Learning in Structured Domains Traditional machine learning and data mining approaches assume: A random sample of homogeneous

More information

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes

More information

Learning Discrete Representations via Information Maximizing Self-Augmented Training

Learning Discrete Representations via Information Maximizing Self-Augmented Training A. Relation to Denoising and Contractive Auto-encoders Our method is related to denoising auto-encoders (Vincent et al., 2008). Auto-encoders maximize a lower bound of mutual information (Cover & Thomas,

More information

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic SEMANTIC COMPUTING Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 23 November 2018 Overview Unsupervised Machine Learning overview Association

More information

Toward a rigorous statistical framework for brain mapping

Toward a rigorous statistical framework for brain mapping Toward a rigorous statistical framework for brain mapping Bertrand Thirion, bertrand.thirion@inria.fr 1 Cognitive neuroscience How are cognitive activities affected or controlled by neural circuits in

More information

node2vec: Scalable Feature Learning for Networks

node2vec: Scalable Feature Learning for Networks node2vec: Scalable Feature Learning for Networks Aditya Grover Stanford University adityag@cs.stanford.edu Jure Leskovec Stanford University jure@cs.stanford.edu ABSTRACT Prediction tasks over nodes and

More information

A United Approach to Learning Sparse Attributed Network Embedding

A United Approach to Learning Sparse Attributed Network Embedding 2018 IEEE International Conference on Data Mining A United Approach to Learning Sparse Attributed Network Embedding Hao Wang 1, Enhong Chen 1,*,QiLiu 1, Tong Xu 1, Dongfang Du 2, Wen Su 2, Xiaopeng Zhang

More information

Machine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart

Machine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart Machine Learning The Breadth of ML Neural Networks & Deep Learning Marc Toussaint University of Stuttgart Duy Nguyen-Tuong Bosch Center for Artificial Intelligence Summer 2017 Neural Networks Consider

More information

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction Akarsh Pokkunuru EECS Department 03-16-2017 Contractive Auto-Encoders: Explicit Invariance During Feature Extraction 1 AGENDA Introduction to Auto-encoders Types of Auto-encoders Analysis of different

More information