This Talk. 1) Node embeddings. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured
|
|
- Chloe Simon
- 5 years ago
- Views:
Transcription
1 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures for graphstructured data 3) pplications
2 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Part 2: Graph Neural Networks
3 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Embedding Nodes Goal is to encode nodes so that similarity in the embedding space (e.g., dot product) approximates similarity in the original network.
4 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Embedding Nodes Goal: similarity(u, v) z > v z u Need to define!
5 Two Key omponents Encoder maps each node to a lowdimensional vector. enc(v) =z v node in the input graph Similarity function specifies how relationships in vector space map to relationships in the original network. Similarity of u and v in the original network d-dimensional embedding similarity(u, v) z > v z u dot product between node embeddings Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW
6 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW From Shallow to Deep So far we have focused on shallow encoders, i.e. embedding lookups: embedding matrix embedding vector for a specific node Z = Dimension/size of embeddings one column per node
7 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW From Shallow to Deep Limitations of shallow encoding: O( V ) parameters are needed: there no parameter sharing and every node has its own unique embedding vector. Inherently transductive : It is impossible to generate embeddings for nodes that were not seen during training. Do not incorporate node features: Many graphs have features that we can and should leverage.
8 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW From Shallow to Deep We will now discuss deeper methods based on graph neural networks. enc(v) = complex function that depends on graph structure. In general, all of these more complex encoders can be combined with the similarity functions from the previous section.
9 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section We will now discuss deeper methods based on graph neural networks. 1. The asics 2. Graph onvolutional Networks (GNs) 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings
10 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW The asics: Graph Neural Networks ased on material from: Hamilton et al Representation Learning on Graphs: Methods and pplications. IEEE Data Engineering ulletin on Graph Systems. Scarselli et al The Graph Neural Network Model. IEEE Transactions on Neural Networks.
11 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Setup ssume we have a graph G: V is the vertex set. is the adjacency matrix (assume binary). X R m V is a matrix of node features. ategorical attributes, text, image data E.g., profile information in a social network. Node degrees, clustering coefficients, etc. Indicator vectors (i.e., one-hot encoding of each node)
12 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Key idea: Generate node embeddings based on local neighborhoods. TRGET NODE D E F D E F INPUT GRPH
13 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Intuition: Nodes aggregate information from their neighbors using neural networks TRGET NODE D E F D E F INPUT GRPH
14 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Intuition: Network neighborhood defines a computation graph Every node defines a unique computation graph!
15 Neighborhood ggregation Nodes have embeddings at each layer. Model can be arbitrary depth. layer-0 embedding of node u is its input feature, i.e. x u. TRGET NODE D E INPUT GRPH F Layer-2 Layer-1 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW D Layer-0 x F E x x x x E x F x
16 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood onvolutions Neighborhood aggregation can be viewed as a center-surround filter. Mathematically related to spectral graph convolutions (see ronstein et al., 2017)
17 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Key distinctions are in how different approaches aggregate information across the layers. what s in the box!? TRGET NODE? D E INPUT GRPH F??? D?? E F
18 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation asic approach: verage neighbor information and apply a neural network. TRGET NODE 1) average messages from neighbors D E F D E F INPUT GRPH 2) apply neural network
19 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW The Math asic approach: verage neighbor messages and apply a neural network. h 0 v = x v 0 h k v k kth layer embedding non-linearity (e.g., of v ReLU or tanh) Initial layer 0 embeddings are equal to node features X u2n(v) h k 1 u average of neighbor s previous layer embeddings previous layer embedding of v 1 N(v) + kh k v 1, 8k >0
20 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model How do we train the model to generate high-quality embeddings? z Need to define a loss function on the embeddings, L(z u )!
21 Training the Model h 0 v = x v 0 h k v k X u2n(v) trainable matrices (i.e., what we learn) h k 1 u 1 N(v) + kh k v 1, 8k 2 {1,...,K} z v = h K v fter K-layers of neighborhood aggregation, we get output embeddings for each node. We can feed these embeddings into any loss function and run stochastic gradient descent to train the aggregation parameters. Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW
22 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model Train in an unsupervised manner using only the graph structure. Unsupervised loss function can be anything from the last section, e.g., based on Random walks (node2vec, DeepWalk) Graph factorization i.e., train the model so that similar nodes have similar embeddings.
23 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model lternative: Directly train the model for a supervised task (e.g., node classification): Human or bot? Human or bot? e.g., an online social network
24 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model lternative: Directly train the model for a supervised task (e.g., node classification): Human or bot? L = X v2v classification weights y v log( (z > v )) + (1 y v )log(1 (z > v )) output node embedding node class label
25 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Overview of Model Design 1) Define a neighborhood aggregation function. z 2) Define a loss function on the embeddings, L(z u )
26 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Overview of Model Design 3) Train on a set of nodes, i.e., a batch of compute graphs
27 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Overview of Model 4) Generate embeddings for nodes as needed Even for nodes we never trained on!!!!
28 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Inductive apability The same aggregation parameters are shared for all nodes. The number of model parameters is sublinear in V and we can generalize to unseen nodes! shared parameters W k k D E F shared parameters INPUT GRPH ompute graph for node ompute graph for node
29 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Inductive apability z u train on one graph Inductive node embedding generalize to new graph generalize to entirely unseen graphs e.g., train on protein interaction graph from model organism and generate embeddings on newly collected data about organism
30 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Inductive apability z u train with snapshot new node arrives generate embedding for new node Many application settings constantly encounter previously unseen nodes. e.g., Reddit, YouTube, GoogleScholar,. Need to generate new embeddings on the fly
31 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Quick Recap Recap: Generate node embeddings by aggregating neighborhood information. llows for parameter sharing in the encoder. llows for inductive learning. We saw a basic variant of this idea now we will cover some state of the art variants from the literature.
32 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Key distinctions are in how different approaches aggregate messages TRGET NODE D E INPUT GRPH What else can we put in the box? F??? D??? E F
33 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings
34 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Graph onvolutional Networks ased on material from: Kipf et al., Semisupervised lassification with Graph onvolutional Networks. ILR.
35 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Graph onvolutional Networks Kipf et al. s Graph onvolutional Networks (GNs) are a slight variation on the neighborhood aggregation idea: h k v = k X u2n(v)[v 1 h k u 1 p N(u) N(v)
36 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Graph onvolutional Networks asic Neighborhood ggregation h k v = h k v = k k X u2n(v) X u2n(v)[v h k 1 u VS. GN Neighborhood ggregation 1 N(v) + kh k v 1 1 h k u 1 p N(u) N(v) same matrix for self and neighbor embeddings per-neighbor normalization
37 Graph onvolutional Networks Empirically, they found this configuration to give the best results. More parameter sharing. Down-weights high degree neighbors. h k v = k X u2n(v)[v 1 h k u 1 p N(u) N(v) use the same transformation matrix for self and neighbor embeddings instead of simple average, normalization varies across neighbors Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW
38 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW atch Implementation an be efficiently implemented using sparse batch operations: H (k+1) = D 1 2 ÃD 1 2 H (k) W k where à = + I D ii = X j i,j O( E ) time complexity overall.
39 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings
40 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE ased on material from: Hamilton et al., Inductive Representation Learning on Large Graphs. NIPS.
41 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Idea So far we have aggregated the neighbor messages by taking their (weighted) average, can we do better? TRGET NODE? D E INPUT GRPH F??? D?? E F
42 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Idea TRGET NODE D E INPUT GRPH F ny differentiable function that maps set of vectors to a single vector. D E F h k v = k agg({h k u 1, 8u 2 N(v)}), k h k v 1
43 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Differences Simple neighborhood aggregation: 0 1 X h k h v k u 1 k N(v) + kh k v 1 GraphSGE: u2n(v) generalized aggregation concatenate self embedding and neighbor embedding h k v = W k agg {h k u 1, 8u 2 N(v)}, k h k v 1
44 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Variants Mean: Pool Transform neighbor vectors and apply symmetric vector function. element-wise mean/max agg = {Qh k u 1, 8u 2 N(v)} LSTM: agg = X u2n(v) h k 1 u N(v) pply LSTM to random permutation of neighbors. agg = LSTM [h k 1 u, 8u 2 (N(v))]
45 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings
46 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks ased on material from: Li et al., Gated Graph Sequence Neural Networks. ILR.
47 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation asic idea: Nodes aggregate messages from their neighbors using neural networks TRGET NODE D E F D E F INPUT GRPH
48 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation GNs and GraphSGE generally only 2-3 layers deep. TRGET NODE D E F D E F INPUT GRPH
49 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation ut what if we want to go deeper? TRGET NODE 10+ layers!? D E F D.. INPUT GRPH
50 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks How can we build models with many layers of neighborhood aggregation? hallenges: Overfitting from too many parameters. Vanishing/exploding gradients during backpropagation. Idea: Use techniques from modern recurrent neural networks!
51 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks Idea 1: Parameter sharing across layers. same neural network across layers TRGET NODE D E F E F. INPUT GRPH D
52 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks Idea 2: Recurrent state update. TRGET NODE RNN module! RNN module D E F E F. INPUT GRPH D
53 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW The Math Intuition: Neighborhood aggregation with RNN state update. 1. Get message from neighbors at step k: m k v = W X u2n(v) h k 1 u aggregation function does not depend on k 2. Update node state using Gated Recurrent Unit (GRU). New node state depends on the old state and the message from neighbors: h k v =GRU(h k 1 v, m k v)
54 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks TRGET NODE D E F RNN module E F. INPUT GRPH D an handle models with >20 layers. Most real-world networks have small diameters (e.g., less than 7). llows for complex information about global graph structure to be propagated to all nodes.
55 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks TRGET NODE D E F RNN module E F. INPUT GRPH D Useful for complex networks representing: Logical formulas. Programs.
56 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings
57 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Summary so far Key idea: Generate node embeddings based on local neighborhoods. TRGET NODE D E F D E F INPUT GRPH
58 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Summary so far Graph convolutional networks verage neighborhood information and stack neural networks. GraphSGE Generalized neighborhood aggregation. Gated Graph Neural Networks Neighborhood aggregation + RNNs
59 Recent advances in graph neural nets (not covered in detail here) Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW ttention-based neighborhood aggregation: Graph ttention Networks (Velickovic et al., 2018) GeniePath (Liu et al., 2018) Generalizations based on spectral convolutions: Geometric Deep Learning (ronstein et al., 2017) Mixture Model NNs (Monti et al., 2017) Speed improvements via subsampling: FastGNs (hen et al., 2018) Stochastic GNs (hen et al., 2017)
60 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings
61 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Subgraph Embeddings ased on material from: Duvenaud et al onvolutional Networks on Graphs for Learning Molecular Fingerprints. IML. Li et al Gated Graph Sequence Neural Networks. ILR.
62 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW (Sub)graph Embeddings So far we have focused on nodelevel embeddings
63 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW (Sub)graph Embeddings ut what about subgraph embeddings?
64 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW pproach 1 Simple idea: Just sum (or average) the node embeddings in the (sub)graph z S = X v2s z v Used by Duvenaud et al., 2016 to classify molecules based on their graph structure.
65 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW pproach 2 Idea: Introduce a virtual node to represent the subgraph and run a standard graph neural network. Proposed by Li et al., 2016 as a general technique for subgraph embedding.
66 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW (Sub)graph Embeddings Still an open research area! How to embed (sub)graphs with millions or billions of nodes? How to do analogue of NN pooling on networks?
Deep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 6 April 2017 joint work with Max Welling (University of Amsterdam) The success story of deep
More informationDeep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016
More informationMachine Learning 13. week
Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of
More informationThis Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured
Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 1 This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures
More informationNatural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu
Natural Language Processing CS 6320 Lecture 6 Neural Language Models Instructor: Sanda Harabagiu In this lecture We shall cover: Deep Neural Models for Natural Language Processing Introduce Feed Forward
More informationLecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa
Instructors: Parth Shah, Riju Pahwa Lecture 2 Notes Outline 1. Neural Networks The Big Idea Architecture SGD and Backpropagation 2. Convolutional Neural Networks Intuition Architecture 3. Recurrent Neural
More informationDEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla
DEEP LEARNING REVIEW Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature 2015 -Presented by Divya Chitimalla What is deep learning Deep learning allows computational models that are composed of multiple
More informationCSC 578 Neural Networks and Deep Learning
CSC 578 Neural Networks and Deep Learning Fall 2018/19 7. Recurrent Neural Networks (Some figures adapted from NNDL book) 1 Recurrent Neural Networks 1. Recurrent Neural Networks (RNNs) 2. RNN Training
More informationCOMP 551 Applied Machine Learning Lecture 16: Deep Learning
COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all
More informationMachine Learning. MGS Lecture 3: Deep Learning
Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ Machine Learning MGS Lecture 3: Deep Learning Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ WHAT IS DEEP LEARNING? Shallow network: Only one hidden layer
More informationTable of Contents. What Really is a Hidden Unit? Visualizing Feed-Forward NNs. Visualizing Convolutional NNs. Visualizing Recurrent NNs
Table of Contents What Really is a Hidden Unit? Visualizing Feed-Forward NNs Visualizing Convolutional NNs Visualizing Recurrent NNs Visualizing Attention Visualizing High Dimensional Data What do visualizations
More informationRepresentation Learning on Graphs: Methods and Applications
Representation Learning on Graphs: Methods and Applications William L. Hamilton wleif@stanford.edu Rex Ying rexying@stanford.edu Department of Computer Science Stanford University Stanford, CA, 94305 Jure
More informationDeepWalk: Online Learning of Social Representations
DeepWalk: Online Learning of Social Representations ACM SIG-KDD August 26, 2014, Rami Al-Rfou, Steven Skiena Stony Brook University Outline Introduction: Graphs as Features Language Modeling DeepWalk Evaluation:
More informationLecture 20: Neural Networks for NLP. Zubin Pahuja
Lecture 20: Neural Networks for NLP Zubin Pahuja zpahuja2@illinois.edu courses.engr.illinois.edu/cs447 CS447: Natural Language Processing 1 Today s Lecture Feed-forward neural networks as classifiers simple
More informationMachine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,
Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image
More informationResidual Networks And Attention Models. cs273b Recitation 11/11/2016. Anna Shcherbina
Residual Networks And Attention Models cs273b Recitation 11/11/2016 Anna Shcherbina Introduction to ResNets Introduced in 2015 by Microsoft Research Deep Residual Learning for Image Recognition (He, Zhang,
More informationInception and Residual Networks. Hantao Zhang. Deep Learning with Python.
Inception and Residual Networks Hantao Zhang Deep Learning with Python https://en.wikipedia.org/wiki/residual_neural_network Deep Neural Network Progress from Large Scale Visual Recognition Challenge (ILSVRC)
More informationKnow your data - many types of networks
Architectures Know your data - many types of networks Fixed length representation Variable length representation Online video sequences, or samples of different sizes Images Specific architectures for
More informationLSTM and its variants for visual recognition. Xiaodan Liang Sun Yat-sen University
LSTM and its variants for visual recognition Xiaodan Liang xdliang328@gmail.com Sun Yat-sen University Outline Context Modelling with CNN LSTM and its Variants LSTM Architecture Variants Application in
More informationDeep Learning Applications
October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning
More informationINTRODUCTION TO DEEP LEARNING
INTRODUCTION TO DEEP LEARNING CONTENTS Introduction to deep learning Contents 1. Examples 2. Machine learning 3. Neural networks 4. Deep learning 5. Convolutional neural networks 6. Conclusion 7. Additional
More informationDeep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group
Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group 1D Input, 1D Output target input 2 2D Input, 1D Output: Data Distribution Complexity Imagine many dimensions (data occupies
More informationConvolutional Neural Networks
Lecturer: Barnabas Poczos Introduction to Machine Learning (Lecture Notes) Convolutional Neural Networks Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications.
More informationCMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro
CMU 15-781 Lecture 18: Deep learning and Vision: Convolutional neural networks Teacher: Gianni A. Di Caro DEEP, SHALLOW, CONNECTED, SPARSE? Fully connected multi-layer feed-forward perceptrons: More powerful
More informationIndex. Springer Nature Switzerland AG 2019 B. Moons et al., Embedded Deep Learning,
Index A Algorithmic noise tolerance (ANT), 93 94 Application specific instruction set processors (ASIPs), 115 116 Approximate computing application level, 95 circuits-levels, 93 94 DAS and DVAS, 107 110
More informationDECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS
DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS Deep Neural Decision Forests Microsoft Research Cambridge UK, ICCV 2015 Decision Forests, Convolutional Networks and the Models in-between
More information27: Hybrid Graphical Models and Neural Networks
10-708: Probabilistic Graphical Models 10-708 Spring 2016 27: Hybrid Graphical Models and Neural Networks Lecturer: Matt Gormley Scribes: Jakob Bauer Otilia Stretcu Rohan Varma 1 Motivation We first look
More informationGraph neural networks February Visin Francesco
Graph neural networks February 2018 Visin Francesco Who I am Outline Motivation and examples Graph nets (Semi)-formal definition Interaction network Relation network Gated graph sequence neural network
More informationDeep Learning. Deep Learning. Practical Application Automatically Adding Sounds To Silent Movies
http://blog.csdn.net/zouxy09/article/details/8775360 Automatic Colorization of Black and White Images Automatically Adding Sounds To Silent Movies Traditionally this was done by hand with human effort
More informationDiffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting
Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting Yaguang Li Joint work with Rose Yu, Cyrus Shahabi, Yan Liu Page 1 Introduction Traffic congesting is wasteful of time,
More informationNeural Network Neurons
Neural Networks Neural Network Neurons 1 Receives n inputs (plus a bias term) Multiplies each input by its weight Applies activation function to the sum of results Outputs result Activation Functions Given
More informationNetwork embedding. Cheng Zheng
Network embedding Cheng Zheng Outline Problem definition Factorization based algorithms --- Laplacian Eigenmaps(NIPS, 2001) Random walk based algorithms ---DeepWalk(KDD, 2014), node2vec(kdd, 2016) Deep
More informationAsynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features
Asynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features Xu SUN ( 孙栩 ) Peking University xusun@pku.edu.cn Motivation Neural networks -> Good Performance CNN, RNN, LSTM
More informationGraphNet: Recommendation system based on language and network structure
GraphNet: Recommendation system based on language and network structure Rex Ying Stanford University rexying@stanford.edu Yuanfang Li Stanford University yli03@stanford.edu Xin Li Stanford University xinli16@stanford.edu
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Image Data: Classification via Neural Networks Instructor: Yizhou Sun yzsun@ccs.neu.edu November 19, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining
More informationSequence Modeling: Recurrent and Recursive Nets. By Pyry Takala 14 Oct 2015
Sequence Modeling: Recurrent and Recursive Nets By Pyry Takala 14 Oct 2015 Agenda Why Recurrent neural networks? Anatomy and basic training of an RNN (10.2, 10.2.1) Properties of RNNs (10.2.2, 8.2.6) Using
More informationSemi-supervised Methods for Graph Representation
Semi-supervised Methods for Graph Representation Modeling Data With Networks + Network Embedding: Problems, Methodologies and Frontiers Ivan Brugere (University of Illinois at Chicago) Peng Cui (Tsinghua
More informationDeep Learning. Architecture Design for. Sargur N. Srihari
Architecture Design for Deep Learning Sargur N. srihari@cedar.buffalo.edu 1 Topics Overview 1. Example: Learning XOR 2. Gradient-Based Learning 3. Hidden Units 4. Architecture Design 5. Backpropagation
More informationCENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 11. Sinan Kalkan
CENG 783 Special topics in Deep Learning AlchemyAPI Week 11 Sinan Kalkan TRAINING A CNN Fig: http://www.robots.ox.ac.uk/~vgg/practicals/cnn/ Feed-forward pass Note that this is written in terms of the
More informationAdvanced Introduction to Machine Learning, CMU-10715
Advanced Introduction to Machine Learning, CMU-10715 Deep Learning Barnabás Póczos, Sept 17 Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio
More informationNeural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani
Neural Networks CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Biological and artificial neural networks Feed-forward neural networks Single layer
More informationA Higher-Order Graph Convolutional Layer
A Higher-Order Graph Convolutional Layer Sami Abu-El-Haija 1, Nazanin Alipourfard 1, Hrayr Harutyunyan 1, Amol Kapoor 2, Bryan Perozzi 2 1 Information Sciences Institute University of Southern California
More informationDeep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?
Data Mining Deep Learning Deep Learning provided breakthrough results in speech recognition and image classification. Why? Because Speech recognition and image classification are two basic examples of
More informationPerceptron: This is convolution!
Perceptron: This is convolution! v v v Shared weights v Filter = local perceptron. Also called kernel. By pooling responses at different locations, we gain robustness to the exact spatial location of image
More informationCombining Neural Networks and Log-linear Models to Improve Relation Extraction
Combining Neural Networks and Log-linear Models to Improve Relation Extraction Thien Huu Nguyen and Ralph Grishman Computer Science Department, New York University {thien,grishman}@cs.nyu.edu Outline Relation
More informationDeep Learning. Volker Tresp Summer 2014
Deep Learning Volker Tresp Summer 2014 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there
More informationRecurrent Convolutional Neural Networks for Scene Labeling
Recurrent Convolutional Neural Networks for Scene Labeling Pedro O. Pinheiro, Ronan Collobert Reviewed by Yizhe Zhang August 14, 2015 Scene labeling task Scene labeling: assign a class label to each pixel
More informationCan Active Memory Replace Attention?
Google Brain NIPS 2016 Presenter: Chao Jiang NIPS 2016 Presenter: Chao Jiang 1 / Outline 1 Introduction 2 Active Memory 3 Step by Step to Neural GPU 4 Another two steps: 1. the Markovian Neural GPU 5 Another
More informationSu et al. Shape Descriptors - III
Su et al. Shape Descriptors - III Siddhartha Chaudhuri http://www.cse.iitb.ac.in/~cs749 Funkhouser; Feng, Liu, Gong Recap Global A shape descriptor is a set of numbers that describes a shape in a way that
More informationFace Recognition A Deep Learning Approach
Face Recognition A Deep Learning Approach Lihi Shiloh Tal Perl Deep Learning Seminar 2 Outline What about Cat recognition? Classical face recognition Modern face recognition DeepFace FaceNet Comparison
More informationImageNet Classification with Deep Convolutional Neural Networks
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky Ilya Sutskever Geoffrey Hinton University of Toronto Canada Paper with same name to appear in NIPS 2012 Main idea Architecture
More informationDeep Learning in Visual Recognition. Thanks Da Zhang for the slides
Deep Learning in Visual Recognition Thanks Da Zhang for the slides Deep Learning is Everywhere 2 Roadmap Introduction Convolutional Neural Network Application Image Classification Object Detection Object
More informationDistributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability
Distributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability Janis Keuper Itwm.fraunhofer.de/ml Competence Center High Performance Computing Fraunhofer ITWM, Kaiserslautern,
More informationXES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework
XES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework Demo Paper Joerg Evermann 1, Jana-Rebecca Rehse 2,3, and Peter Fettke 2,3 1 Memorial University of Newfoundland 2 German Research
More informationConvolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech
Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:
More informationEmpirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling
Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling Authors: Junyoung Chung, Caglar Gulcehre, KyungHyun Cho and Yoshua Bengio Presenter: Yu-Wei Lin Background: Recurrent Neural
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationCS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes
CS 6501: Deep Learning for Computer Graphics Training Neural Networks II Connelly Barnes Overview Preprocessing Initialization Vanishing/exploding gradients problem Batch normalization Dropout Additional
More informationSEMANTIC COMPUTING. Lecture 9: Deep Learning: Recurrent Neural Networks (RNNs) TU Dresden, 21 December 2018
SEMANTIC COMPUTING Lecture 9: Deep Learning: Recurrent Neural Networks (RNNs) Dagmar Gromann International Center For Computational Logic TU Dresden, 21 December 2018 Overview Handling Overfitting Recurrent
More informationCOMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017
COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization
More informationLecture 10 CNNs on Graphs
Lecture 10 CNNs on Graphs CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago April 26, 2017 Two Scenarios For CNNs on graphs, we have two distinct scenarios: Scenario 1: Each
More informationCode Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:
Code Mania 2019 Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: 1. Introduction to Artificial Intelligence 2. Introduction to python programming and Environment
More informationCS489/698: Intro to ML
CS489/698: Intro to ML Lecture 14: Training of Deep NNs Instructor: Sun Sun 1 Outline Activation functions Regularization Gradient-based optimization 2 Examples of activation functions 3 5/28/18 Sun Sun
More informationDeepFace: Closing the Gap to Human-Level Performance in Face Verification
DeepFace: Closing the Gap to Human-Level Performance in Face Verification Report on the paper Artem Komarichev February 7, 2016 Outline New alignment technique New DNN architecture New large dataset with
More informationArtificial Intelligence
Artificial Intelligence AI & Machine Learning & Neural Nets Marc Toussaint University of Stuttgart Winter 2018/19 Motivation: Neural networks became a central topic for Machine Learning and AI. But in
More informationData Mining and Analytics
Data Mining and Analytics Aik Choon Tan, Ph.D. Associate Professor of Bioinformatics Division of Medical Oncology Department of Medicine aikchoon.tan@ucdenver.edu 9/22/2017 http://tanlab.ucdenver.edu/labhomepage/teaching/bsbt6111/
More informationEECS 496 Statistical Language Models. Winter 2018
EECS 496 Statistical Language Models Winter 2018 Introductions Professor: Doug Downey Course web site: www.cs.northwestern.edu/~ddowney/courses/496_winter2018 (linked off prof. home page) Logistics Grading
More informationObject Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal
Object Detection Lecture 10.3 - Introduction to deep learning (CNN) Idar Dyrdal Deep Learning Labels Computational models composed of multiple processing layers (non-linear transformations) Used to learn
More informationIntroduction to Deep Learning
ENEE698A : Machine Learning Seminar Introduction to Deep Learning Raviteja Vemulapalli Image credit: [LeCun 1998] Resources Unsupervised feature learning and deep learning (UFLDL) tutorial (http://ufldl.stanford.edu/wiki/index.php/ufldl_tutorial)
More informationSlide credit from Hung-Yi Lee & Richard Socher
Slide credit from Hung-Yi Lee & Richard Socher 1 Review Word Vector 2 Word2Vec Variants Skip-gram: predicting surrounding words given the target word (Mikolov+, 2013) CBOW (continuous bag-of-words): predicting
More informationFastText. Jon Koss, Abhishek Jindal
FastText Jon Koss, Abhishek Jindal FastText FastText is on par with state-of-the-art deep learning classifiers in terms of accuracy But it is way faster: FastText can train on more than one billion words
More informationBayesian model ensembling using meta-trained recurrent neural networks
Bayesian model ensembling using meta-trained recurrent neural networks Luca Ambrogioni l.ambrogioni@donders.ru.nl Umut Güçlü u.guclu@donders.ru.nl Yağmur Güçlütürk y.gucluturk@donders.ru.nl Julia Berezutskaya
More informationDay 3 Lecture 1. Unsupervised Learning
Day 3 Lecture 1 Unsupervised Learning Semi-supervised and transfer learning Myth: you can t do deep learning unless you have a million labelled examples for your problem. Reality You can learn useful representations
More informationEncoding RNNs, 48 End of sentence (EOS) token, 207 Exploding gradient, 131 Exponential function, 42 Exponential Linear Unit (ELU), 44
A Activation potential, 40 Annotated corpus add padding, 162 check versions, 158 create checkpoints, 164, 166 create input, 160 create train and validation datasets, 163 dropout, 163 DRUG-AE.rel file,
More informationDeep Learning for Computer Vision II
IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L
More informationDEEP LEARNING IN PYTHON. The need for optimization
DEEP LEARNING IN PYTHON The need for optimization A baseline neural network Input 2 Hidden Layer 5 2 Output - 9-3 Actual Value of Target: 3 Error: Actual - Predicted = 4 A baseline neural network Input
More informationHierarchical Video Frame Sequence Representation with Deep Convolutional Graph Network
Hierarchical Video Frame Sequence Representation with Deep Convolutional Graph Network Feng Mao [0000 0001 6171 3168], Xiang Wu [0000 0003 2698 2156], Hui Xue, and Rong Zhang Alibaba Group, Hangzhou, China
More informationDeep Neural Networks Applications in Handwriting Recognition
Deep Neural Networks Applications in Handwriting Recognition 2 Who am I? Théodore Bluche PhD defended at Université Paris-Sud last year Deep Neural Networks for Large Vocabulary Handwritten
More informationInductive Representation Learning on Large Graphs
Inductive Representation Learning on Large Graphs William L. Hamilton wleif@stanford.edu Rex Ying rexying@stanford.edu Jure Leskovec jure@cs.stanford.edu Department of Computer Science Stanford University
More informationDeep Learning and Its Applications
Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent
More informationDeep (1) Matthieu Cord LIP6 / UPMC Paris 6
Deep (1) Matthieu Cord LIP6 / UPMC Paris 6 Syllabus 1. Whole traditional (old) visual recognition pipeline 2. Introduction to Neural Nets 3. Deep Nets for image classification To do : Voir la leçon inaugurale
More informationEmpirical Evaluation of RNN Architectures on Sentence Classification Task
Empirical Evaluation of RNN Architectures on Sentence Classification Task Lei Shen, Junlin Zhang Chanjet Information Technology lorashen@126.com, zhangjlh@chanjet.com Abstract. Recurrent Neural Networks
More informationThe Mathematics Behind Neural Networks
The Mathematics Behind Neural Networks Pattern Recognition and Machine Learning by Christopher M. Bishop Student: Shivam Agrawal Mentor: Nathaniel Monson Courtesy of xkcd.com The Black Box Training the
More informationMachine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15
Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15 Slides adapted from Jordan Boyd-Graber Machine Learning: Chenhao Tan Boulder 1 of 21 Logistics HW3 available on Github, due on October
More informationLayerwise Interweaving Convolutional LSTM
Layerwise Interweaving Convolutional LSTM Tiehang Duan and Sargur N. Srihari Department of Computer Science and Engineering The State University of New York at Buffalo Buffalo, NY 14260, United States
More informationAkarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction
Akarsh Pokkunuru EECS Department 03-16-2017 Contractive Auto-Encoders: Explicit Invariance During Feature Extraction 1 AGENDA Introduction to Auto-encoders Types of Auto-encoders Analysis of different
More informationAdvanced Video Analysis & Imaging
Advanced Video Analysis & Imaging (5LSH0), Module 09B Machine Learning with Convolutional Neural Networks (CNNs) - Workout Farhad G. Zanjani, Clint Sebastian, Egor Bondarev, Peter H.N. de With ( p.h.n.de.with@tue.nl
More informationNatural Language Processing with Deep Learning CS224N/Ling284. Christopher Manning Lecture 4: Backpropagation and computation graphs
Natural Language Processing with Deep Learning CS4N/Ling84 Christopher Manning Lecture 4: Backpropagation and computation graphs Lecture Plan Lecture 4: Backpropagation and computation graphs 1. Matrix
More informationarxiv: v1 [cond-mat.dis-nn] 30 Dec 2018
A General Deep Learning Framework for Structure and Dynamics Reconstruction from Time Series Data arxiv:1812.11482v1 [cond-mat.dis-nn] 30 Dec 2018 Zhang Zhang, Jing Liu, Shuo Wang, Ruyue Xin, Jiang Zhang
More informationReport: Privacy-Preserving Classification on Deep Neural Network
Report: Privacy-Preserving Classification on Deep Neural Network Janno Veeorg Supervised by Helger Lipmaa and Raul Vicente Zafra May 25, 2017 1 Introduction In this report we consider following task: how
More informationLARGE MARGIN CLASSIFIERS
Admin Assignment 5 LARGE MARGIN CLASSIFIERS David Kauchak CS 451 Fall 2013 Midterm Download from course web page when you re ready to take it 2 hours to complete Must hand-in (or e-mail in) by 11:59pm
More informationFacial Expression Classification with Random Filters Feature Extraction
Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle
More informationECG782: Multidimensional Digital Signal Processing
ECG782: Multidimensional Digital Signal Processing Object Recognition http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Knowledge Representation Statistical Pattern Recognition Neural Networks Boosting
More informationRestricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version
Shallow vs. deep networks Restricted Boltzmann Machines Shallow: one hidden layer Features can be learned more-or-less independently Arbitrary function approximator (with enough hidden units) Deep: two
More informationLecture 7: Neural network acoustic models in speech recognition
CS 224S / LINGUIST 285 Spoken Language Processing Andrew Maas Stanford University Spring 2017 Lecture 7: Neural network acoustic models in speech recognition Outline Hybrid acoustic modeling overview Basic
More informationKnow your neighbours: Machine Learning on Graphs
Know your neighbours: Machine Learning on Graphs Andrew Docherty Senior Research Engineer andrew.docherty@data61.csiro.au www.data61.csiro.au 2 Graphs are Everywhere Online Social Networks Transportation
More informationCSE 559A: Computer Vision
CSE 559A: Computer Vision Fall 2018: T-R: 11:30-1pm @ Lopata 101 Instructor: Ayan Chakrabarti (ayan@wustl.edu). Course Staff: Zhihao Xia, Charlie Wu, Han Liu http://www.cse.wustl.edu/~ayan/courses/cse559a/
More informationArtificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu
Artificial Neural Networks Introduction to Computational Neuroscience Ardi Tampuu 7.0.206 Artificial neural network NB! Inspired by biology, not based on biology! Applications Automatic speech recognition
More informationDeep Learning with Tensorflow AlexNet
Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification
More informationIndex. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,
A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,
More information