This Talk. 1) Node embeddings. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured

Size: px
Start display at page:

Download "This Talk. 1) Node embeddings. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured"

Transcription

1 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures for graphstructured data 3) pplications

2 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Part 2: Graph Neural Networks

3 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Embedding Nodes Goal is to encode nodes so that similarity in the embedding space (e.g., dot product) approximates similarity in the original network.

4 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Embedding Nodes Goal: similarity(u, v) z > v z u Need to define!

5 Two Key omponents Encoder maps each node to a lowdimensional vector. enc(v) =z v node in the input graph Similarity function specifies how relationships in vector space map to relationships in the original network. Similarity of u and v in the original network d-dimensional embedding similarity(u, v) z > v z u dot product between node embeddings Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW

6 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW From Shallow to Deep So far we have focused on shallow encoders, i.e. embedding lookups: embedding matrix embedding vector for a specific node Z = Dimension/size of embeddings one column per node

7 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW From Shallow to Deep Limitations of shallow encoding: O( V ) parameters are needed: there no parameter sharing and every node has its own unique embedding vector. Inherently transductive : It is impossible to generate embeddings for nodes that were not seen during training. Do not incorporate node features: Many graphs have features that we can and should leverage.

8 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW From Shallow to Deep We will now discuss deeper methods based on graph neural networks. enc(v) = complex function that depends on graph structure. In general, all of these more complex encoders can be combined with the similarity functions from the previous section.

9 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section We will now discuss deeper methods based on graph neural networks. 1. The asics 2. Graph onvolutional Networks (GNs) 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings

10 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW The asics: Graph Neural Networks ased on material from: Hamilton et al Representation Learning on Graphs: Methods and pplications. IEEE Data Engineering ulletin on Graph Systems. Scarselli et al The Graph Neural Network Model. IEEE Transactions on Neural Networks.

11 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Setup ssume we have a graph G: V is the vertex set. is the adjacency matrix (assume binary). X R m V is a matrix of node features. ategorical attributes, text, image data E.g., profile information in a social network. Node degrees, clustering coefficients, etc. Indicator vectors (i.e., one-hot encoding of each node)

12 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Key idea: Generate node embeddings based on local neighborhoods. TRGET NODE D E F D E F INPUT GRPH

13 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Intuition: Nodes aggregate information from their neighbors using neural networks TRGET NODE D E F D E F INPUT GRPH

14 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Intuition: Network neighborhood defines a computation graph Every node defines a unique computation graph!

15 Neighborhood ggregation Nodes have embeddings at each layer. Model can be arbitrary depth. layer-0 embedding of node u is its input feature, i.e. x u. TRGET NODE D E INPUT GRPH F Layer-2 Layer-1 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW D Layer-0 x F E x x x x E x F x

16 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood onvolutions Neighborhood aggregation can be viewed as a center-surround filter. Mathematically related to spectral graph convolutions (see ronstein et al., 2017)

17 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Key distinctions are in how different approaches aggregate information across the layers. what s in the box!? TRGET NODE? D E INPUT GRPH F??? D?? E F

18 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation asic approach: verage neighbor information and apply a neural network. TRGET NODE 1) average messages from neighbors D E F D E F INPUT GRPH 2) apply neural network

19 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW The Math asic approach: verage neighbor messages and apply a neural network. h 0 v = x v 0 h k v k kth layer embedding non-linearity (e.g., of v ReLU or tanh) Initial layer 0 embeddings are equal to node features X u2n(v) h k 1 u average of neighbor s previous layer embeddings previous layer embedding of v 1 N(v) + kh k v 1, 8k >0

20 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model How do we train the model to generate high-quality embeddings? z Need to define a loss function on the embeddings, L(z u )!

21 Training the Model h 0 v = x v 0 h k v k X u2n(v) trainable matrices (i.e., what we learn) h k 1 u 1 N(v) + kh k v 1, 8k 2 {1,...,K} z v = h K v fter K-layers of neighborhood aggregation, we get output embeddings for each node. We can feed these embeddings into any loss function and run stochastic gradient descent to train the aggregation parameters. Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW

22 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model Train in an unsupervised manner using only the graph structure. Unsupervised loss function can be anything from the last section, e.g., based on Random walks (node2vec, DeepWalk) Graph factorization i.e., train the model so that similar nodes have similar embeddings.

23 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model lternative: Directly train the model for a supervised task (e.g., node classification): Human or bot? Human or bot? e.g., an online social network

24 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Training the Model lternative: Directly train the model for a supervised task (e.g., node classification): Human or bot? L = X v2v classification weights y v log( (z > v )) + (1 y v )log(1 (z > v )) output node embedding node class label

25 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Overview of Model Design 1) Define a neighborhood aggregation function. z 2) Define a loss function on the embeddings, L(z u )

26 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Overview of Model Design 3) Train on a set of nodes, i.e., a batch of compute graphs

27 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Overview of Model 4) Generate embeddings for nodes as needed Even for nodes we never trained on!!!!

28 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Inductive apability The same aggregation parameters are shared for all nodes. The number of model parameters is sublinear in V and we can generalize to unseen nodes! shared parameters W k k D E F shared parameters INPUT GRPH ompute graph for node ompute graph for node

29 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Inductive apability z u train on one graph Inductive node embedding generalize to new graph generalize to entirely unseen graphs e.g., train on protein interaction graph from model organism and generate embeddings on newly collected data about organism

30 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Inductive apability z u train with snapshot new node arrives generate embedding for new node Many application settings constantly encounter previously unseen nodes. e.g., Reddit, YouTube, GoogleScholar,. Need to generate new embeddings on the fly

31 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Quick Recap Recap: Generate node embeddings by aggregating neighborhood information. llows for parameter sharing in the encoder. llows for inductive learning. We saw a basic variant of this idea now we will cover some state of the art variants from the literature.

32 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation Key distinctions are in how different approaches aggregate messages TRGET NODE D E INPUT GRPH What else can we put in the box? F??? D??? E F

33 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings

34 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Graph onvolutional Networks ased on material from: Kipf et al., Semisupervised lassification with Graph onvolutional Networks. ILR.

35 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Graph onvolutional Networks Kipf et al. s Graph onvolutional Networks (GNs) are a slight variation on the neighborhood aggregation idea: h k v = k X u2n(v)[v 1 h k u 1 p N(u) N(v)

36 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Graph onvolutional Networks asic Neighborhood ggregation h k v = h k v = k k X u2n(v) X u2n(v)[v h k 1 u VS. GN Neighborhood ggregation 1 N(v) + kh k v 1 1 h k u 1 p N(u) N(v) same matrix for self and neighbor embeddings per-neighbor normalization

37 Graph onvolutional Networks Empirically, they found this configuration to give the best results. More parameter sharing. Down-weights high degree neighbors. h k v = k X u2n(v)[v 1 h k u 1 p N(u) N(v) use the same transformation matrix for self and neighbor embeddings instead of simple average, normalization varies across neighbors Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW

38 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW atch Implementation an be efficiently implemented using sparse batch operations: H (k+1) = D 1 2 ÃD 1 2 H (k) W k where à = + I D ii = X j i,j O( E ) time complexity overall.

39 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings

40 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE ased on material from: Hamilton et al., Inductive Representation Learning on Large Graphs. NIPS.

41 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Idea So far we have aggregated the neighbor messages by taking their (weighted) average, can we do better? TRGET NODE? D E INPUT GRPH F??? D?? E F

42 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Idea TRGET NODE D E INPUT GRPH F ny differentiable function that maps set of vectors to a single vector. D E F h k v = k agg({h k u 1, 8u 2 N(v)}), k h k v 1

43 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Differences Simple neighborhood aggregation: 0 1 X h k h v k u 1 k N(v) + kh k v 1 GraphSGE: u2n(v) generalized aggregation concatenate self embedding and neighbor embedding h k v = W k agg {h k u 1, 8u 2 N(v)}, k h k v 1

44 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW GraphSGE Variants Mean: Pool Transform neighbor vectors and apply symmetric vector function. element-wise mean/max agg = {Qh k u 1, 8u 2 N(v)} LSTM: agg = X u2n(v) h k 1 u N(v) pply LSTM to random permutation of neighbors. agg = LSTM [h k 1 u, 8u 2 (N(v))]

45 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings

46 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks ased on material from: Li et al., Gated Graph Sequence Neural Networks. ILR.

47 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation asic idea: Nodes aggregate messages from their neighbors using neural networks TRGET NODE D E F D E F INPUT GRPH

48 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation GNs and GraphSGE generally only 2-3 layers deep. TRGET NODE D E F D E F INPUT GRPH

49 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Neighborhood ggregation ut what if we want to go deeper? TRGET NODE 10+ layers!? D E F D.. INPUT GRPH

50 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks How can we build models with many layers of neighborhood aggregation? hallenges: Overfitting from too many parameters. Vanishing/exploding gradients during backpropagation. Idea: Use techniques from modern recurrent neural networks!

51 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks Idea 1: Parameter sharing across layers. same neural network across layers TRGET NODE D E F E F. INPUT GRPH D

52 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks Idea 2: Recurrent state update. TRGET NODE RNN module! RNN module D E F E F. INPUT GRPH D

53 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW The Math Intuition: Neighborhood aggregation with RNN state update. 1. Get message from neighbors at step k: m k v = W X u2n(v) h k 1 u aggregation function does not depend on k 2. Update node state using Gated Recurrent Unit (GRU). New node state depends on the old state and the message from neighbors: h k v =GRU(h k 1 v, m k v)

54 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks TRGET NODE D E F RNN module E F. INPUT GRPH D an handle models with >20 layers. Most real-world networks have small diameters (e.g., less than 7). llows for complex information about global graph structure to be propagated to all nodes.

55 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Gated Graph Neural Networks TRGET NODE D E F RNN module E F. INPUT GRPH D Useful for complex networks representing: Logical formulas. Programs.

56 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings

57 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Summary so far Key idea: Generate node embeddings based on local neighborhoods. TRGET NODE D E F D E F INPUT GRPH

58 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Summary so far Graph convolutional networks verage neighborhood information and stack neural networks. GraphSGE Generalized neighborhood aggregation. Gated Graph Neural Networks Neighborhood aggregation + RNNs

59 Recent advances in graph neural nets (not covered in detail here) Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW ttention-based neighborhood aggregation: Graph ttention Networks (Velickovic et al., 2018) GeniePath (Liu et al., 2018) Generalizations based on spectral convolutions: Geometric Deep Learning (ronstein et al., 2017) Mixture Model NNs (Monti et al., 2017) Speed improvements via subsampling: FastGNs (hen et al., 2018) Stochastic GNs (hen et al., 2017)

60 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Outline for this Section 1. The asics 2. Graph onvolutional Networks 3. GraphSGE 4. Gated Graph Neural Networks 5. Subgraph Embeddings

61 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW Subgraph Embeddings ased on material from: Duvenaud et al onvolutional Networks on Graphs for Learning Molecular Fingerprints. IML. Li et al Gated Graph Sequence Neural Networks. ILR.

62 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW (Sub)graph Embeddings So far we have focused on nodelevel embeddings

63 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW (Sub)graph Embeddings ut what about subgraph embeddings?

64 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW pproach 1 Simple idea: Just sum (or average) the node embeddings in the (sub)graph z S = X v2s z v Used by Duvenaud et al., 2016 to classify molecules based on their graph structure.

65 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW pproach 2 Idea: Introduce a virtual node to represent the subgraph and run a standard graph neural network. Proposed by Li et al., 2016 as a general technique for subgraph embedding.

66 Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW (Sub)graph Embeddings Still an open research area! How to embed (sub)graphs with millions or billions of nodes? How to do analogue of NN pooling on networks?

Deep Learning on Graphs

Deep Learning on Graphs Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 6 April 2017 joint work with Max Welling (University of Amsterdam) The success story of deep

More information

Deep Learning on Graphs

Deep Learning on Graphs Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016

More information

Machine Learning 13. week

Machine Learning 13. week Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of

More information

This Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured

This Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 1 This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures

More information

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu Natural Language Processing CS 6320 Lecture 6 Neural Language Models Instructor: Sanda Harabagiu In this lecture We shall cover: Deep Neural Models for Natural Language Processing Introduce Feed Forward

More information

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa Instructors: Parth Shah, Riju Pahwa Lecture 2 Notes Outline 1. Neural Networks The Big Idea Architecture SGD and Backpropagation 2. Convolutional Neural Networks Intuition Architecture 3. Recurrent Neural

More information

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla DEEP LEARNING REVIEW Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature 2015 -Presented by Divya Chitimalla What is deep learning Deep learning allows computational models that are composed of multiple

More information

CSC 578 Neural Networks and Deep Learning

CSC 578 Neural Networks and Deep Learning CSC 578 Neural Networks and Deep Learning Fall 2018/19 7. Recurrent Neural Networks (Some figures adapted from NNDL book) 1 Recurrent Neural Networks 1. Recurrent Neural Networks (RNNs) 2. RNN Training

More information

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all

More information

Machine Learning. MGS Lecture 3: Deep Learning

Machine Learning. MGS Lecture 3: Deep Learning Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ Machine Learning MGS Lecture 3: Deep Learning Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ WHAT IS DEEP LEARNING? Shallow network: Only one hidden layer

More information

Table of Contents. What Really is a Hidden Unit? Visualizing Feed-Forward NNs. Visualizing Convolutional NNs. Visualizing Recurrent NNs

Table of Contents. What Really is a Hidden Unit? Visualizing Feed-Forward NNs. Visualizing Convolutional NNs. Visualizing Recurrent NNs Table of Contents What Really is a Hidden Unit? Visualizing Feed-Forward NNs Visualizing Convolutional NNs Visualizing Recurrent NNs Visualizing Attention Visualizing High Dimensional Data What do visualizations

More information

Representation Learning on Graphs: Methods and Applications

Representation Learning on Graphs: Methods and Applications Representation Learning on Graphs: Methods and Applications William L. Hamilton wleif@stanford.edu Rex Ying rexying@stanford.edu Department of Computer Science Stanford University Stanford, CA, 94305 Jure

More information

DeepWalk: Online Learning of Social Representations

DeepWalk: Online Learning of Social Representations DeepWalk: Online Learning of Social Representations ACM SIG-KDD August 26, 2014, Rami Al-Rfou, Steven Skiena Stony Brook University Outline Introduction: Graphs as Features Language Modeling DeepWalk Evaluation:

More information

Lecture 20: Neural Networks for NLP. Zubin Pahuja

Lecture 20: Neural Networks for NLP. Zubin Pahuja Lecture 20: Neural Networks for NLP Zubin Pahuja zpahuja2@illinois.edu courses.engr.illinois.edu/cs447 CS447: Natural Language Processing 1 Today s Lecture Feed-forward neural networks as classifiers simple

More information

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU, Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image

More information

Residual Networks And Attention Models. cs273b Recitation 11/11/2016. Anna Shcherbina

Residual Networks And Attention Models. cs273b Recitation 11/11/2016. Anna Shcherbina Residual Networks And Attention Models cs273b Recitation 11/11/2016 Anna Shcherbina Introduction to ResNets Introduced in 2015 by Microsoft Research Deep Residual Learning for Image Recognition (He, Zhang,

More information

Inception and Residual Networks. Hantao Zhang. Deep Learning with Python.

Inception and Residual Networks. Hantao Zhang. Deep Learning with Python. Inception and Residual Networks Hantao Zhang Deep Learning with Python https://en.wikipedia.org/wiki/residual_neural_network Deep Neural Network Progress from Large Scale Visual Recognition Challenge (ILSVRC)

More information

Know your data - many types of networks

Know your data - many types of networks Architectures Know your data - many types of networks Fixed length representation Variable length representation Online video sequences, or samples of different sizes Images Specific architectures for

More information

LSTM and its variants for visual recognition. Xiaodan Liang Sun Yat-sen University

LSTM and its variants for visual recognition. Xiaodan Liang Sun Yat-sen University LSTM and its variants for visual recognition Xiaodan Liang xdliang328@gmail.com Sun Yat-sen University Outline Context Modelling with CNN LSTM and its Variants LSTM Architecture Variants Application in

More information

Deep Learning Applications

Deep Learning Applications October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning

More information

INTRODUCTION TO DEEP LEARNING

INTRODUCTION TO DEEP LEARNING INTRODUCTION TO DEEP LEARNING CONTENTS Introduction to deep learning Contents 1. Examples 2. Machine learning 3. Neural networks 4. Deep learning 5. Convolutional neural networks 6. Conclusion 7. Additional

More information

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group 1D Input, 1D Output target input 2 2D Input, 1D Output: Data Distribution Complexity Imagine many dimensions (data occupies

More information

Convolutional Neural Networks

Convolutional Neural Networks Lecturer: Barnabas Poczos Introduction to Machine Learning (Lecture Notes) Convolutional Neural Networks Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications.

More information

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro CMU 15-781 Lecture 18: Deep learning and Vision: Convolutional neural networks Teacher: Gianni A. Di Caro DEEP, SHALLOW, CONNECTED, SPARSE? Fully connected multi-layer feed-forward perceptrons: More powerful

More information

Index. Springer Nature Switzerland AG 2019 B. Moons et al., Embedded Deep Learning,

Index. Springer Nature Switzerland AG 2019 B. Moons et al., Embedded Deep Learning, Index A Algorithmic noise tolerance (ANT), 93 94 Application specific instruction set processors (ASIPs), 115 116 Approximate computing application level, 95 circuits-levels, 93 94 DAS and DVAS, 107 110

More information

DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS

DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS Deep Neural Decision Forests Microsoft Research Cambridge UK, ICCV 2015 Decision Forests, Convolutional Networks and the Models in-between

More information

27: Hybrid Graphical Models and Neural Networks

27: Hybrid Graphical Models and Neural Networks 10-708: Probabilistic Graphical Models 10-708 Spring 2016 27: Hybrid Graphical Models and Neural Networks Lecturer: Matt Gormley Scribes: Jakob Bauer Otilia Stretcu Rohan Varma 1 Motivation We first look

More information

Graph neural networks February Visin Francesco

Graph neural networks February Visin Francesco Graph neural networks February 2018 Visin Francesco Who I am Outline Motivation and examples Graph nets (Semi)-formal definition Interaction network Relation network Gated graph sequence neural network

More information

Deep Learning. Deep Learning. Practical Application Automatically Adding Sounds To Silent Movies

Deep Learning. Deep Learning. Practical Application Automatically Adding Sounds To Silent Movies http://blog.csdn.net/zouxy09/article/details/8775360 Automatic Colorization of Black and White Images Automatically Adding Sounds To Silent Movies Traditionally this was done by hand with human effort

More information

Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting

Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting Yaguang Li Joint work with Rose Yu, Cyrus Shahabi, Yan Liu Page 1 Introduction Traffic congesting is wasteful of time,

More information

Neural Network Neurons

Neural Network Neurons Neural Networks Neural Network Neurons 1 Receives n inputs (plus a bias term) Multiplies each input by its weight Applies activation function to the sum of results Outputs result Activation Functions Given

More information

Network embedding. Cheng Zheng

Network embedding. Cheng Zheng Network embedding Cheng Zheng Outline Problem definition Factorization based algorithms --- Laplacian Eigenmaps(NIPS, 2001) Random walk based algorithms ---DeepWalk(KDD, 2014), node2vec(kdd, 2016) Deep

More information

Asynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features

Asynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features Asynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features Xu SUN ( 孙栩 ) Peking University xusun@pku.edu.cn Motivation Neural networks -> Good Performance CNN, RNN, LSTM

More information

GraphNet: Recommendation system based on language and network structure

GraphNet: Recommendation system based on language and network structure GraphNet: Recommendation system based on language and network structure Rex Ying Stanford University rexying@stanford.edu Yuanfang Li Stanford University yli03@stanford.edu Xin Li Stanford University xinli16@stanford.edu

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Image Data: Classification via Neural Networks Instructor: Yizhou Sun yzsun@ccs.neu.edu November 19, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining

More information

Sequence Modeling: Recurrent and Recursive Nets. By Pyry Takala 14 Oct 2015

Sequence Modeling: Recurrent and Recursive Nets. By Pyry Takala 14 Oct 2015 Sequence Modeling: Recurrent and Recursive Nets By Pyry Takala 14 Oct 2015 Agenda Why Recurrent neural networks? Anatomy and basic training of an RNN (10.2, 10.2.1) Properties of RNNs (10.2.2, 8.2.6) Using

More information

Semi-supervised Methods for Graph Representation

Semi-supervised Methods for Graph Representation Semi-supervised Methods for Graph Representation Modeling Data With Networks + Network Embedding: Problems, Methodologies and Frontiers Ivan Brugere (University of Illinois at Chicago) Peng Cui (Tsinghua

More information

Deep Learning. Architecture Design for. Sargur N. Srihari

Deep Learning. Architecture Design for. Sargur N. Srihari Architecture Design for Deep Learning Sargur N. srihari@cedar.buffalo.edu 1 Topics Overview 1. Example: Learning XOR 2. Gradient-Based Learning 3. Hidden Units 4. Architecture Design 5. Backpropagation

More information

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 11. Sinan Kalkan

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 11. Sinan Kalkan CENG 783 Special topics in Deep Learning AlchemyAPI Week 11 Sinan Kalkan TRAINING A CNN Fig: http://www.robots.ox.ac.uk/~vgg/practicals/cnn/ Feed-forward pass Note that this is written in terms of the

More information

Advanced Introduction to Machine Learning, CMU-10715

Advanced Introduction to Machine Learning, CMU-10715 Advanced Introduction to Machine Learning, CMU-10715 Deep Learning Barnabás Póczos, Sept 17 Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio

More information

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Neural Networks CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Biological and artificial neural networks Feed-forward neural networks Single layer

More information

A Higher-Order Graph Convolutional Layer

A Higher-Order Graph Convolutional Layer A Higher-Order Graph Convolutional Layer Sami Abu-El-Haija 1, Nazanin Alipourfard 1, Hrayr Harutyunyan 1, Amol Kapoor 2, Bryan Perozzi 2 1 Information Sciences Institute University of Southern California

More information

Deep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?

Deep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why? Data Mining Deep Learning Deep Learning provided breakthrough results in speech recognition and image classification. Why? Because Speech recognition and image classification are two basic examples of

More information

Perceptron: This is convolution!

Perceptron: This is convolution! Perceptron: This is convolution! v v v Shared weights v Filter = local perceptron. Also called kernel. By pooling responses at different locations, we gain robustness to the exact spatial location of image

More information

Combining Neural Networks and Log-linear Models to Improve Relation Extraction

Combining Neural Networks and Log-linear Models to Improve Relation Extraction Combining Neural Networks and Log-linear Models to Improve Relation Extraction Thien Huu Nguyen and Ralph Grishman Computer Science Department, New York University {thien,grishman}@cs.nyu.edu Outline Relation

More information

Deep Learning. Volker Tresp Summer 2014

Deep Learning. Volker Tresp Summer 2014 Deep Learning Volker Tresp Summer 2014 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there

More information

Recurrent Convolutional Neural Networks for Scene Labeling

Recurrent Convolutional Neural Networks for Scene Labeling Recurrent Convolutional Neural Networks for Scene Labeling Pedro O. Pinheiro, Ronan Collobert Reviewed by Yizhe Zhang August 14, 2015 Scene labeling task Scene labeling: assign a class label to each pixel

More information

Can Active Memory Replace Attention?

Can Active Memory Replace Attention? Google Brain NIPS 2016 Presenter: Chao Jiang NIPS 2016 Presenter: Chao Jiang 1 / Outline 1 Introduction 2 Active Memory 3 Step by Step to Neural GPU 4 Another two steps: 1. the Markovian Neural GPU 5 Another

More information

Su et al. Shape Descriptors - III

Su et al. Shape Descriptors - III Su et al. Shape Descriptors - III Siddhartha Chaudhuri http://www.cse.iitb.ac.in/~cs749 Funkhouser; Feng, Liu, Gong Recap Global A shape descriptor is a set of numbers that describes a shape in a way that

More information

Face Recognition A Deep Learning Approach

Face Recognition A Deep Learning Approach Face Recognition A Deep Learning Approach Lihi Shiloh Tal Perl Deep Learning Seminar 2 Outline What about Cat recognition? Classical face recognition Modern face recognition DeepFace FaceNet Comparison

More information

ImageNet Classification with Deep Convolutional Neural Networks

ImageNet Classification with Deep Convolutional Neural Networks ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky Ilya Sutskever Geoffrey Hinton University of Toronto Canada Paper with same name to appear in NIPS 2012 Main idea Architecture

More information

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides Deep Learning in Visual Recognition Thanks Da Zhang for the slides Deep Learning is Everywhere 2 Roadmap Introduction Convolutional Neural Network Application Image Classification Object Detection Object

More information

Distributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability

Distributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability Distributed Training of Deep Neural Networks: Theoretical and Practical Limits of Parallel Scalability Janis Keuper Itwm.fraunhofer.de/ml Competence Center High Performance Computing Fraunhofer ITWM, Kaiserslautern,

More information

XES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework

XES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework XES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework Demo Paper Joerg Evermann 1, Jana-Rebecca Rehse 2,3, and Peter Fettke 2,3 1 Memorial University of Newfoundland 2 German Research

More information

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:

More information

Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling

Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling Authors: Junyoung Chung, Caglar Gulcehre, KyungHyun Cho and Yoshua Bengio Presenter: Yu-Wei Lin Background: Recurrent Neural

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes CS 6501: Deep Learning for Computer Graphics Training Neural Networks II Connelly Barnes Overview Preprocessing Initialization Vanishing/exploding gradients problem Batch normalization Dropout Additional

More information

SEMANTIC COMPUTING. Lecture 9: Deep Learning: Recurrent Neural Networks (RNNs) TU Dresden, 21 December 2018

SEMANTIC COMPUTING. Lecture 9: Deep Learning: Recurrent Neural Networks (RNNs) TU Dresden, 21 December 2018 SEMANTIC COMPUTING Lecture 9: Deep Learning: Recurrent Neural Networks (RNNs) Dagmar Gromann International Center For Computational Logic TU Dresden, 21 December 2018 Overview Handling Overfitting Recurrent

More information

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017 COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization

More information

Lecture 10 CNNs on Graphs

Lecture 10 CNNs on Graphs Lecture 10 CNNs on Graphs CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago April 26, 2017 Two Scenarios For CNNs on graphs, we have two distinct scenarios: Scenario 1: Each

More information

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: Code Mania 2019 Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: 1. Introduction to Artificial Intelligence 2. Introduction to python programming and Environment

More information

CS489/698: Intro to ML

CS489/698: Intro to ML CS489/698: Intro to ML Lecture 14: Training of Deep NNs Instructor: Sun Sun 1 Outline Activation functions Regularization Gradient-based optimization 2 Examples of activation functions 3 5/28/18 Sun Sun

More information

DeepFace: Closing the Gap to Human-Level Performance in Face Verification

DeepFace: Closing the Gap to Human-Level Performance in Face Verification DeepFace: Closing the Gap to Human-Level Performance in Face Verification Report on the paper Artem Komarichev February 7, 2016 Outline New alignment technique New DNN architecture New large dataset with

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence AI & Machine Learning & Neural Nets Marc Toussaint University of Stuttgart Winter 2018/19 Motivation: Neural networks became a central topic for Machine Learning and AI. But in

More information

Data Mining and Analytics

Data Mining and Analytics Data Mining and Analytics Aik Choon Tan, Ph.D. Associate Professor of Bioinformatics Division of Medical Oncology Department of Medicine aikchoon.tan@ucdenver.edu 9/22/2017 http://tanlab.ucdenver.edu/labhomepage/teaching/bsbt6111/

More information

EECS 496 Statistical Language Models. Winter 2018

EECS 496 Statistical Language Models. Winter 2018 EECS 496 Statistical Language Models Winter 2018 Introductions Professor: Doug Downey Course web site: www.cs.northwestern.edu/~ddowney/courses/496_winter2018 (linked off prof. home page) Logistics Grading

More information

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal Object Detection Lecture 10.3 - Introduction to deep learning (CNN) Idar Dyrdal Deep Learning Labels Computational models composed of multiple processing layers (non-linear transformations) Used to learn

More information

Introduction to Deep Learning

Introduction to Deep Learning ENEE698A : Machine Learning Seminar Introduction to Deep Learning Raviteja Vemulapalli Image credit: [LeCun 1998] Resources Unsupervised feature learning and deep learning (UFLDL) tutorial (http://ufldl.stanford.edu/wiki/index.php/ufldl_tutorial)

More information

Slide credit from Hung-Yi Lee & Richard Socher

Slide credit from Hung-Yi Lee & Richard Socher Slide credit from Hung-Yi Lee & Richard Socher 1 Review Word Vector 2 Word2Vec Variants Skip-gram: predicting surrounding words given the target word (Mikolov+, 2013) CBOW (continuous bag-of-words): predicting

More information

FastText. Jon Koss, Abhishek Jindal

FastText. Jon Koss, Abhishek Jindal FastText Jon Koss, Abhishek Jindal FastText FastText is on par with state-of-the-art deep learning classifiers in terms of accuracy But it is way faster: FastText can train on more than one billion words

More information

Bayesian model ensembling using meta-trained recurrent neural networks

Bayesian model ensembling using meta-trained recurrent neural networks Bayesian model ensembling using meta-trained recurrent neural networks Luca Ambrogioni l.ambrogioni@donders.ru.nl Umut Güçlü u.guclu@donders.ru.nl Yağmur Güçlütürk y.gucluturk@donders.ru.nl Julia Berezutskaya

More information

Day 3 Lecture 1. Unsupervised Learning

Day 3 Lecture 1. Unsupervised Learning Day 3 Lecture 1 Unsupervised Learning Semi-supervised and transfer learning Myth: you can t do deep learning unless you have a million labelled examples for your problem. Reality You can learn useful representations

More information

Encoding RNNs, 48 End of sentence (EOS) token, 207 Exploding gradient, 131 Exponential function, 42 Exponential Linear Unit (ELU), 44

Encoding RNNs, 48 End of sentence (EOS) token, 207 Exploding gradient, 131 Exponential function, 42 Exponential Linear Unit (ELU), 44 A Activation potential, 40 Annotated corpus add padding, 162 check versions, 158 create checkpoints, 164, 166 create input, 160 create train and validation datasets, 163 dropout, 163 DRUG-AE.rel file,

More information

Deep Learning for Computer Vision II

Deep Learning for Computer Vision II IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L

More information

DEEP LEARNING IN PYTHON. The need for optimization

DEEP LEARNING IN PYTHON. The need for optimization DEEP LEARNING IN PYTHON The need for optimization A baseline neural network Input 2 Hidden Layer 5 2 Output - 9-3 Actual Value of Target: 3 Error: Actual - Predicted = 4 A baseline neural network Input

More information

Hierarchical Video Frame Sequence Representation with Deep Convolutional Graph Network

Hierarchical Video Frame Sequence Representation with Deep Convolutional Graph Network Hierarchical Video Frame Sequence Representation with Deep Convolutional Graph Network Feng Mao [0000 0001 6171 3168], Xiang Wu [0000 0003 2698 2156], Hui Xue, and Rong Zhang Alibaba Group, Hangzhou, China

More information

Deep Neural Networks Applications in Handwriting Recognition

Deep Neural Networks Applications in Handwriting Recognition Deep Neural Networks Applications in Handwriting Recognition 2 Who am I? Théodore Bluche PhD defended at Université Paris-Sud last year Deep Neural Networks for Large Vocabulary Handwritten

More information

Inductive Representation Learning on Large Graphs

Inductive Representation Learning on Large Graphs Inductive Representation Learning on Large Graphs William L. Hamilton wleif@stanford.edu Rex Ying rexying@stanford.edu Jure Leskovec jure@cs.stanford.edu Department of Computer Science Stanford University

More information

Deep Learning and Its Applications

Deep Learning and Its Applications Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent

More information

Deep (1) Matthieu Cord LIP6 / UPMC Paris 6

Deep (1) Matthieu Cord LIP6 / UPMC Paris 6 Deep (1) Matthieu Cord LIP6 / UPMC Paris 6 Syllabus 1. Whole traditional (old) visual recognition pipeline 2. Introduction to Neural Nets 3. Deep Nets for image classification To do : Voir la leçon inaugurale

More information

Empirical Evaluation of RNN Architectures on Sentence Classification Task

Empirical Evaluation of RNN Architectures on Sentence Classification Task Empirical Evaluation of RNN Architectures on Sentence Classification Task Lei Shen, Junlin Zhang Chanjet Information Technology lorashen@126.com, zhangjlh@chanjet.com Abstract. Recurrent Neural Networks

More information

The Mathematics Behind Neural Networks

The Mathematics Behind Neural Networks The Mathematics Behind Neural Networks Pattern Recognition and Machine Learning by Christopher M. Bishop Student: Shivam Agrawal Mentor: Nathaniel Monson Courtesy of xkcd.com The Black Box Training the

More information

Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15

Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15 Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 15 Slides adapted from Jordan Boyd-Graber Machine Learning: Chenhao Tan Boulder 1 of 21 Logistics HW3 available on Github, due on October

More information

Layerwise Interweaving Convolutional LSTM

Layerwise Interweaving Convolutional LSTM Layerwise Interweaving Convolutional LSTM Tiehang Duan and Sargur N. Srihari Department of Computer Science and Engineering The State University of New York at Buffalo Buffalo, NY 14260, United States

More information

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction Akarsh Pokkunuru EECS Department 03-16-2017 Contractive Auto-Encoders: Explicit Invariance During Feature Extraction 1 AGENDA Introduction to Auto-encoders Types of Auto-encoders Analysis of different

More information

Advanced Video Analysis & Imaging

Advanced Video Analysis & Imaging Advanced Video Analysis & Imaging (5LSH0), Module 09B Machine Learning with Convolutional Neural Networks (CNNs) - Workout Farhad G. Zanjani, Clint Sebastian, Egor Bondarev, Peter H.N. de With ( p.h.n.de.with@tue.nl

More information

Natural Language Processing with Deep Learning CS224N/Ling284. Christopher Manning Lecture 4: Backpropagation and computation graphs

Natural Language Processing with Deep Learning CS224N/Ling284. Christopher Manning Lecture 4: Backpropagation and computation graphs Natural Language Processing with Deep Learning CS4N/Ling84 Christopher Manning Lecture 4: Backpropagation and computation graphs Lecture Plan Lecture 4: Backpropagation and computation graphs 1. Matrix

More information

arxiv: v1 [cond-mat.dis-nn] 30 Dec 2018

arxiv: v1 [cond-mat.dis-nn] 30 Dec 2018 A General Deep Learning Framework for Structure and Dynamics Reconstruction from Time Series Data arxiv:1812.11482v1 [cond-mat.dis-nn] 30 Dec 2018 Zhang Zhang, Jing Liu, Shuo Wang, Ruyue Xin, Jiang Zhang

More information

Report: Privacy-Preserving Classification on Deep Neural Network

Report: Privacy-Preserving Classification on Deep Neural Network Report: Privacy-Preserving Classification on Deep Neural Network Janno Veeorg Supervised by Helger Lipmaa and Raul Vicente Zafra May 25, 2017 1 Introduction In this report we consider following task: how

More information

LARGE MARGIN CLASSIFIERS

LARGE MARGIN CLASSIFIERS Admin Assignment 5 LARGE MARGIN CLASSIFIERS David Kauchak CS 451 Fall 2013 Midterm Download from course web page when you re ready to take it 2 hours to complete Must hand-in (or e-mail in) by 11:59pm

More information

Facial Expression Classification with Random Filters Feature Extraction

Facial Expression Classification with Random Filters Feature Extraction Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle

More information

ECG782: Multidimensional Digital Signal Processing

ECG782: Multidimensional Digital Signal Processing ECG782: Multidimensional Digital Signal Processing Object Recognition http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Knowledge Representation Statistical Pattern Recognition Neural Networks Boosting

More information

Restricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version

Restricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version Shallow vs. deep networks Restricted Boltzmann Machines Shallow: one hidden layer Features can be learned more-or-less independently Arbitrary function approximator (with enough hidden units) Deep: two

More information

Lecture 7: Neural network acoustic models in speech recognition

Lecture 7: Neural network acoustic models in speech recognition CS 224S / LINGUIST 285 Spoken Language Processing Andrew Maas Stanford University Spring 2017 Lecture 7: Neural network acoustic models in speech recognition Outline Hybrid acoustic modeling overview Basic

More information

Know your neighbours: Machine Learning on Graphs

Know your neighbours: Machine Learning on Graphs Know your neighbours: Machine Learning on Graphs Andrew Docherty Senior Research Engineer andrew.docherty@data61.csiro.au www.data61.csiro.au 2 Graphs are Everywhere Online Social Networks Transportation

More information

CSE 559A: Computer Vision

CSE 559A: Computer Vision CSE 559A: Computer Vision Fall 2018: T-R: 11:30-1pm @ Lopata 101 Instructor: Ayan Chakrabarti (ayan@wustl.edu). Course Staff: Zhihao Xia, Charlie Wu, Han Liu http://www.cse.wustl.edu/~ayan/courses/cse559a/

More information

Artificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu

Artificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu Artificial Neural Networks Introduction to Computational Neuroscience Ardi Tampuu 7.0.206 Artificial neural network NB! Inspired by biology, not based on biology! Applications Automatic speech recognition

More information

Deep Learning with Tensorflow AlexNet

Deep Learning with Tensorflow   AlexNet Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification

More information

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning, A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,

More information