Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting
|
|
- Arnold Hensley
- 5 years ago
- Views:
Transcription
1 Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting Yaguang Li Joint work with Rose Yu, Cyrus Shahabi, Yan Liu Page 1
2 Introduction Traffic congesting is wasteful of time, money and energy Traffic congestion costs Americans $124 billion + direct/indirect loss in Accurate traffic forecasting could substantially improve route planning and mitigate traffic congestion. + * Image from LaLa Land: Page 2
3 The Problem Existing Solutions Introduction Route 1: Best route according to current traffic conditions Route 2: Let s see. Route 1 Route 2 Destination 7:00AM Page 3
4 The Problem Existing Solutions Introduction Predictive vs. Real-Time Path-Planning Route 1 Route 2 Destination Evolution of traffic over time 7:15AM Page 4
5 The Problem Existing Solutions Introduction Traffic forecasting enables better route planning. Route 1 Route 2 Destination Stuck in traffic 7:30AM Page 5
6 Traffic Prediction Input: road network and past T traffic speed observed at sensors Output: traffic speed for the next T steps Input: Observations Output: Predictions... 7:00 AM... 8:00 AM 8:10AM, 8:20AM,, 9:00 AM Page 6
7 Challenges for Traffic Prediction Speed (mile/h) Complex Spatial Dependency Non-linear, non-stationary Temporal Dynamics Sensor 1 Sensor 2 Sensor 3 Page 7
8 Related Work Traffic Prediction without spatial dependency modeling Simulation and queuing theory [Drew 1968] Kalman Filter: [Okutani et al. TRB 83] [Wang et al. TRB 05] ARIMA: [Williams et al. TRB 98] [Pan et al. ICDM 12] Support Vector Regression (SVR): [Muller et al, ICANN' 97] [Wu et al. ITS 04] Gaussian process [Xie et al. TRB 10] [Zhou et al. SIGMOD 15] Recurrent neural networks and deep learning: [Lv et al ITS 15] [Ma et al. TRC 15] [Li et al SDM 17] Model each sensor independently Fail to capture spatial correlation Page 8
9 Related Work Traffic Prediction with spatial dependency modeling Vector ARIMA [Williams and Hoel JTE 03], [Chandra et al. ITS 09] Spatiotemporal ARIMA [Kamarianakis et al., TRB 03] [Min and Wynter, TRC 11] k-nearest Neighbor [Li et al. ITS 12] [Rice et al. ITS 13] Latent Space Model [Deng et al.kdd 17] Convolutional Neural Network [Ma et al. ITS 17] Either assume linear temporal dependency or fail to capture the non-euclidean spatial dependency Page 9
10 Big picture Model spatial dependency with proposed diffusion convolution. Model temporal dependency with augmented recurrent neural network * Li, Yaguang et al. Diffusion Convolutional Recurrent Neural Network: Data-driven Traffic Forecasting, ICLR Page 10
11 Spatial Dependency Modeling Model spatial dependency with Convolutional Neural Networks (CNN) CNN extracts meaningful spatial patterns using filters. State-of-the-art results on image related tasks Image CNN is only applicable to Euclidean grid graph. Convolutional Filter * Y LeCun et al. Gradient-based learning applied to document recognition. Proc. IEEE Image from: Page 11
12 Spatial Dependency in Traffic Prediction Spatial dependency among traffic flow is non-euclidean and directed Sensor 1 Sensor 2 Close in Euclidean space Similar traffic speed Sensor 3 dist net v i v j dist net v i v j Page 12
13 Spatial Dependency Modeling Model the network of traffic sensors, i.e., loop detectors, as a directed graph Graph G = (V, A) Vertices V: o sensors Adjacency matrix A: weight between vertices A ij = exp dist net v i, v j 2 σ 2 if dist net v i, v j κ dist net v i, v j : road network distance from v i to v j, κ: threshold to ensure sparsity, σ 2 variance of all pairwise road network distances Page 13
14 Problem Statement Graph signal: X t R V P, observation on G at time t V : number of vertices P : feature dimension of each vertex. Problem Statement: Learn a function g( ) to map T historical graph signals to future T graph signals X t T +1 X t X t+1 X t+t g. Page 14
15 Generalize Convolution to Graph Convolution as a weighted combination of neighborhood vertices. l w 6,4 h 7 l Filter weight Convolutional filter on graph centered at v 6 Learning complexity is too high: O V N h 2 l h 8 l Min h 6 l h 1 l h 4 l h 9 l h 5 l h 3 l l h 10 Max V h i l+1 = j N i w l l ij h j N i : neighbor of vertex i w l ij : filter weight of vertex j centered at vertex i layer l h l j : feature of vertex j in layer l h 1 j = X j,:, i.e., the input. Page 15
16 Generalize Convolution to Graph Diffusion convolution filter: combination of diffusion processes with different steps on the graph. Transition matrices of the diffusion process K 1 X :,p G f θ = θ k D O 1 A k X :,p Learning complexity: O K k=0 = θ 0 + θ 1 + θ θ K Example diffusion filter Centered at 0 Step Diffusion G : diffusion convolution, D o : diagonal out-degree matrix. 1 Step Diffusion 2 Step Diffusion Filter weight Min K Step Diffusion Max Page 16
17 Generalize Convolution to Graph Diffusion convolution filter: combination of diffusion processes with different steps on the graph. Dual directional diffusion to model upstream and downstream separately K 1 X :,p G f θ = k=0 θ k,1 D O 1 A k + θ k,2 D I 1 A k X :,p = θ 0 + θ 1 + θ θ K Example diffusion filter Centered at 0 Step Diffusion 1 Step Diffusion G : diffusion convolution, D o : diagonal out-degree matrix, D I : diagonal in-degree matrix 2 Step Diffusion Weight Min K Step Diffusion Max Page 17
18 Advantage of Diffusion Convolution Efficient K 1 X :,p G f θ = θ k,1 D 1 O A k + θ k,2 D 1 I A k k=0 Learning complexity: O K Time complexity: O K E, E number of edges X :,p Expressive Many popular convolution operations, including the ChebNet [Defferrard et al., NIPS 16], can be represented using the diffusion convolution [Li et al. ICLR 18]. G : diffusion convolution, D o : diagonal out-degree matrix, D I : diagonal in-degree matrix + Defferrard, M., Bresson, X., Vandergheynst, P., Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering, NIPS, 2016 * Li, Yaguang et al. Diffusion Convolutional Recurrent Neural Network: Data-driven Traffic Forecasting, ICLR, 2018 Page 18
19 Big picture Model spatial dependency with proposed diffusion convolution Model temporal dependency with augmented recurrent neural network * Li, Yaguang et al. Diffusion Convolutional Recurrent Neural Network: Data-driven Traffic Forecasting, ICLR Page 19
20 Model Temporal Dynamics using Recurrent Neural Networks Recurrent Neural Networks (RNN) Non-linear, non-stationary auto-regression State-of-the-art performance in sequence modeling h h 1 h 2 h 3 RNN Unroll RNN RNN RNN Popular example of RNN Long Short-Term Memory unit (LSTM) Gated Recurrent Unit (GRU) x GRU Diffusion + Convolution x 1 x 2 x 3 Page 20
21 Model Temporal Dynamics using Recurrent Neural Network Multi-step ahead prediction with RNN Current Time Error Propagation Teach the model to deal with its own error. x 4 x 5 x 6 x 1 x 2 x 3 x 4 x 5 x x Model prediction Observation or ground truth Previous model output is fed into the network Page 21
22 Improve Multi-step ahead Forecasting Traffic prediction as a sequence to sequence learning problem Encoder-decoder framework x 1, x 2, x 3 x 4 x 1, x 2, x 3 x 4, x 5, x 6 Current Time Backprop errors from multiple steps. x 4 x 5 x 6 δ 4 δ 5 δ 6 x 4 x 5 x 6 x 1 x 2 x 3 <GO> x 4 x 5 x x Model prediction Observation or ground truth Encoder * Sutskever et al. Sequence to sequence learning with neural networks, NIPS 2014 Decoder Ground truth becomes unavailable in testing. Page 22
23 Improve Multi-step ahead Forecasting Improve multi-step ahead forecasting with scheduled sampling Current Time x 4 x 5 x 6 x x x 1 Model prediction Observation or ground truth x 2 Encoder x 3 * Bengio,Samy, et al. Scheduled sampling for sequence prediction with recurrent neural networks. NIPS 2015 <GO> x 4 x 4 x 5 x 5 Decoder Scheduled sampling: Choose to use the previous ground truth or model prediction by flipping a coin Page 23
24 Improve Multi-step ahead Forecasting probability Improve multi-step ahead forecasting with scheduled sampling Curriculum learning: gradually enables the model to deal with its own error. Easy: Only feed ground truth x t+1 x t x t x x Model prediction Observation or ground truth # iteration * Bengio,Samy, et al. Scheduled sampling for sequence prediction with recurrent neural networks. NIPS 2015 Hard: Only feed model prediction Page 24
25 Diffusion Convolutional Recurrent Neural Network Diffusion Convolutional Recurrent Neural Network (DCRNN) Model spatial dependency with diffusion convolution Sequence to sequence learning with encoder-decoder framework Improve multi-step ahead forecasting with scheduled sampling * Li, Yaguang et al. Diffusion Convolutional Recurrent Neural Network: Data-driven Traffic Forecasting, ICLR 2018 Page 25
26 Experiments Datasets METR-LA: 207 traffic sensors in Los Angeles 4 months in M observations PEMS-BAY: 345 traffic sensors in Bay Area 6 months in M observations Page 26
27 Experiments Baselines Historical Average (HA) Autoregressive Integrated Moving Average (ARIMA) Support Vector Regression (SVR) Vector Auto-Regression (VAR) Feed forward Neural network (FNN) Fully connected LSTM with Sequence to Sequence framework (FC-LSTM) Task Multi-step ahead traffic speed forecasting Page 27
28 Experimental Results Mean Absolute Error (MAE) Mean Absolute Error (MAE) DCRNN achieves the best performance for all forecasting horizons for both datasets 7.00 METR-LA 3.50 PEMS-BAY Min 30 Min 1 Hour Min 30 Min 1 Hour HA ARIMA VAR SVR FNN FC-LSTM DCRNN HA ARIMA VAR SVR FNN FC-LSTM DCRNN Page 28
29 Effects of Spatiotemporal Dependency Modeling Mean Absolute Error (MAE) w/o temporal: removing sequence to sequence learning. w/o spatial: remove the diffusion convolution Removing either spatial or temporal modeling results in significantly worse results Min 30 Min 1 Hour DCRNN w/o Temporal DCRNN w/o Spatial DCRNN METR-LA Page 29
30 Example: Prediction Results mile/h DCRNN is more likely to accurately predict abrupt changes in the traffic speed than the best baseline method. Example Prediction Results on METR-LA Page 30
31 Example: Filter Visualization Visualization of learned filters Filters are localized around the center. Weights diffuse alongside the road network. METR-LA Max 0 Min center Learned filters centered at different verticices. Page 31
32 Summary Propose diffusion convolution to model the spatial dependency of traffic flow. Propose Diffusion Convolutional Recurrent Neural Network (DCRNN) that captures both spatial and temporal dependencies. DCRNN obtains consistent improvement over state-of-the-art baseline methods. ICLR * Li, Yaguang et al. Diffusion Convolutional Recurrent Neural Network: Data-driven Traffic Forecasting, ICLR Page 32
33 Thank You! Q & A Page 33
Spatiotemporal Multi-Graph Convolution Network for Ride-hailing Demand Forecasting
Spatiotemporal Multi-Graph Convolution Network for Ride-hailing Demand Forecasting Xu Geng 1, Yaguang Li 2, Leye Wang 1,3, Lingyu Zhang 4, Qiang Yang 1, Jieping Ye 4, Yan Liu 2,4 1 Hong Kong University
More informationLatent Space Model for Road Networks to Predict Time-Varying Traffic. Presented by: Rob Fitzgerald Spring 2017
Latent Space Model for Road Networks to Predict Time-Varying Traffic Presented by: Rob Fitzgerald Spring 2017 Definition of Latent https://en.oxforddictionaries.com/definition/latent Latent Space Model?
More informationDeep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016
More informationAsynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features
Asynchronous Parallel Learning for Neural Networks and Structured Models with Dense Features Xu SUN ( 孙栩 ) Peking University xusun@pku.edu.cn Motivation Neural networks -> Good Performance CNN, RNN, LSTM
More informationLSTM and its variants for visual recognition. Xiaodan Liang Sun Yat-sen University
LSTM and its variants for visual recognition Xiaodan Liang xdliang328@gmail.com Sun Yat-sen University Outline Context Modelling with CNN LSTM and its Variants LSTM Architecture Variants Application in
More informationImage Question Answering using Convolutional Neural Network with Dynamic Parameter Prediction
Image Question Answering using Convolutional Neural Network with Dynamic Parameter Prediction by Noh, Hyeonwoo, Paul Hongsuck Seo, and Bohyung Han.[1] Presented : Badri Patro 1 1 Computer Vision Reading
More informationDeep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 6 April 2017 joint work with Max Welling (University of Amsterdam) The success story of deep
More informationMachine Learning 13. week
Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of
More informationRecurrent Neural Network (RNN) Industrial AI Lab.
Recurrent Neural Network (RNN) Industrial AI Lab. For example (Deterministic) Time Series Data Closed- form Linear difference equation (LDE) and initial condition High order LDEs 2 (Stochastic) Time Series
More information27: Hybrid Graphical Models and Neural Networks
10-708: Probabilistic Graphical Models 10-708 Spring 2016 27: Hybrid Graphical Models and Neural Networks Lecturer: Matt Gormley Scribes: Jakob Bauer Otilia Stretcu Rohan Varma 1 Motivation We first look
More informationLearning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li
Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,
More informationEnd-To-End Spam Classification With Neural Networks
End-To-End Spam Classification With Neural Networks Christopher Lennan, Bastian Naber, Jan Reher, Leon Weber 1 Introduction A few years ago, the majority of the internet s network traffic was due to spam
More informationSequence Modeling: Recurrent and Recursive Nets. By Pyry Takala 14 Oct 2015
Sequence Modeling: Recurrent and Recursive Nets By Pyry Takala 14 Oct 2015 Agenda Why Recurrent neural networks? Anatomy and basic training of an RNN (10.2, 10.2.1) Properties of RNNs (10.2.2, 8.2.6) Using
More informationBidirectional Recurrent Convolutional Networks for Video Super-Resolution
Bidirectional Recurrent Convolutional Networks for Video Super-Resolution Qi Zhang & Yan Huang Center for Research on Intelligent Perception and Computing (CRIPAC) National Laboratory of Pattern Recognition
More informationData Quality Network for Spatiotemporal Forecasting
Data Quality Network for Spatiotemporal Forecasting Sungyong Seo, Yan Liu Department of Computer Science University of Southern California {sungyons, yanliu.cs}@usc.edu Arash Mohegh, George Ban-Weiss Dept
More informationRestricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version
Shallow vs. deep networks Restricted Boltzmann Machines Shallow: one hidden layer Features can be learned more-or-less independently Arbitrary function approximator (with enough hidden units) Deep: two
More informationConvolutional Sequence to Sequence Learning. Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research
Convolutional Sequence to Sequence Learning Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research Sequence generation Need to model a conditional distribution
More informationLatent Space Model for Road Networks to Predict Time-Varying Traffic
Latent Space Model for Road Networks to Predict Time-Varying Traffic Dingxiong Deng 1, Cyrus Shahabi 1, Ugur Demiryurek 1, Linhong Zhu 2,RoseYu 1, Yan Liu 1 Department of Computer Science, University of
More informationDisguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601
Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network Nathan Sun CIS601 Introduction Face ID is complicated by alterations to an individual s appearance Beard,
More informationJOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation
JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS Puyang Xu, Ruhi Sarikaya Microsoft Corporation ABSTRACT We describe a joint model for intent detection and slot filling based
More informationAlternatives to Direct Supervision
CreativeAI: Deep Learning for Graphics Alternatives to Direct Supervision Niloy Mitra Iasonas Kokkinos Paul Guerrero Nils Thuerey Tobias Ritschel UCL UCL UCL TUM UCL Timetable Theory and Basics State of
More informationImage Captioning with Object Detection and Localization
Image Captioning with Object Detection and Localization Zhongliang Yang, Yu-Jin Zhang, Sadaqat ur Rehman, Yongfeng Huang, Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
More informationLayerwise Interweaving Convolutional LSTM
Layerwise Interweaving Convolutional LSTM Tiehang Duan and Sargur N. Srihari Department of Computer Science and Engineering The State University of New York at Buffalo Buffalo, NY 14260, United States
More informationDeep Learning Applications
October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning
More informationThis Talk. 1) Node embeddings. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured
Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 1 This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures
More informationDeep Learning in Visual Recognition. Thanks Da Zhang for the slides
Deep Learning in Visual Recognition Thanks Da Zhang for the slides Deep Learning is Everywhere 2 Roadmap Introduction Convolutional Neural Network Application Image Classification Object Detection Object
More informationLecture 10 CNNs on Graphs
Lecture 10 CNNs on Graphs CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago April 26, 2017 Two Scenarios For CNNs on graphs, we have two distinct scenarios: Scenario 1: Each
More informationStructured Attention Networks
Structured Attention Networks Yoon Kim Carl Denton Luong Hoang Alexander M. Rush HarvardNLP ICLR, 2017 Presenter: Chao Jiang ICLR, 2017 Presenter: Chao Jiang 1 / Outline 1 Deep Neutral Networks for Text
More informationarxiv: v1 [cs.lg] 13 Mar 2019
ST-UNet: A Spatio-Temporal U-Network for Graph-structured Time Series Modeling arxiv:1903.05631v1 [cs.lg] 13 Mar 2019 Bing Yu byu@pku.edu.cn School of Mathematical Sciences, Peking University Beijing,
More informationA. Further experimental analysis. B. Simulation data
A. Further experimental analysis A.1. Kuramoto LSTM vs. NRI comparison From the results in our main paper it became evident that a simple LSTM model excels at predicting the dynamics of a network of phase-coupled
More informationSpatial Localization and Detection. Lecture 8-1
Lecture 8: Spatial Localization and Detection Lecture 8-1 Administrative - Project Proposals were due on Saturday Homework 2 due Friday 2/5 Homework 1 grades out this week Midterm will be in-class on Wednesday
More informationEmpirical Evaluation of RNN Architectures on Sentence Classification Task
Empirical Evaluation of RNN Architectures on Sentence Classification Task Lei Shen, Junlin Zhang Chanjet Information Technology lorashen@126.com, zhangjlh@chanjet.com Abstract. Recurrent Neural Networks
More informationChannel Locality Block: A Variant of Squeeze-and-Excitation
Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan
More informationPart Localization by Exploiting Deep Convolutional Networks
Part Localization by Exploiting Deep Convolutional Networks Marcel Simon, Erik Rodner, and Joachim Denzler Computer Vision Group, Friedrich Schiller University of Jena, Germany www.inf-cv.uni-jena.de Abstract.
More informationContent-Based Image Recovery
Content-Based Image Recovery Hong-Yu Zhou and Jianxin Wu National Key Laboratory for Novel Software Technology Nanjing University, China zhouhy@lamda.nju.edu.cn wujx2001@nju.edu.cn Abstract. We propose
More informationCrowd Scene Understanding with Coherent Recurrent Neural Networks
Crowd Scene Understanding with Coherent Recurrent Neural Networks Hang Su, Yinpeng Dong, Jun Zhu May 22, 2016 Hang Su, Yinpeng Dong, Jun Zhu IJCAI 2016 May 22, 2016 1 / 26 Outline 1 Introduction 2 LSTM
More informationRecurrent Neural Networks. Nand Kishore, Audrey Huang, Rohan Batra
Recurrent Neural Networks Nand Kishore, Audrey Huang, Rohan Batra Roadmap Issues Motivation 1 Application 1: Sequence Level Training 2 Basic Structure 3 4 Variations 5 Application 3: Image Classification
More informationDeep Learning for Computer Vision II
IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L
More informationA Hybrid Method for Traffic Flow Forecasting Using Multimodal Deep Learning
A Hybrid Method for Traffic Flow Forecasting Using Multimodal Deep Learning Shengdong Du, Tianrui Li, Xun Gong, Zeng Yu and Shi-Jinn Horng Abstract Traffic flow forecasting has been regarded as a key problem
More informationEarly detection of Crossfire attacks using deep learning
Early detection of Crossfire attacks using deep learning Saurabh Misra, Mengxuan Tan, Mostafa Rezazad, Ngai-Man Cheung Singapore University of Technology and Design Content The Crossfire Attack A brief
More informationOptimization for Machine Learning
with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex
More informationCSC 578 Neural Networks and Deep Learning
CSC 578 Neural Networks and Deep Learning Fall 2018/19 7. Recurrent Neural Networks (Some figures adapted from NNDL book) 1 Recurrent Neural Networks 1. Recurrent Neural Networks (RNNs) 2. RNN Training
More informationarxiv: v1 [cs.cv] 6 Jul 2016
arxiv:607.079v [cs.cv] 6 Jul 206 Deep CORAL: Correlation Alignment for Deep Domain Adaptation Baochen Sun and Kate Saenko University of Massachusetts Lowell, Boston University Abstract. Deep neural networks
More informationRecurrent Convolutional Neural Networks for Scene Labeling
Recurrent Convolutional Neural Networks for Scene Labeling Pedro O. Pinheiro, Ronan Collobert Reviewed by Yizhe Zhang August 14, 2015 Scene labeling task Scene labeling: assign a class label to each pixel
More informationLearning-based Localization
Learning-based Localization Eric Brachmann ECCV 2018 Tutorial on Visual Localization - Feature-based vs. Learned Approaches Torsten Sattler, Eric Brachmann Roadmap Machine Learning Basics [10min] Convolutional
More informationPixel-level Generative Model
Pixel-level Generative Model Generative Image Modeling Using Spatial LSTMs (2015NIPS) L. Theis and M. Bethge University of Tübingen, Germany Pixel Recurrent Neural Networks (2016ICML) A. van den Oord,
More informationGRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION
GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION Nasehe Jamshidpour a, Saeid Homayouni b, Abdolreza Safari a a Dept. of Geomatics Engineering, College of Engineering,
More informationCS 2750: Machine Learning. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh April 13, 2016
CS 2750: Machine Learning Neural Networks Prof. Adriana Kovashka University of Pittsburgh April 13, 2016 Plan for today Neural network definition and examples Training neural networks (backprop) Convolutional
More informationLecture 21 : A Hybrid: Deep Learning and Graphical Models
10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation
More informationGaAN: Gated Attention Networks for Learning on Large and Spatiotemporal Graphs
GaAN: Gated Attention Networks for Learning on Large and Spatiotemporal Graphs Jiani Zhang, Xingjian Shi, Junyuan Xie, Hao Ma, Irwin King, Dit-Yan Yeung The Chinese University of Hong Kong, Hong Kong,
More informationLarge-Scale Face Manifold Learning
Large-Scale Face Manifold Learning Sanjiv Kumar Google Research New York, NY * Joint work with A. Talwalkar, H. Rowley and M. Mohri 1 Face Manifold Learning 50 x 50 pixel faces R 2500 50 x 50 pixel random
More informationarxiv: v1 [cs.lg] 20 Apr 2017
Dynamic Graph Convolutional Networks Franco Manessi 1, Alessandro Rozza 1, and Mario Manzo 2 arxiv:1704.06199v1 [cs.lg] 20 Apr 2017 1 Research Team - Waynaut {name.surname}@waynaut.com 2 Servizi IT - Università
More informationScaling Distributed Machine Learning
Scaling Distributed Machine Learning with System and Algorithm Co-design Mu Li Thesis Defense CSD, CMU Feb 2nd, 2017 nx min w f i (w) Distributed systems i=1 Large scale optimization methods Large-scale
More informationDeep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?
Data Mining Deep Learning Deep Learning provided breakthrough results in speech recognition and image classification. Why? Because Speech recognition and image classification are two basic examples of
More informationEfficient Algorithms may not be those we think
Efficient Algorithms may not be those we think Yann LeCun, Computational and Biological Learning Lab The Courant Institute of Mathematical Sciences New York University http://yann.lecun.com http://www.cs.nyu.edu/~yann
More informationMarkov Random Fields and Gibbs Sampling for Image Denoising
Markov Random Fields and Gibbs Sampling for Image Denoising Chang Yue Electrical Engineering Stanford University changyue@stanfoed.edu Abstract This project applies Gibbs Sampling based on different Markov
More informationLC-RNN: A Deep Learning Model for Traffic Speed Prediction
LC-RNN: A Deep Learning Model for Traffic Speed Prediction Zhongjian Lv 1, Jiajie Xu 1,2,3, Kai Zheng 4, Hongzhi Yin 5, Pengpeng Zhao 1, Xiaofang Zhou 5,1 1 School of Computer Science and Technology, Soochow
More informationarxiv: v1 [cond-mat.dis-nn] 30 Dec 2018
A General Deep Learning Framework for Structure and Dynamics Reconstruction from Time Series Data arxiv:1812.11482v1 [cond-mat.dis-nn] 30 Dec 2018 Zhang Zhang, Jing Liu, Shuo Wang, Ruyue Xin, Jiang Zhang
More informationTutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY
Tutorial on Keras CAP 6412 - ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Deep learning packages TensorFlow Google PyTorch Facebook AI research Keras Francois Chollet (now at Google) Chainer Company
More informationDoes the Brain do Inverse Graphics?
Does the Brain do Inverse Graphics? Geoffrey Hinton, Alex Krizhevsky, Navdeep Jaitly, Tijmen Tieleman & Yichuan Tang Department of Computer Science University of Toronto The representation used by the
More informationPractical Methodology. Lecture slides for Chapter 11 of Deep Learning Ian Goodfellow
Practical Methodology Lecture slides for Chapter 11 of Deep Learning www.deeplearningbook.org Ian Goodfellow 2016-09-26 What drives success in ML? Arcane knowledge of dozens of obscure algorithms? Mountains
More informationHello Edge: Keyword Spotting on Microcontrollers
Hello Edge: Keyword Spotting on Microcontrollers Yundong Zhang, Naveen Suda, Liangzhen Lai and Vikas Chandra ARM Research, Stanford University arxiv.org, 2017 Presented by Mohammad Mofrad University of
More informationShow, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks
Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks Zelun Luo Department of Computer Science Stanford University zelunluo@stanford.edu Te-Lin Wu Department of
More informationReddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011
Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011 1. Introduction Reddit is one of the most popular online social news websites with millions
More informationApplication of Deep Learning Techniques in Satellite Telemetry Analysis.
Application of Deep Learning Techniques in Satellite Telemetry Analysis. Greg Adamski, Member of Technical Staff L3 Technologies Telemetry and RF Products Julian Spencer Jones, Spacecraft Engineer Telenor
More informationMachine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,
Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image
More informationDeepIM: Deep Iterative Matching for 6D Pose Estimation - Supplementary Material
DeepIM: Deep Iterative Matching for 6D Pose Estimation - Supplementary Material Yi Li 1, Gu Wang 1, Xiangyang Ji 1, Yu Xiang 2, and Dieter Fox 2 1 Tsinghua University, BNRist 2 University of Washington
More informationSemantic Segmentation. Zhongang Qi
Semantic Segmentation Zhongang Qi qiz@oregonstate.edu Semantic Segmentation "Two men riding on a bike in front of a building on the road. And there is a car." Idea: recognizing, understanding what's in
More informationSpectral Active Clustering of Remote Sensing Images
Spectral Active Clustering of Remote Sensing Images Zifeng Wang, Gui-Song Xia, Caiming Xiong, Liangpei Zhang To cite this version: Zifeng Wang, Gui-Song Xia, Caiming Xiong, Liangpei Zhang. Spectral Active
More informationLiangjie Hong*, Dawei Yin*, Jian Guo, Brian D. Davison*
Tracking Trends: Incorporating Term Volume into Temporal Topic Models Liangjie Hong*, Dawei Yin*, Jian Guo, Brian D. Davison* Dept. of Computer Science and Engineering, Lehigh University, Bethlehem, PA,
More information16-785: Integrated Intelligence in Robotics: Vision, Language, and Planning. Spring 2018 Lecture 14. Image to Text
16-785: Integrated Intelligence in Robotics: Vision, Language, and Planning Spring 2018 Lecture 14. Image to Text Input Output Classification tasks 4/1/18 CMU 16-785: Integrated Intelligence in Robotics
More informationDeep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks
Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin
More informationAutoencoder. Representation learning (related to dictionary learning) Both the input and the output are x
Deep Learning 4 Autoencoder, Attention (spatial transformer), Multi-modal learning, Neural Turing Machine, Memory Networks, Generative Adversarial Net Jian Li IIIS, Tsinghua Autoencoder Autoencoder Unsupervised
More informationRecommender Systems New Approaches with Netflix Dataset
Recommender Systems New Approaches with Netflix Dataset Robert Bell Yehuda Koren AT&T Labs ICDM 2007 Presented by Matt Rodriguez Outline Overview of Recommender System Approaches which are Content based
More informationA Dynamic Shortest Path Algorithm Using Multi-Step Ahead Link Travel Time Prediction
Journal ofadvanced Transportation, Vol. 39, No. I, pp. 5-18 www, advanced-transport. corn A Dynamic Shortest Path Algorithm Using Multi-Step Ahead Link Travel Time Prediction Young-Ihn Lee Seungiae Lee
More informationPredicting Station-level Hourly Demand in a Large-scale Bike-sharing Network: A Graph Convolutional Neural Network Approach
Predicting Station-level Hourly Demand in a Large-scale Bike-sharing Network: A Graph Convolutional Neural Network Approach Lei Lin a, Zhengbing He b and Srinivas Peeta c* a NEXTRANS Center, Purdue University,
More informationXES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework
XES Tensorflow Process Prediction using the Tensorflow Deep-Learning Framework Demo Paper Joerg Evermann 1, Jana-Rebecca Rehse 2,3, and Peter Fettke 2,3 1 Memorial University of Newfoundland 2 German Research
More informationNeural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /10/2017
3/0/207 Neural Networks Emily Fox University of Washington March 0, 207 Slides adapted from Ali Farhadi (via Carlos Guestrin and Luke Zettlemoyer) Single-layer neural network 3/0/207 Perceptron as a neural
More informationA Dendrogram. Bioinformatics (Lec 17)
A Dendrogram 3/15/05 1 Hierarchical Clustering [Johnson, SC, 1967] Given n points in R d, compute the distance between every pair of points While (not done) Pick closest pair of points s i and s j and
More informationModeling Sequences Conditioned on Context with RNNs
Modeling Sequences Conditioned on Context with RNNs Sargur Srihari srihari@buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 10. Topics in Sequence
More informationECS289: Scalable Machine Learning
ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 22, 2016 Course Information Website: http://www.stat.ucdavis.edu/~chohsieh/teaching/ ECS289G_Fall2016/main.html My office: Mathematical Sciences
More informationUnsupervised Learning
Deep Learning for Graphics Unsupervised Learning Niloy Mitra Iasonas Kokkinos Paul Guerrero Vladimir Kim Kostas Rematas Tobias Ritschel UCL UCL/Facebook UCL Adobe Research U Washington UCL Timetable Niloy
More informationContrast adjustment via Bayesian sequential partitioning
Contrast adjustment via Bayesian sequential partitioning Zhiyu Wang, Shuo Xie, Bai Jiang Abstract Photographs taken in dim light have low color contrast. However, traditional methods for adjusting contrast
More informationRNNs as Directed Graphical Models
RNNs as Directed Graphical Models Sargur Srihari srihari@buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 10. Topics in Sequence Modeling Overview
More informationSentiment Classification of Food Reviews
Sentiment Classification of Food Reviews Hua Feng Department of Electrical Engineering Stanford University Stanford, CA 94305 fengh15@stanford.edu Ruixi Lin Department of Electrical Engineering Stanford
More informationSensor-based Semantic-level Human Activity Recognition using Temporal Classification
Sensor-based Semantic-level Human Activity Recognition using Temporal Classification Weixuan Gao gaow@stanford.edu Chuanwei Ruan chuanwei@stanford.edu Rui Xu ray1993@stanford.edu I. INTRODUCTION Human
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationProceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong
, March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong TABLE I CLASSIFICATION ACCURACY OF DIFFERENT PRE-TRAINED MODELS ON THE TEST DATA
More informationHouse Price Prediction Using LSTM
House Price Prediction Using LSTM Xiaochen Chen Lai Wei The Hong Kong University of Science and Technology Jiaxin Xu ABSTRACT In this paper, we use the house price data ranging from January 2004 to October
More informationPrediction of Pedestrian Trajectories Final Report
Prediction of Pedestrian Trajectories Final Report Mingchen Li (limc), Yiyang Li (yiyang7), Gendong Zhang (zgdsh29) December 15, 2017 1 Introduction As the industry of automotive vehicles growing rapidly,
More informationOverall Description. Goal: to improve spatial invariance to the input data. Translation, Rotation, Scale, Clutter, Elastic
Philippe Giguère Overall Description Goal: to improve spatial invariance to the input data Translation, Rotation, Scale, Clutter, Elastic How: add a learnable module which explicitly manipulate spatially
More informationTHE free electron laser(fel) at Stanford Linear
1 Beam Detection Based on Machine Learning Algorithms Haoyuan Li, hyli16@stanofrd.edu Qing Yin, qingyin@stanford.edu Abstract The positions of free electron laser beams on screens are precisely determined
More informationTemporal HeartNet: Towards Human-Level Automatic Analysis of Fetal Cardiac Screening Video
Temporal HeartNet: Towards Human-Level Automatic Analysis of Fetal Cardiac Screening Video Weilin Huang, Christopher P. Bridge, J. Alison Noble, and Andrew Zisserman Department of Engineering Science,
More informationWhy equivariance is better than premature invariance
1 Why equivariance is better than premature invariance Geoffrey Hinton Canadian Institute for Advanced Research & Department of Computer Science University of Toronto with contributions from Sida Wang
More informationS7348: Deep Learning in Ford's Autonomous Vehicles. Bryan Goodman Argo AI 9 May 2017
S7348: Deep Learning in Ford's Autonomous Vehicles Bryan Goodman Argo AI 9 May 2017 1 Ford s 12 Year History in Autonomous Driving Today: examples from Stereo image processing Object detection Using RNN
More informationDeep Character-Level Click-Through Rate Prediction for Sponsored Search
Deep Character-Level Click-Through Rate Prediction for Sponsored Search Bora Edizel - Phd Student UPF Amin Mantrach - Criteo Research Xiao Bai - Oath This work was done at Yahoo and will be presented as
More informationCapsule Networks. Eric Mintun
Capsule Networks Eric Mintun Motivation An improvement* to regular Convolutional Neural Networks. Two goals: Replace max-pooling operation with something more intuitive. Keep more info about an activated
More informationResearch on Pruning Convolutional Neural Network, Autoencoder and Capsule Network
Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Tianyu Wang Australia National University, Colledge of Engineering and Computer Science u@anu.edu.au Abstract. Some tasks,
More informationConvolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech
Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:
More informationDEEP LEARNING APPROACH FOR REMOTE SENSING IMAGE ANALYSIS
DEEP LEARNING APPROACH FOR REMOTE SENSING IMAGE ANALYSIS Amina Ben Hamida*,** Alexandre Benoit*, Patrick Lambert*, Chokri Ben Amar** * LISTIC, Université Savoie Mont Blanc, France {amina.ben-hamida,alexandre.benoit,
More information