Developing Machine Learning Models. Kevin Tsai

Size: px
Start display at page:

Download "Developing Machine Learning Models. Kevin Tsai"

Transcription

1 Developing Machine Learning Models Kevin Tsai

2

3 GOTO Chicago 2018 Developing Machine Learning Models Kevin Tsai, Google

4 Agenda Machine Learning with tf.estimator Performance pipelines with TFRecords and tf.data Distributed Training with GPUs and TPUs

5 Machine Learning with tf.estimator

6 Machine Learning Pipeline Training Data Transform & Feature Engineering Model Building Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

7 Machine Learning Pipeline Training Data Transform & Feature Engineering split() train() Model evaluate() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

8 Machine Learning Pipeline Training Data Transform & Feature Engineering split() train() Logs Model evaluate() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

9 Machine Learning Pipeline Training Data Transform & Feature Engineering split() train() Logs Model export() evaluate() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

10 Machine Learning Pipeline with tf.estimator Training Data Transform & Feature Engineering model_fn estimator tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

11 Machine Learning Pipeline with tf.estimator train() Training Data Transform & Feature Engineering model_fn evaluate() predict() estimator tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

12 Machine Learning Pipeline with tf.estimator train_input_fn train() Training Data Transform & Feature Engineering eval_input_fn model_fn evaluate() predict_input_fn predict() estimator tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

13 Machine Learning Pipeline with tf.estimator train_input_fn train() Training Data Transform & Feature Engineering eval_input_fn model_fn evaluate() predict_input_fn predict() estimator Logs tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

14 Machine Learning Pipeline with tf.estimator train_input_fn train() Training Data Transform & Feature Engineering eval_input_fn model_fn evaluate() predict_input_fn predict() serving_input_rec_fn estimator Logs tf.estimator export_savedmodel() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions

15 Estimator in TensorFlow Canned estimators Models in a box Estimator Keras model Train and evaluate models Layers Build models Python C++ Java Go... TensorFlow Distributed Execution Engine CPU GPU TPU Android ios...

16 Example: MNIST? 1 MNIST: Yann LeCunn

17 Steps to Run Training on Premade Estimator Training: 1. Import data 2. Create input functions 3. Create estimator 4. Create train and evaluate specs 5. Run train and evaluate Test Inference: 6. Test run predict() 7. Export model

18 Premade and Custom Estimators Premade Estimators DNNClassifier DNNRegressor LinearClassifier LinearRegressor DNNLinearCombinedClassifier DNNLinearCombinedRegressor + tf.contrib.estimator or Custom Estimator

19 Creating Custom Estimator

20 #1 Inference() #1 Inference() (-1,784)

21 #1 Inference() #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool

22 #1 Inference() #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) (-1,3136) (-1,1024) Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool Reshape Dense ReLU Dropout

23 #1 Inference() #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) (-1,3136) (-1,1024) (-1,10) (-1,10) Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool Reshape Dense ReLU Dropout Dense Softmax

24 #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) (-1,3136) (-1,1024) (-1,10) (-1,10) #1 Inference() Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool tf.name_scope log tf.nn.conv2d tf.nn.relu tf.nn.max_pool tf.summary.histogram( tf.summary.histogram tf.summary.histogram Reshape Dense ReLU Dropout Dense Softmax tf.name_scope log relu=false tf.add tf.matmul tf.nn.relu tf.summary.histogram tf.summary.histogram tf.summary.histogram Hands-on TensorBoard: Dandelion Mane

25 #1 Inference() and #2 Calculations and Metrics #### 1 INFERENCE MODEL input_layer = tf.reshape(features["x"], [-1, 28, 28, 1]) conv1 = _conv(input_layer,kernel=[5,5,1,64],name='conv1',log=params['log']) conv2 = _conv(conv1,kernel=[5,5,64,64],name='conv2',log=params['log']) dense = _dense(conv2,size_in=7*7*64,size_out=params['dense_units'], name='dense',relu=true,log=params['log']) if mode==tf.estimator.modekeys.train: dense = tf.nn.dropout(dense,params['drop_out']) logits = _dense(dense,size_in=params['dense_units'], size_out=10,name='output',relu=false,log=params['log']) #1 Inference()

26 #1 Inference() and #2 Calculations and Metrics #### 2 CALCULATIONS AND METRICS predictions = {"classes": tf.argmax(input=logits,axis=1), "logits": logits, "probabilities": tf.nn.softmax(logits,name='softmax')} export_outputs = {'predictions': tf.estimator.export.predictoutput(predictions)} if (mode==tf.estimator.modekeys.train or mode==tf.estimator.modekeys.eval): loss = tf.losses.sparse_softmax_cross_entropy(labels=labels,logits=logits) accuracy = tf.metrics.accuracy( labels=labels, predictions=tf.argmax(logits,axis=1)) metrics = {'accuracy':accuracy} tf.summary.scalar('accuracy',accuracy[1]) #2 Calculations

27 #3 Predict, #4 Train, and #5 Eval #### 3 MODE = PREDICT if mode == tf.estimator.modekeys.predict: return tf.estimator.estimatorspec( mode=mode, predictions=predictions, export_outputs=export_outputs) #3 MODE = PREDICT

28 #3 Predict, #4 Train, and #5 Eval #### 4 MODE = TRAIN if mode == tf.estimator.modekeys.train: learning_rate = tf.train.exponential_decay( params['learning_rate'],tf.train.get_global_step(), decay_steps=100000,decay_rate=0.96) optimizer = tf.train.gradientdescentoptimizer(learning_rate=learning_rate) if params['replicate']==true: optimizer = tf.contrib.estimator.toweroptimizer(optimizer) train_op = optimizer.minimize(loss=loss,global_step=tf.train.get_global_step()) tf.summary.scalar('learning_rate', learning_rate) return tf.estimator.estimatorspec( mode=mode, loss=loss, train_op=train_op) #4 MODE = TRAIN

29 #3 Predict, #4 Train, and #5 Eval #5 MODE = EVAL #### 5 MODE = EVAL if mode == tf.estimator.modekeys.eval: return tf.estimator.estimatorspec( mode=mode,loss=loss,eval_metric_ops=metrics)

30 TFRecords and tf.data

31 (Inefficient) Pipeline # Load Data mnist = tf.contrib.learn.datasets.load_dataset("mnist") train_data = mnist.train.images train_labels = np.asarray(mnist.train.labels, dtype=np.int32) eval_data = mnist.test.images eval_labels = np.asarray(mnist.test.labels, dtype=np.int32) # Create Input Functions train_input_fn = tf.estimator.inputs.numpy_input_fn( x={"x": train_data},y=train_labels,batch_size=100,num_epochs=none,shuffle=true) eval_input_fn = tf.estimator.inputs.numpy_input_fn( x={"x": eval_data},y=eval_labels,num_epochs=1,shuffle=false) pred_input_fn = tf.estimator.inputs.numpy_input_fn( x={"x": eval_data},num_epochs=1,shuffle=false)

32 TFRecord tf.train.example tf.train.example tf.train.example tf.train.example tf.train.example... train004.tfrecords

33 TFRecord tf.train.example tf.train.example tf.train.example tf.train.example tf.train.example train004.tfrecords tf.python_io.tfrecordwriter tf.train.example tf.train.feature tf.train.features tf.train.feature tf.train.feature tf.train.feature tf.train.feature... tf.train.example tf.train.features...

34 tf.data tf.train.example tf.train.features tf.train.feature tf.train.feature tf.train.feature... tf.fixedlenfeature tf.fixedlenfeature tf.fixedlenfeature tf.parse_single_example

35 tf.data input_fn() dataset = tf.data.tfrecorddataset(c['files'],num_parallel_reads=c['threads'])

36 tf.data input_fn() dataset = dataset.map(parse_tfrecord,num_parallel_calls=c['threads']) dataset = dataset.map(lambda x,y: (image_scaling(x),y),num_parallel_calls=c['threads'])

37 tf.data input_fn() if c['mode']==tf.estimator.modekeys.train: dataset = dataset.map(lambda x,y: (distort(x),y),num_parallel_calls=c['threads']) dataset = dataset.shuffle(buffer_size=c['shuffle_buff'])

38 tf.data input_fn() dataset = dataset.repeat() dataset = dataset.batch(c['batch']) dataset = dataset.prefetch(2*c['batch'])

39 tf.data input_fn() iterator = dataset.make_one_shot_iterator() images, labels = iterator.get_next() features = {'images': images} return features, labels

40 tf.data input_fn() train_spec = tf.estimator.trainspec(input_fn=lambda: dataset_input_fn(train_params)) eval_spec = tf.estimator.evalspec(input_fn=lambda: dataset_input_fn(eval_params))

41 Distributed Training

42 Parallelism is Not Automatic

43 Data and Model Parallelism ypred = WX + b X mat mul add ypred graph W b

44 Data and Model Parallelism Data Parallelism X split mat mul graph add gpu:0 ypred = WX + b mat mul graph add gpu:1 X mat mul add ypred graph W b

45 Data and Model Parallelism Data Parallelism X split mat mul graph add gpu:0 ypred = WX + b mat mul graph add gpu:1 Model Parallelism X mat mul add graph ypred X mat mul graph add gpu:0 gpu:1 W b

46 Data and Model Parallelism Data Parallelism X split mat mul graph add gpu:0 ypred = WX + b mat mul graph add gpu:1 Model Parallelism X mat mul add graph ypred X split mat mul graph add gpu:0 gpu:1 W b mat mul graph add gpu:2 gpu:3

47 Output Reduction X split mat mul graph add gpu:0 [ grad(gpu:0), ypred(gpu:0), loss(gpu:0) ] mat mul graph add gpu:1 [ grad(gpu:1), ypred(gpu:1), loss(gpu:1) ]

48 Output Reduction X split mat mul graph add gpu:0 [ grad(gpu:0), ypred(gpu:0), loss(gpu:0) ] mat mul graph add gpu:1 [ grad(gpu:1), ypred(gpu:1), loss(gpu:1) ] avg concat sum [ grad(model), ypred(model), loss(model) ] Update Model Params

49 Output Reduction gpu:0 gpu:0 gpu:1 gpu:2 cpu/gpu gpu:3 gpu:1 gpu:3 gpu:2 Naive Consolidation Ring All-Reduce Baidu All-Reduce: Andrew Ng

50 Distributed Accelerator Options replicate_model_fn (TF1.5+; Naive Consolidation; Single Server) distribute.mirroredstrategy (TF1.8+; All-Reduce; Single Server for now) TPUEstimator (TF1.4+; All-Reduce; 64-TPU pod) Horovod (TF1.8+; All-Reduce; Multi Server) multi_gpu_model (Keras; Naive Consolidation; Single Server) Meet Horovod: Alex Sergeev, Mike Del Balso

51 TPU / TPU Pod TPUs Batch Size Time to 90 Epochs Accuracy : % : % : % min 76.1% min 75.0% > min 76.1% ImageNet is the New MNIST

52 Summary tf.estimator Premade Estimators for quick and dirty jobs (5-part recipe) Custom models w/o the plumbing (5-part function) Keras Functional API for Inference() in custom model_fn() Use Checkpoints/Preemptible VMs/Cloud Storage to reduce cost Scalable model building from laptop to GPUs to TPU Pods TFRecords & tf.data for pipeline performance Parallelize creation of TFRecord files ~ MB Use tf.data input_fn() Distributed Accelerator options Input pipeline before bigger/faster/more accelerators Scale Up before Out

53 Thank you.

54

55

Keras: Handwritten Digit Recognition using MNIST Dataset

Keras: Handwritten Digit Recognition using MNIST Dataset Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA February 9, 2017 1 / 24 OUTLINE 1 Introduction Keras: Deep Learning library for Theano and TensorFlow 2 Installing Keras Installation

More information

Keras: Handwritten Digit Recognition using MNIST Dataset

Keras: Handwritten Digit Recognition using MNIST Dataset Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA January 31, 2018 1 / 30 OUTLINE 1 Keras: Introduction 2 Installing Keras 3 Keras: Building, Testing, Improving A Simple Network 2 / 30

More information

POINT CLOUD DEEP LEARNING

POINT CLOUD DEEP LEARNING POINT CLOUD DEEP LEARNING Innfarn Yoo, 3/29/28 / 57 Introduction AGENDA Previous Work Method Result Conclusion 2 / 57 INTRODUCTION 3 / 57 2D OBJECT CLASSIFICATION Deep Learning for 2D Object Classification

More information

Deep Learning for Computer Vision with MATLAB By Jon Cherrie

Deep Learning for Computer Vision with MATLAB By Jon Cherrie Deep Learning for Computer Vision with MATLAB By Jon Cherrie 2015 The MathWorks, Inc. 1 Deep learning is getting a lot of attention "Dahl and his colleagues won $22,000 with a deeplearning system. 'We

More information

SVM multiclass classification in 10 steps 17/32

SVM multiclass classification in 10 steps 17/32 SVM multiclass classification in 10 steps import numpy as np # load digits dataset from sklearn import datasets digits = datasets. load_digits () # define training set size n_samples = len ( digits. images

More information

Deep Learning with Tensorflow AlexNet

Deep Learning with Tensorflow   AlexNet Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification

More information

Hardware and Software. Fei-Fei Li & Justin Johnson & Serena Yeung. Lecture 6-1

Hardware and Software. Fei-Fei Li & Justin Johnson & Serena Yeung. Lecture 6-1 Lecture 6: Hardware and Software Lecture 6-1 Administrative Assignment 1 was due yesterday. Assignment 2 is out, due Wed May 1. Project proposal due Wed April 24. Project-only office hours leading up to

More information

OPENSEQ2SEQ: A DEEP LEARNING TOOLKIT FOR SPEECH RECOGNITION, SPEECH SYNTHESIS, AND NLP

OPENSEQ2SEQ: A DEEP LEARNING TOOLKIT FOR SPEECH RECOGNITION, SPEECH SYNTHESIS, AND NLP OPENSEQ2SEQ: A DEEP LEARNING TOOLKIT FOR SPEECH RECOGNITION, SPEECH SYNTHESIS, AND NLP Boris Ginsburg, Vitaly Lavrukhin, Igor Gitman, Oleksii Kuchaiev, Jason Li, Vahid Noroozi, Ravi Teja Gadde, Chip Nguyen

More information

Machine Learning Workshop

Machine Learning Workshop Machine Learning Workshop {Presenters} Feb. 20th, 2018 Theory of Neural Networks Architecture and Types of Layers: Fully Connected (FC) Convolutional Neural Network (CNN) Pooling Drop out Residual Recurrent

More information

A Quick Guide on Training a neural network using Keras.

A Quick Guide on Training a neural network using Keras. A Quick Guide on Training a neural network using Keras. TensorFlow and Keras Keras Open source High level, less flexible Easy to learn Perfect for quick implementations Starts by François Chollet from

More information

Machine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center

Machine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center Machine Learning With Python Bin Chen Nov. 7, 2017 Research Computing Center Outline Introduction to Machine Learning (ML) Introduction to Neural Network (NN) Introduction to Deep Learning NN Introduction

More information

CNN Basics. Chongruo Wu

CNN Basics. Chongruo Wu CNN Basics Chongruo Wu Overview 1. 2. 3. Forward: compute the output of each layer Back propagation: compute gradient Updating: update the parameters with computed gradient Agenda 1. Forward Conv, Fully

More information

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn Intro to Deep Learning Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn Why this class? Deep Features Have been able to harness the big data in the most efficient and effective

More information

Inference Optimization Using TensorRT with Use Cases. Jack Han / 한재근 Solutions Architect NVIDIA

Inference Optimization Using TensorRT with Use Cases. Jack Han / 한재근 Solutions Architect NVIDIA Inference Optimization Using TensorRT with Use Cases Jack Han / 한재근 Solutions Architect NVIDIA Search Image NLP Maps TensorRT 4 Adoption Use Cases Speech Video AI Inference is exploding 1 Billion Videos

More information

Tutorial on Machine Learning Tools

Tutorial on Machine Learning Tools Tutorial on Machine Learning Tools Yanbing Xue Milos Hauskrecht Why do we need these tools? Widely deployed classical models No need to code from scratch Easy-to-use GUI Outline Matlab Apps Weka 3 UI TensorFlow

More information

R for SQListas, a Continuation

R for SQListas, a Continuation 3-2 - 1-0: Classifying Digits with R R for SQListas, a Continuation R for SQListas: Now that we're in the tidyverse... what can we do now? Machine Learning MNIST - the Drosophila of Machine Learning (attributed

More information

An introduction to TensorFlow! Chip Huyen CS224N 1/25/2018

An introduction to TensorFlow! Chip Huyen CS224N 1/25/2018 An introduction to TensorFlow! Chip Huyen (chiphuyen@cs.stanford.edu) CS224N 1/25/2018 1 2 Agenda Why TensorFlow Graphs and Sessions Linear Regression tf.data word2vec Structuring your model Managing experiments

More information

Deep Neural Network Hyperparameter Optimization with Genetic Algorithms

Deep Neural Network Hyperparameter Optimization with Genetic Algorithms Deep Neural Network Hyperparameter Optimization with Genetic Algorithms EvoDevo A Genetic Algorithm Framework Aaron Vose, Jacob Balma, Geert Wenes, and Rangan Sukumar Cray Inc. October 2017 Presenter Vose,

More information

Deep Learning Frameworks. COSC 7336: Advanced Natural Language Processing Fall 2017

Deep Learning Frameworks. COSC 7336: Advanced Natural Language Processing Fall 2017 Deep Learning Frameworks COSC 7336: Advanced Natural Language Processing Fall 2017 Today s lecture Deep learning software overview TensorFlow Keras Practical Graphical Processing Unit (GPU) From graphical

More information

MoonRiver: Deep Neural Network in C++

MoonRiver: Deep Neural Network in C++ MoonRiver: Deep Neural Network in C++ Chung-Yi Weng Computer Science & Engineering University of Washington chungyi@cs.washington.edu Abstract Artificial intelligence resurges with its dramatic improvement

More information

Assignment4. November 29, Follow the directions on https://www.tensorflow.org/install/ to install Tensorflow on your computer.

Assignment4. November 29, Follow the directions on https://www.tensorflow.org/install/ to install Tensorflow on your computer. Assignment4 November 29, 2017 1 CSE 252A Computer Vision I Fall 2017 1.1 Assignment 4 1.2 Problem 1: Install Tensorflow [2 pts] Follow the directions on https://www.tensorflow.org/install/ to install Tensorflow

More information

Simplifying ML Workflows with Apache Beam & TensorFlow Extended

Simplifying ML Workflows with Apache Beam & TensorFlow Extended Simplifying ML Workflows with Apache Beam & TensorFlow Extended Tyler Akidau @takidau Software Engineer at Google Apache Beam PMC Apache Beam Portable data-processing pipelines Example pipelines Python

More information

Defense Data Generation in Distributed Deep Learning System Se-Yoon Oh / ADD-IDAR

Defense Data Generation in Distributed Deep Learning System Se-Yoon Oh / ADD-IDAR Defense Data Generation in Distributed Deep Learning System Se-Yoon Oh / 2017. 10. 31 syoh@add.re.kr Page 1/36 Overview 1. Introduction 2. Data Generation Synthesis 3. Distributed Deep Learning 4. Conclusions

More information

NVIDIA FOR DEEP LEARNING. Bill Veenhuis

NVIDIA FOR DEEP LEARNING. Bill Veenhuis NVIDIA FOR DEEP LEARNING Bill Veenhuis bveenhuis@nvidia.com Nvidia is the world s leading ai platform ONE ARCHITECTURE CUDA 2 GPU: Perfect Companion for Accelerating Apps & A.I. CPU GPU 3 Intro to AI AGENDA

More information

Day 1 Lecture 6. Software Frameworks for Deep Learning

Day 1 Lecture 6. Software Frameworks for Deep Learning Day 1 Lecture 6 Software Frameworks for Deep Learning Packages Caffe Theano NVIDIA Digits Lasagne Keras Blocks Torch TensorFlow MxNet MatConvNet Nervana Neon Leaf Caffe Deep learning framework from Berkeley

More information

Deep Learning and Its Applications

Deep Learning and Its Applications Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent

More information

Deep Learning Training System: Adam & Google Cloud ML. CSE 5194: Introduction to High- Performance Deep Learning Qinghua Zhou

Deep Learning Training System: Adam & Google Cloud ML. CSE 5194: Introduction to High- Performance Deep Learning Qinghua Zhou Deep Learning Training System: Adam & Google Cloud ML CSE 5194: Introduction to High- Performance Deep Learning Qinghua Zhou Outline Project Adam: Building an Efficient and Scalable Deep Learning Training

More information

BAYESIAN GLOBAL OPTIMIZATION

BAYESIAN GLOBAL OPTIMIZATION BAYESIAN GLOBAL OPTIMIZATION Using Optimal Learning to Tune Deep Learning Pipelines Scott Clark scott@sigopt.com OUTLINE 1. Why is Tuning AI Models Hard? 2. Comparison of Tuning Methods 3. Bayesian Global

More information

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research An Introduction to Deep Learning with RapidMiner Philipp Schlunder - RapidMiner Research What s in store for today?. Things to know before getting started 2. What s Deep Learning anyway? 3. How to use

More information

All You Want To Know About CNNs. Yukun Zhu

All You Want To Know About CNNs. Yukun Zhu All You Want To Know About CNNs Yukun Zhu Deep Learning Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image

More information

EFFICIENT INFERENCE WITH TENSORRT. Han Vanholder

EFFICIENT INFERENCE WITH TENSORRT. Han Vanholder EFFICIENT INFERENCE WITH TENSORRT Han Vanholder AI INFERENCING IS EXPLODING 2 Trillion Messages Per Day On LinkedIn 500M Daily active users of iflytek 140 Billion Words Per Day Translated by Google 60

More information

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Tutorial on Keras CAP 6412 - ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Deep learning packages TensorFlow Google PyTorch Facebook AI research Keras Francois Chollet (now at Google) Chainer Company

More information

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:

More information

TensorFlow. Research at Scale. Rajat Monga

TensorFlow. Research at Scale. Rajat Monga TensorFlow Research at Scale Rajat Monga Decision Trees Signal Processing Neural Nets Linear Algebra BayesFlow Random Forests C++ Frontend Python Frontend... TensorFlow Distributed Execution Engine CPU

More information

CNN optimization. Rassadin A

CNN optimization. Rassadin A CNN optimization Rassadin A. 01.2017-02.2017 What to optimize? Training stage time consumption (CPU / GPU) Inference stage time consumption (CPU / GPU) Training stage memory consumption Inference stage

More information

If you installed VM and Linux libraries as in the tutorial, you should not get any errors. Otherwise, you may need to install wget or gunzip.

If you installed VM and Linux libraries as in the tutorial, you should not get any errors. Otherwise, you may need to install wget or gunzip. MNIST 1- Prepare Dataset cd $CAFFE_ROOT./data/mnist/get_mnist.sh./examples/mnist/create_mnist.sh If you installed VM and Linux libraries as in the tutorial, you should not get any errors. Otherwise, you

More information

Application of Deep Learning Techniques in Satellite Telemetry Analysis.

Application of Deep Learning Techniques in Satellite Telemetry Analysis. Application of Deep Learning Techniques in Satellite Telemetry Analysis. Greg Adamski, Member of Technical Staff L3 Technologies Telemetry and RF Products Julian Spencer Jones, Spacecraft Engineer Telenor

More information

Inception and Residual Networks. Hantao Zhang. Deep Learning with Python.

Inception and Residual Networks. Hantao Zhang. Deep Learning with Python. Inception and Residual Networks Hantao Zhang Deep Learning with Python https://en.wikipedia.org/wiki/residual_neural_network Deep Neural Network Progress from Large Scale Visual Recognition Challenge (ILSVRC)

More information

Two FPGA-DNN Projects: 1. Low Latency Multi-Layer Perceptrons using FPGAs 2. Acceleration of CNN Training on FPGA-based Clusters

Two FPGA-DNN Projects: 1. Low Latency Multi-Layer Perceptrons using FPGAs 2. Acceleration of CNN Training on FPGA-based Clusters Two FPGA-DNN Projects: 1. Low Latency Multi-Layer Perceptrons using FPGAs 2. Acceleration of CNN Training on FPGA-based Clusters *Argonne National Lab +BU & USTC Presented by Martin Herbordt Work by Ahmed

More information

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: Code Mania 2019 Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: 1. Introduction to Artificial Intelligence 2. Introduction to python programming and Environment

More information

Characterization and Benchmarking of Deep Learning. Natalia Vassilieva, PhD Sr. Research Manager

Characterization and Benchmarking of Deep Learning. Natalia Vassilieva, PhD Sr. Research Manager Characterization and Benchmarking of Deep Learning Natalia Vassilieva, PhD Sr. Research Manager Deep learning applications Vision Speech Text Other Search & information extraction Security/Video surveillance

More information

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017 COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization

More information

Deep Learning Frameworks with Spark and GPUs

Deep Learning Frameworks with Spark and GPUs Deep Learning Frameworks with Spark and GPUs Abstract Spark is a powerful, scalable, real-time data analytics engine that is fast becoming the de facto hub for data science and big data. However, in parallel,

More information

OPERATIONALIZING MACHINE LEARNING USING GPU ACCELERATED, IN-DATABASE ANALYTICS

OPERATIONALIZING MACHINE LEARNING USING GPU ACCELERATED, IN-DATABASE ANALYTICS OPERATIONALIZING MACHINE LEARNING USING GPU ACCELERATED, IN-DATABASE ANALYTICS 1 Why GPUs? A Tale of Numbers 100x Performance Increase Infrastructure Cost Savings Performance 100x gains over traditional

More information

Inception Network Overview. David White CS793

Inception Network Overview. David White CS793 Inception Network Overview David White CS793 So, Leonardo DiCaprio dreams about dreaming... https://m.media-amazon.com/images/m/mv5bmjaxmzy3njcxnf5bml5banbnxkftztcwnti5otm0mw@@._v1_sy1000_cr0,0,675,1 000_AL_.jpg

More information

Deep Learning Workshop. Nov. 20, 2015 Andrew Fishberg, Rowan Zellers

Deep Learning Workshop. Nov. 20, 2015 Andrew Fishberg, Rowan Zellers Deep Learning Workshop Nov. 20, 2015 Andrew Fishberg, Rowan Zellers Why deep learning? The ImageNet Challenge Goal: image classification with 1000 categories Top 5 error rate of 15%. Krizhevsky, Alex,

More information

Research Faculty Summit Systems Fueling future disruptions

Research Faculty Summit Systems Fueling future disruptions Research Faculty Summit 2018 Systems Fueling future disruptions Wolong: A Back-end Optimizer for Deep Learning Computation Jilong Xue Researcher, Microsoft Research Asia System Challenge in Deep Learning

More information

Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets)

Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets) Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets) Sayan D. Pathak, Ph.D., Principal ML Scientist, Microsoft Roland Fernandez, Senior Researcher, Microsoft Module Outline

More information

Deep Learning for Computer Vision II

Deep Learning for Computer Vision II IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L

More information

Accelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs

Accelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs Accelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs Ritchie Zhao 1, Weinan Song 2, Wentao Zhang 2, Tianwei Xing 3, Jeng-Hau Lin 4, Mani Srivastava 3, Rajesh Gupta 4, Zhiru

More information

DECISION SUPPORT SYSTEM USING TENSOR FLOW

DECISION SUPPORT SYSTEM USING TENSOR FLOW Volume 118 No. 24 2018 ISSN: 1314-3395 (on-line version) url: http://www.acadpubl.eu/hub/ http://www.acadpubl.eu/hub/ DECISION SUPPORT SYSTEM USING TENSOR FLOW D.Anji Reddy 1, G.Narasimha 2, K.Srinivas

More information

Fei-Fei Li & Justin Johnson & Serena Yeung

Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9-1 Administrative A2 due Wed May 2 Midterm: In-class Tue May 8. Covers material through Lecture 10 (Thu May 3). Sample midterm released on piazza. Midterm review session: Fri May 4 discussion

More information

tf-lenet Documentation

tf-lenet Documentation tf-lenet Documentation Release a1 Ragav Venkatesan Sep 27, 2017 Contents 1 Tutorial 3 1.1 Introduction............................................... 3 1.2 Dot-Product Layers...........................................

More information

Multi-Task Self-Supervised Visual Learning

Multi-Task Self-Supervised Visual Learning Multi-Task Self-Supervised Visual Learning Sikai Zhong March 4, 2018 COMPUTER SCIENCE Table of contents 1. Introduction 2. Self-supervised Tasks 3. Architectures 4. Experiments 1 Introduction Self-supervised

More information

BinaryMatcher2. Preparing the Data. Preparing the Raw Data: 32 rows Binary and 1-Hot. D. Thiebaut. March 19, 2017

BinaryMatcher2. Preparing the Data. Preparing the Raw Data: 32 rows Binary and 1-Hot. D. Thiebaut. March 19, 2017 BinaryMatcher2 D. Thiebaut March 19, 2017 This Jupyter Notebook illustrates how to design a simple multi-layer Tensorflow Neural Net to recognize integers coded in binary and output them as 1-hot vector.

More information

IMPROVEMENT OF DEEP LEARNING MODELS ON CLASSIFICATION TASKS USING HAAR TRANSFORM AND MODEL ENSEMBLE

IMPROVEMENT OF DEEP LEARNING MODELS ON CLASSIFICATION TASKS USING HAAR TRANSFORM AND MODEL ENSEMBLE IMPROVEMENT OF DEEP LEARNING MODELS ON CLASSIFICATION TASKS USING HAAR TRANSFORM AND MODEL ENSEMBLE Bachelor s thesis Valkeakoski, Automation Engineering Spring 2017 Tung Son Nguyen 1 ABSTRACT Automation

More information

S8873 GBM INFERENCING ON GPU. Shankara Rao Thejaswi Nanditale, Vinay Deshpande

S8873 GBM INFERENCING ON GPU. Shankara Rao Thejaswi Nanditale, Vinay Deshpande S8873 GBM INFERENCING ON GPU Shankara Rao Thejaswi Nanditale, Vinay Deshpande Introduction AGENDA Objective Experimental Results Implementation Details Conclusion 2 INTRODUCTION 3 BOOSTING What is it?

More information

Face Recognition A Deep Learning Approach

Face Recognition A Deep Learning Approach Face Recognition A Deep Learning Approach Lihi Shiloh Tal Perl Deep Learning Seminar 2 Outline What about Cat recognition? Classical face recognition Modern face recognition DeepFace FaceNet Comparison

More information

GPU FOR DEEP LEARNING. 周国峰 Wuhan University 2017/10/13

GPU FOR DEEP LEARNING. 周国峰 Wuhan University 2017/10/13 GPU FOR DEEP LEARNING chandlerz@nvidia.com 周国峰 Wuhan University 2017/10/13 Why Deep Learning Boost Today? Nvidia SDK for Deep Learning? Agenda CUDA 8.0 cudnn TensorRT (GIE) NCCL DIGITS 2 Why Deep Learning

More information

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning, A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,

More information

Convolutional Sequence to Sequence Learning. Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research

Convolutional Sequence to Sequence Learning. Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research Convolutional Sequence to Sequence Learning Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research Sequence generation Need to model a conditional distribution

More information

Training Deep Neural Networks (in parallel)

Training Deep Neural Networks (in parallel) Lecture 9: Training Deep Neural Networks (in parallel) Visual Computing Systems How would you describe this professor? Easy? Mean? Boring? Nerdy? Professor classification task Classifies professors as

More information

Convolutional Neural Network Layer Reordering for Acceleration

Convolutional Neural Network Layer Reordering for Acceleration R1-15 SASIMI 2016 Proceedings Convolutional Neural Network Layer Reordering for Acceleration Vijay Daultani Subhajit Chaudhury Kazuhisa Ishizaka System Platform Labs Value Co-creation Center System Platform

More information

Tensorflow Example: Fizzbuzz. Sargur N. Srihari

Tensorflow Example: Fizzbuzz. Sargur N. Srihari Tensorflow Example: Fizzbuzz Sargur N. srihari@cedar.buffalo.edu 1 Fizzbuzz in Tensorflow Fizzbuzz: Print the numbers from 1 to 100, except that if the number is divisible by 3 print "fizz", if it's divisible

More information

A Deep Learning Approach to Vehicle Speed Estimation

A Deep Learning Approach to Vehicle Speed Estimation A Deep Learning Approach to Vehicle Speed Estimation Benjamin Penchas bpenchas@stanford.edu Tobin Bell tbell@stanford.edu Marco Monteiro marcorm@stanford.edu ABSTRACT Given car dashboard video footage,

More information

S8822 OPTIMIZING NMT WITH TENSORRT Micah Villmow Senior TensorRT Software Engineer

S8822 OPTIMIZING NMT WITH TENSORRT Micah Villmow Senior TensorRT Software Engineer S8822 OPTIMIZING NMT WITH TENSORRT Micah Villmow Senior TensorRT Software Engineer 2 100 倍以上速く 本当に可能ですか? 2 DOUGLAS ADAMS BABEL FISH Neural Machine Translation Unit 3 4 OVER 100X FASTER, IS IT REALLY POSSIBLE?

More information

Deep Nets with. Keras

Deep Nets with. Keras docs https://keras.io Deep Nets with Keras κέρας http://vem.quantumunlimited.org/the-gates-of-horn/ Professor Marie Roch These slides only cover enough to get started with feed-forward networks and do

More information

Hello Edge: Keyword Spotting on Microcontrollers

Hello Edge: Keyword Spotting on Microcontrollers Hello Edge: Keyword Spotting on Microcontrollers Yundong Zhang, Naveen Suda, Liangzhen Lai and Vikas Chandra ARM Research, Stanford University arxiv.org, 2017 Presented by Mohammad Mofrad University of

More information

App Engine MapReduce. Mike Aizatsky 11 May Hashtags: #io2011 #AppEngine Feedback:

App Engine MapReduce. Mike Aizatsky 11 May Hashtags: #io2011 #AppEngine Feedback: App Engine MapReduce Mike Aizatsky 11 May 2011 Hashtags: #io2011 #AppEngine Feedback: http://goo.gl/snv2i Agenda MapReduce Computational Model Mapper library Announcement Technical bits: Files API User-space

More information

Convolutional Neural Networks

Convolutional Neural Networks NPFL114, Lecture 4 Convolutional Neural Networks Milan Straka March 25, 2019 Charles University in Prague Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics unless otherwise

More information

Real-time convolutional networks for sonar image classification in low-power embedded systems

Real-time convolutional networks for sonar image classification in low-power embedded systems Real-time convolutional networks for sonar image classification in low-power embedded systems Matias Valdenegro-Toro Ocean Systems Laboratory - School of Engineering & Physical Sciences Heriot-Watt University,

More information

Brainchip OCTOBER

Brainchip OCTOBER Brainchip OCTOBER 2017 1 Agenda Neuromorphic computing background Akida Neuromorphic System-on-Chip (NSoC) Brainchip OCTOBER 2017 2 Neuromorphic Computing Background Brainchip OCTOBER 2017 3 A Brief History

More information

TENSORFLOW: LARGE-SCALE MACHINE LEARNING ON HETEROGENEOUS DISTRIBUTED SYSTEMS. By Sanjay Surendranath Girija

TENSORFLOW: LARGE-SCALE MACHINE LEARNING ON HETEROGENEOUS DISTRIBUTED SYSTEMS. By Sanjay Surendranath Girija TENSORFLOW: LARGE-SCALE MACHINE LEARNING ON HETEROGENEOUS DISTRIBUTED SYSTEMS By Sanjay Surendranath Girija WHAT IS TENSORFLOW? TensorFlow is an interface for expressing machine learning algorithms, and

More information

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University. Visualizing and Understanding Convolutional Networks Christopher Pennsylvania State University February 23, 2015 Some Slide Information taken from Pierre Sermanet (Google) presentation on and Computer

More information

Deep Learning for Embedded Security Evaluation

Deep Learning for Embedded Security Evaluation Deep Learning for Embedded Security Evaluation Emmanuel Prouff 1 1 Laboratoire de Sécurité des Composants, ANSSI, France April 2018, CISCO April 2018, CISCO E. Prouff 1/22 Contents 1. Context and Motivation

More information

Deep Learning Applications

Deep Learning Applications October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning

More information

Basic Models in TensorFlow. CS 20SI: TensorFlow for Deep Learning Research Lecture 3 1/20/2017

Basic Models in TensorFlow. CS 20SI: TensorFlow for Deep Learning Research Lecture 3 1/20/2017 Basic Models in TensorFlow CS 20SI: TensorFlow for Deep Learning Research Lecture 3 1/20/2017 1 2 Agenda Review Linear regression in TensorFlow Optimizers Logistic regression on MNIST Loss functions 3

More information

Onto Petaflops with Kubernetes

Onto Petaflops with Kubernetes Onto Petaflops with Kubernetes Vishnu Kannan Google Inc. vishh@google.com Key Takeaways Kubernetes can manage hardware accelerators at Scale Kubernetes provides a playground for ML ML journey with Kubernetes

More information

NVIDIA DATA LOADING LIBRARY (DALI)

NVIDIA DATA LOADING LIBRARY (DALI) NVIDIA DATA LOADING LIBRARY (DALI) RN-09096-001 _v01 September 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter Chapter 1. 2. 3. 4. 5. DALI DALI DALI DALI DALI Overview...1 Release

More information

Parallel Deep Network Training

Parallel Deep Network Training Lecture 19: Parallel Deep Network Training Parallel Computer Architecture and Programming How would you describe this professor? Easy? Mean? Boring? Nerdy? Professor classification task Classifies professors

More information

Lecture 3: Overview of Deep Learning System. CSE599W: Spring 2018

Lecture 3: Overview of Deep Learning System. CSE599W: Spring 2018 Lecture 3: Overview of Deep Learning System CSE599W: Spring 2018 The Deep Learning Systems Juggle We won t focus on a specific one, but will discuss the common and useful elements of these systems Typical

More information

MIXED PRECISION TRAINING OF NEURAL NETWORKS. Carl Case, Senior Architect, NVIDIA

MIXED PRECISION TRAINING OF NEURAL NETWORKS. Carl Case, Senior Architect, NVIDIA MIXED PRECISION TRAINING OF NEURAL NETWORKS Carl Case, Senior Architect, NVIDIA OUTLINE 1. What is mixed precision training with FP16? 2. Considerations and methodology for mixed precision training 3.

More information

DEEP NEURAL NETWORKS AND GPUS. Julie Bernauer

DEEP NEURAL NETWORKS AND GPUS. Julie Bernauer DEEP NEURAL NETWORKS AND GPUS Julie Bernauer GPU Computing GPU Computing Run Computations on GPUs x86 CUDA Framework to Program NVIDIA GPUs A simple sum of two vectors (arrays) in C void vector_add(int

More information

Dynamic Routing Between Capsules

Dynamic Routing Between Capsules Report Explainable Machine Learning Dynamic Routing Between Capsules Author: Michael Dorkenwald Supervisor: Dr. Ullrich Köthe 28. Juni 2018 Inhaltsverzeichnis 1 Introduction 2 2 Motivation 2 3 CapusleNet

More information

Street Address Image Recognition (CNN) Machine Learning Engineer Nanodegree

Street Address Image Recognition (CNN) Machine Learning Engineer Nanodegree Street Address Image Recognition (CNN) Machine Learning Engineer Nanodegree YOONSU PARK Oct 22 2016 Definition Project overview With the big advance in computing power with the enhanced deep learning technique,

More information

Training Neural Networks with Mixed Precision MICHAEL CARILLI CHRISTIAN SAROFEEN MICHAEL RUBERRY BEN BARSDELL

Training Neural Networks with Mixed Precision MICHAEL CARILLI CHRISTIAN SAROFEEN MICHAEL RUBERRY BEN BARSDELL Training Neural Networks with Mixed Precision MICHAEL CARILLI CHRISTIAN SAROFEEN MICHAEL RUBERRY BEN BARSDELL 1 THIS TALK Using mixed precision and Volta your networks can be: 1. 2-4x faster 2. half the

More information

Two-Stream Convolutional Networks for Action Recognition in Videos

Two-Stream Convolutional Networks for Action Recognition in Videos Two-Stream Convolutional Networks for Action Recognition in Videos Karen Simonyan Andrew Zisserman Cemil Zalluhoğlu Introduction Aim Extend deep Convolution Networks to action recognition in video. Motivation

More information

A performance comparison of Deep Learning frameworks on KNL

A performance comparison of Deep Learning frameworks on KNL A performance comparison of Deep Learning frameworks on KNL R. Zanella, G. Fiameni, M. Rorro Middleware, Data Management - SCAI - CINECA IXPUG Bologna, March 5, 2018 Table of Contents 1. Problem description

More information

Frameworks in Python for Numeric Computation / ML

Frameworks in Python for Numeric Computation / ML Frameworks in Python for Numeric Computation / ML Why use a framework? Why not use the built-in data structures? Why not write our own matrix multiplication function? Frameworks are needed not only because

More information

Convolutional Neural Networks and Supervised Learning

Convolutional Neural Networks and Supervised Learning Convolutional Neural Networks and Supervised Learning Eilif Solberg August 30, 2018 Outline Convolutional Architectures Convolutional neural networks Training Loss Optimization Regularization Hyperparameter

More information

IST 597 Deep Learning Overfitting and Regularization. Sep. 27, 2018

IST 597 Deep Learning Overfitting and Regularization. Sep. 27, 2018 IST 597 Deep Learning Overfitting and Regularization 1. Overfitting Sep. 27, 2018 Regression model y 1 3 x3 13 2 x2 36x10 import numpy as np import matplotlib.pyplot as plt from sklearn.linear_model import

More information

Pouya Kousha Fall 2018 CSE 5194 Prof. DK Panda

Pouya Kousha Fall 2018 CSE 5194 Prof. DK Panda Pouya Kousha Fall 2018 CSE 5194 Prof. DK Panda 1 Motivation And Intro Programming Model Spark Data Transformation Model Construction Model Training Model Inference Execution Model Data Parallel Training

More information

Emel: Deep Learning in One Line using Type Inference, Architecture Selection, and Hyperparameter Tuning

Emel: Deep Learning in One Line using Type Inference, Architecture Selection, and Hyperparameter Tuning Emel: Deep Learning in One Line using Type Inference, Architecture Selection, and Hyperparameter Tuning BARAK OSHRI and NISHITH KHANDWALA We present Emel, a new framework for training baseline supervised

More information

Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class

Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class Guidelines Submission. Submit a hardcopy of the report containing all the figures and printouts of code in class. For readability

More information

PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space

PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space Sikai Zhong February 14, 2018 COMPUTER SCIENCE Table of contents 1. PointNet 2. PointNet++ 3. Experiments 1 PointNet Property

More information

Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018

Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Outline: Introduction Action classification architectures

More information

ImageNet Classification with Deep Convolutional Neural Networks

ImageNet Classification with Deep Convolutional Neural Networks ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky Ilya Sutskever Geoffrey Hinton University of Toronto Canada Paper with same name to appear in NIPS 2012 Main idea Architecture

More information

Analyzing rtweet data with kerasformula

Analyzing rtweet data with kerasformula Analyzing rtweet data with kerasformula Pete Mohanty BARUG March 2018 hosted by Google Overview Now on CRAN, kerasformula package offers a high-level interface for the R interface to Keras. install.packages("kerasformula")

More information

Polytechnic University of Tirana

Polytechnic University of Tirana 1 Polytechnic University of Tirana Department of Computer Engineering SIBORA THEODHOR ELINDA KAJO M ECE 2 Computer Vision OCR AND BEYOND THE PRESENTATION IS ORGANISED IN 3 PARTS : 3 Introduction, previous

More information

Deep Model Compression

Deep Model Compression Deep Model Compression Xin Wang Oct.31.2016 Some of the contents are borrowed from Hinton s and Song s slides. Two papers Distilling the Knowledge in a Neural Network by Geoffrey Hinton et al What s the

More information