Developing Machine Learning Models. Kevin Tsai
|
|
- Adam Copeland
- 5 years ago
- Views:
Transcription
1 Developing Machine Learning Models Kevin Tsai
2
3 GOTO Chicago 2018 Developing Machine Learning Models Kevin Tsai, Google
4 Agenda Machine Learning with tf.estimator Performance pipelines with TFRecords and tf.data Distributed Training with GPUs and TPUs
5 Machine Learning with tf.estimator
6 Machine Learning Pipeline Training Data Transform & Feature Engineering Model Building Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
7 Machine Learning Pipeline Training Data Transform & Feature Engineering split() train() Model evaluate() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
8 Machine Learning Pipeline Training Data Transform & Feature Engineering split() train() Logs Model evaluate() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
9 Machine Learning Pipeline Training Data Transform & Feature Engineering split() train() Logs Model export() evaluate() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
10 Machine Learning Pipeline with tf.estimator Training Data Transform & Feature Engineering model_fn estimator tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
11 Machine Learning Pipeline with tf.estimator train() Training Data Transform & Feature Engineering model_fn evaluate() predict() estimator tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
12 Machine Learning Pipeline with tf.estimator train_input_fn train() Training Data Transform & Feature Engineering eval_input_fn model_fn evaluate() predict_input_fn predict() estimator tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
13 Machine Learning Pipeline with tf.estimator train_input_fn train() Training Data Transform & Feature Engineering eval_input_fn model_fn evaluate() predict_input_fn predict() estimator Logs tf.estimator Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
14 Machine Learning Pipeline with tf.estimator train_input_fn train() Training Data Transform & Feature Engineering eval_input_fn model_fn evaluate() predict_input_fn predict() serving_input_rec_fn estimator Logs tf.estimator export_savedmodel() Training Pipeline Inference Pipeline Inference Requests Transform & Feature Engineering Model Serving Predictions
15 Estimator in TensorFlow Canned estimators Models in a box Estimator Keras model Train and evaluate models Layers Build models Python C++ Java Go... TensorFlow Distributed Execution Engine CPU GPU TPU Android ios...
16 Example: MNIST? 1 MNIST: Yann LeCunn
17 Steps to Run Training on Premade Estimator Training: 1. Import data 2. Create input functions 3. Create estimator 4. Create train and evaluate specs 5. Run train and evaluate Test Inference: 6. Test run predict() 7. Export model
18 Premade and Custom Estimators Premade Estimators DNNClassifier DNNRegressor LinearClassifier LinearRegressor DNNLinearCombinedClassifier DNNLinearCombinedRegressor + tf.contrib.estimator or Custom Estimator
19 Creating Custom Estimator
20 #1 Inference() #1 Inference() (-1,784)
21 #1 Inference() #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool
22 #1 Inference() #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) (-1,3136) (-1,1024) Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool Reshape Dense ReLU Dropout
23 #1 Inference() #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) (-1,3136) (-1,1024) (-1,10) (-1,10) Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool Reshape Dense ReLU Dropout Dense Softmax
24 #1 Inference() (-1,784) (-1,28,28,1) (-1,14,14,64) (-1,7,7,64) (-1,3136) (-1,1024) (-1,10) (-1,10) #1 Inference() Reshape Convolution ReLU Max Pool Convolution ReLU Max Pool tf.name_scope log tf.nn.conv2d tf.nn.relu tf.nn.max_pool tf.summary.histogram( tf.summary.histogram tf.summary.histogram Reshape Dense ReLU Dropout Dense Softmax tf.name_scope log relu=false tf.add tf.matmul tf.nn.relu tf.summary.histogram tf.summary.histogram tf.summary.histogram Hands-on TensorBoard: Dandelion Mane
25 #1 Inference() and #2 Calculations and Metrics #### 1 INFERENCE MODEL input_layer = tf.reshape(features["x"], [-1, 28, 28, 1]) conv1 = _conv(input_layer,kernel=[5,5,1,64],name='conv1',log=params['log']) conv2 = _conv(conv1,kernel=[5,5,64,64],name='conv2',log=params['log']) dense = _dense(conv2,size_in=7*7*64,size_out=params['dense_units'], name='dense',relu=true,log=params['log']) if mode==tf.estimator.modekeys.train: dense = tf.nn.dropout(dense,params['drop_out']) logits = _dense(dense,size_in=params['dense_units'], size_out=10,name='output',relu=false,log=params['log']) #1 Inference()
26 #1 Inference() and #2 Calculations and Metrics #### 2 CALCULATIONS AND METRICS predictions = {"classes": tf.argmax(input=logits,axis=1), "logits": logits, "probabilities": tf.nn.softmax(logits,name='softmax')} export_outputs = {'predictions': tf.estimator.export.predictoutput(predictions)} if (mode==tf.estimator.modekeys.train or mode==tf.estimator.modekeys.eval): loss = tf.losses.sparse_softmax_cross_entropy(labels=labels,logits=logits) accuracy = tf.metrics.accuracy( labels=labels, predictions=tf.argmax(logits,axis=1)) metrics = {'accuracy':accuracy} tf.summary.scalar('accuracy',accuracy[1]) #2 Calculations
27 #3 Predict, #4 Train, and #5 Eval #### 3 MODE = PREDICT if mode == tf.estimator.modekeys.predict: return tf.estimator.estimatorspec( mode=mode, predictions=predictions, export_outputs=export_outputs) #3 MODE = PREDICT
28 #3 Predict, #4 Train, and #5 Eval #### 4 MODE = TRAIN if mode == tf.estimator.modekeys.train: learning_rate = tf.train.exponential_decay( params['learning_rate'],tf.train.get_global_step(), decay_steps=100000,decay_rate=0.96) optimizer = tf.train.gradientdescentoptimizer(learning_rate=learning_rate) if params['replicate']==true: optimizer = tf.contrib.estimator.toweroptimizer(optimizer) train_op = optimizer.minimize(loss=loss,global_step=tf.train.get_global_step()) tf.summary.scalar('learning_rate', learning_rate) return tf.estimator.estimatorspec( mode=mode, loss=loss, train_op=train_op) #4 MODE = TRAIN
29 #3 Predict, #4 Train, and #5 Eval #5 MODE = EVAL #### 5 MODE = EVAL if mode == tf.estimator.modekeys.eval: return tf.estimator.estimatorspec( mode=mode,loss=loss,eval_metric_ops=metrics)
30 TFRecords and tf.data
31 (Inefficient) Pipeline # Load Data mnist = tf.contrib.learn.datasets.load_dataset("mnist") train_data = mnist.train.images train_labels = np.asarray(mnist.train.labels, dtype=np.int32) eval_data = mnist.test.images eval_labels = np.asarray(mnist.test.labels, dtype=np.int32) # Create Input Functions train_input_fn = tf.estimator.inputs.numpy_input_fn( x={"x": train_data},y=train_labels,batch_size=100,num_epochs=none,shuffle=true) eval_input_fn = tf.estimator.inputs.numpy_input_fn( x={"x": eval_data},y=eval_labels,num_epochs=1,shuffle=false) pred_input_fn = tf.estimator.inputs.numpy_input_fn( x={"x": eval_data},num_epochs=1,shuffle=false)
32 TFRecord tf.train.example tf.train.example tf.train.example tf.train.example tf.train.example... train004.tfrecords
33 TFRecord tf.train.example tf.train.example tf.train.example tf.train.example tf.train.example train004.tfrecords tf.python_io.tfrecordwriter tf.train.example tf.train.feature tf.train.features tf.train.feature tf.train.feature tf.train.feature tf.train.feature... tf.train.example tf.train.features...
34 tf.data tf.train.example tf.train.features tf.train.feature tf.train.feature tf.train.feature... tf.fixedlenfeature tf.fixedlenfeature tf.fixedlenfeature tf.parse_single_example
35 tf.data input_fn() dataset = tf.data.tfrecorddataset(c['files'],num_parallel_reads=c['threads'])
36 tf.data input_fn() dataset = dataset.map(parse_tfrecord,num_parallel_calls=c['threads']) dataset = dataset.map(lambda x,y: (image_scaling(x),y),num_parallel_calls=c['threads'])
37 tf.data input_fn() if c['mode']==tf.estimator.modekeys.train: dataset = dataset.map(lambda x,y: (distort(x),y),num_parallel_calls=c['threads']) dataset = dataset.shuffle(buffer_size=c['shuffle_buff'])
38 tf.data input_fn() dataset = dataset.repeat() dataset = dataset.batch(c['batch']) dataset = dataset.prefetch(2*c['batch'])
39 tf.data input_fn() iterator = dataset.make_one_shot_iterator() images, labels = iterator.get_next() features = {'images': images} return features, labels
40 tf.data input_fn() train_spec = tf.estimator.trainspec(input_fn=lambda: dataset_input_fn(train_params)) eval_spec = tf.estimator.evalspec(input_fn=lambda: dataset_input_fn(eval_params))
41 Distributed Training
42 Parallelism is Not Automatic
43 Data and Model Parallelism ypred = WX + b X mat mul add ypred graph W b
44 Data and Model Parallelism Data Parallelism X split mat mul graph add gpu:0 ypred = WX + b mat mul graph add gpu:1 X mat mul add ypred graph W b
45 Data and Model Parallelism Data Parallelism X split mat mul graph add gpu:0 ypred = WX + b mat mul graph add gpu:1 Model Parallelism X mat mul add graph ypred X mat mul graph add gpu:0 gpu:1 W b
46 Data and Model Parallelism Data Parallelism X split mat mul graph add gpu:0 ypred = WX + b mat mul graph add gpu:1 Model Parallelism X mat mul add graph ypred X split mat mul graph add gpu:0 gpu:1 W b mat mul graph add gpu:2 gpu:3
47 Output Reduction X split mat mul graph add gpu:0 [ grad(gpu:0), ypred(gpu:0), loss(gpu:0) ] mat mul graph add gpu:1 [ grad(gpu:1), ypred(gpu:1), loss(gpu:1) ]
48 Output Reduction X split mat mul graph add gpu:0 [ grad(gpu:0), ypred(gpu:0), loss(gpu:0) ] mat mul graph add gpu:1 [ grad(gpu:1), ypred(gpu:1), loss(gpu:1) ] avg concat sum [ grad(model), ypred(model), loss(model) ] Update Model Params
49 Output Reduction gpu:0 gpu:0 gpu:1 gpu:2 cpu/gpu gpu:3 gpu:1 gpu:3 gpu:2 Naive Consolidation Ring All-Reduce Baidu All-Reduce: Andrew Ng
50 Distributed Accelerator Options replicate_model_fn (TF1.5+; Naive Consolidation; Single Server) distribute.mirroredstrategy (TF1.8+; All-Reduce; Single Server for now) TPUEstimator (TF1.4+; All-Reduce; 64-TPU pod) Horovod (TF1.8+; All-Reduce; Multi Server) multi_gpu_model (Keras; Naive Consolidation; Single Server) Meet Horovod: Alex Sergeev, Mike Del Balso
51 TPU / TPU Pod TPUs Batch Size Time to 90 Epochs Accuracy : % : % : % min 76.1% min 75.0% > min 76.1% ImageNet is the New MNIST
52 Summary tf.estimator Premade Estimators for quick and dirty jobs (5-part recipe) Custom models w/o the plumbing (5-part function) Keras Functional API for Inference() in custom model_fn() Use Checkpoints/Preemptible VMs/Cloud Storage to reduce cost Scalable model building from laptop to GPUs to TPU Pods TFRecords & tf.data for pipeline performance Parallelize creation of TFRecord files ~ MB Use tf.data input_fn() Distributed Accelerator options Input pipeline before bigger/faster/more accelerators Scale Up before Out
53 Thank you.
54
55
Keras: Handwritten Digit Recognition using MNIST Dataset
Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA February 9, 2017 1 / 24 OUTLINE 1 Introduction Keras: Deep Learning library for Theano and TensorFlow 2 Installing Keras Installation
More informationKeras: Handwritten Digit Recognition using MNIST Dataset
Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA January 31, 2018 1 / 30 OUTLINE 1 Keras: Introduction 2 Installing Keras 3 Keras: Building, Testing, Improving A Simple Network 2 / 30
More informationPOINT CLOUD DEEP LEARNING
POINT CLOUD DEEP LEARNING Innfarn Yoo, 3/29/28 / 57 Introduction AGENDA Previous Work Method Result Conclusion 2 / 57 INTRODUCTION 3 / 57 2D OBJECT CLASSIFICATION Deep Learning for 2D Object Classification
More informationDeep Learning for Computer Vision with MATLAB By Jon Cherrie
Deep Learning for Computer Vision with MATLAB By Jon Cherrie 2015 The MathWorks, Inc. 1 Deep learning is getting a lot of attention "Dahl and his colleagues won $22,000 with a deeplearning system. 'We
More informationSVM multiclass classification in 10 steps 17/32
SVM multiclass classification in 10 steps import numpy as np # load digits dataset from sklearn import datasets digits = datasets. load_digits () # define training set size n_samples = len ( digits. images
More informationDeep Learning with Tensorflow AlexNet
Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification
More informationHardware and Software. Fei-Fei Li & Justin Johnson & Serena Yeung. Lecture 6-1
Lecture 6: Hardware and Software Lecture 6-1 Administrative Assignment 1 was due yesterday. Assignment 2 is out, due Wed May 1. Project proposal due Wed April 24. Project-only office hours leading up to
More informationOPENSEQ2SEQ: A DEEP LEARNING TOOLKIT FOR SPEECH RECOGNITION, SPEECH SYNTHESIS, AND NLP
OPENSEQ2SEQ: A DEEP LEARNING TOOLKIT FOR SPEECH RECOGNITION, SPEECH SYNTHESIS, AND NLP Boris Ginsburg, Vitaly Lavrukhin, Igor Gitman, Oleksii Kuchaiev, Jason Li, Vahid Noroozi, Ravi Teja Gadde, Chip Nguyen
More informationMachine Learning Workshop
Machine Learning Workshop {Presenters} Feb. 20th, 2018 Theory of Neural Networks Architecture and Types of Layers: Fully Connected (FC) Convolutional Neural Network (CNN) Pooling Drop out Residual Recurrent
More informationA Quick Guide on Training a neural network using Keras.
A Quick Guide on Training a neural network using Keras. TensorFlow and Keras Keras Open source High level, less flexible Easy to learn Perfect for quick implementations Starts by François Chollet from
More informationMachine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center
Machine Learning With Python Bin Chen Nov. 7, 2017 Research Computing Center Outline Introduction to Machine Learning (ML) Introduction to Neural Network (NN) Introduction to Deep Learning NN Introduction
More informationCNN Basics. Chongruo Wu
CNN Basics Chongruo Wu Overview 1. 2. 3. Forward: compute the output of each layer Back propagation: compute gradient Updating: update the parameters with computed gradient Agenda 1. Forward Conv, Fully
More informationIntro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn
Intro to Deep Learning Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn Why this class? Deep Features Have been able to harness the big data in the most efficient and effective
More informationInference Optimization Using TensorRT with Use Cases. Jack Han / 한재근 Solutions Architect NVIDIA
Inference Optimization Using TensorRT with Use Cases Jack Han / 한재근 Solutions Architect NVIDIA Search Image NLP Maps TensorRT 4 Adoption Use Cases Speech Video AI Inference is exploding 1 Billion Videos
More informationTutorial on Machine Learning Tools
Tutorial on Machine Learning Tools Yanbing Xue Milos Hauskrecht Why do we need these tools? Widely deployed classical models No need to code from scratch Easy-to-use GUI Outline Matlab Apps Weka 3 UI TensorFlow
More informationR for SQListas, a Continuation
3-2 - 1-0: Classifying Digits with R R for SQListas, a Continuation R for SQListas: Now that we're in the tidyverse... what can we do now? Machine Learning MNIST - the Drosophila of Machine Learning (attributed
More informationAn introduction to TensorFlow! Chip Huyen CS224N 1/25/2018
An introduction to TensorFlow! Chip Huyen (chiphuyen@cs.stanford.edu) CS224N 1/25/2018 1 2 Agenda Why TensorFlow Graphs and Sessions Linear Regression tf.data word2vec Structuring your model Managing experiments
More informationDeep Neural Network Hyperparameter Optimization with Genetic Algorithms
Deep Neural Network Hyperparameter Optimization with Genetic Algorithms EvoDevo A Genetic Algorithm Framework Aaron Vose, Jacob Balma, Geert Wenes, and Rangan Sukumar Cray Inc. October 2017 Presenter Vose,
More informationDeep Learning Frameworks. COSC 7336: Advanced Natural Language Processing Fall 2017
Deep Learning Frameworks COSC 7336: Advanced Natural Language Processing Fall 2017 Today s lecture Deep learning software overview TensorFlow Keras Practical Graphical Processing Unit (GPU) From graphical
More informationMoonRiver: Deep Neural Network in C++
MoonRiver: Deep Neural Network in C++ Chung-Yi Weng Computer Science & Engineering University of Washington chungyi@cs.washington.edu Abstract Artificial intelligence resurges with its dramatic improvement
More informationAssignment4. November 29, Follow the directions on https://www.tensorflow.org/install/ to install Tensorflow on your computer.
Assignment4 November 29, 2017 1 CSE 252A Computer Vision I Fall 2017 1.1 Assignment 4 1.2 Problem 1: Install Tensorflow [2 pts] Follow the directions on https://www.tensorflow.org/install/ to install Tensorflow
More informationSimplifying ML Workflows with Apache Beam & TensorFlow Extended
Simplifying ML Workflows with Apache Beam & TensorFlow Extended Tyler Akidau @takidau Software Engineer at Google Apache Beam PMC Apache Beam Portable data-processing pipelines Example pipelines Python
More informationDefense Data Generation in Distributed Deep Learning System Se-Yoon Oh / ADD-IDAR
Defense Data Generation in Distributed Deep Learning System Se-Yoon Oh / 2017. 10. 31 syoh@add.re.kr Page 1/36 Overview 1. Introduction 2. Data Generation Synthesis 3. Distributed Deep Learning 4. Conclusions
More informationNVIDIA FOR DEEP LEARNING. Bill Veenhuis
NVIDIA FOR DEEP LEARNING Bill Veenhuis bveenhuis@nvidia.com Nvidia is the world s leading ai platform ONE ARCHITECTURE CUDA 2 GPU: Perfect Companion for Accelerating Apps & A.I. CPU GPU 3 Intro to AI AGENDA
More informationDay 1 Lecture 6. Software Frameworks for Deep Learning
Day 1 Lecture 6 Software Frameworks for Deep Learning Packages Caffe Theano NVIDIA Digits Lasagne Keras Blocks Torch TensorFlow MxNet MatConvNet Nervana Neon Leaf Caffe Deep learning framework from Berkeley
More informationDeep Learning and Its Applications
Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent
More informationDeep Learning Training System: Adam & Google Cloud ML. CSE 5194: Introduction to High- Performance Deep Learning Qinghua Zhou
Deep Learning Training System: Adam & Google Cloud ML CSE 5194: Introduction to High- Performance Deep Learning Qinghua Zhou Outline Project Adam: Building an Efficient and Scalable Deep Learning Training
More informationBAYESIAN GLOBAL OPTIMIZATION
BAYESIAN GLOBAL OPTIMIZATION Using Optimal Learning to Tune Deep Learning Pipelines Scott Clark scott@sigopt.com OUTLINE 1. Why is Tuning AI Models Hard? 2. Comparison of Tuning Methods 3. Bayesian Global
More informationAn Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research
An Introduction to Deep Learning with RapidMiner Philipp Schlunder - RapidMiner Research What s in store for today?. Things to know before getting started 2. What s Deep Learning anyway? 3. How to use
More informationAll You Want To Know About CNNs. Yukun Zhu
All You Want To Know About CNNs Yukun Zhu Deep Learning Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image
More informationEFFICIENT INFERENCE WITH TENSORRT. Han Vanholder
EFFICIENT INFERENCE WITH TENSORRT Han Vanholder AI INFERENCING IS EXPLODING 2 Trillion Messages Per Day On LinkedIn 500M Daily active users of iflytek 140 Billion Words Per Day Translated by Google 60
More informationTutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY
Tutorial on Keras CAP 6412 - ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY Deep learning packages TensorFlow Google PyTorch Facebook AI research Keras Francois Chollet (now at Google) Chainer Company
More informationConvolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech
Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:
More informationTensorFlow. Research at Scale. Rajat Monga
TensorFlow Research at Scale Rajat Monga Decision Trees Signal Processing Neural Nets Linear Algebra BayesFlow Random Forests C++ Frontend Python Frontend... TensorFlow Distributed Execution Engine CPU
More informationCNN optimization. Rassadin A
CNN optimization Rassadin A. 01.2017-02.2017 What to optimize? Training stage time consumption (CPU / GPU) Inference stage time consumption (CPU / GPU) Training stage memory consumption Inference stage
More informationIf you installed VM and Linux libraries as in the tutorial, you should not get any errors. Otherwise, you may need to install wget or gunzip.
MNIST 1- Prepare Dataset cd $CAFFE_ROOT./data/mnist/get_mnist.sh./examples/mnist/create_mnist.sh If you installed VM and Linux libraries as in the tutorial, you should not get any errors. Otherwise, you
More informationApplication of Deep Learning Techniques in Satellite Telemetry Analysis.
Application of Deep Learning Techniques in Satellite Telemetry Analysis. Greg Adamski, Member of Technical Staff L3 Technologies Telemetry and RF Products Julian Spencer Jones, Spacecraft Engineer Telenor
More informationInception and Residual Networks. Hantao Zhang. Deep Learning with Python.
Inception and Residual Networks Hantao Zhang Deep Learning with Python https://en.wikipedia.org/wiki/residual_neural_network Deep Neural Network Progress from Large Scale Visual Recognition Challenge (ILSVRC)
More informationTwo FPGA-DNN Projects: 1. Low Latency Multi-Layer Perceptrons using FPGAs 2. Acceleration of CNN Training on FPGA-based Clusters
Two FPGA-DNN Projects: 1. Low Latency Multi-Layer Perceptrons using FPGAs 2. Acceleration of CNN Training on FPGA-based Clusters *Argonne National Lab +BU & USTC Presented by Martin Herbordt Work by Ahmed
More informationCode Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:
Code Mania 2019 Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python: 1. Introduction to Artificial Intelligence 2. Introduction to python programming and Environment
More informationCharacterization and Benchmarking of Deep Learning. Natalia Vassilieva, PhD Sr. Research Manager
Characterization and Benchmarking of Deep Learning Natalia Vassilieva, PhD Sr. Research Manager Deep learning applications Vision Speech Text Other Search & information extraction Security/Video surveillance
More informationCOMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017
COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization
More informationDeep Learning Frameworks with Spark and GPUs
Deep Learning Frameworks with Spark and GPUs Abstract Spark is a powerful, scalable, real-time data analytics engine that is fast becoming the de facto hub for data science and big data. However, in parallel,
More informationOPERATIONALIZING MACHINE LEARNING USING GPU ACCELERATED, IN-DATABASE ANALYTICS
OPERATIONALIZING MACHINE LEARNING USING GPU ACCELERATED, IN-DATABASE ANALYTICS 1 Why GPUs? A Tale of Numbers 100x Performance Increase Infrastructure Cost Savings Performance 100x gains over traditional
More informationInception Network Overview. David White CS793
Inception Network Overview David White CS793 So, Leonardo DiCaprio dreams about dreaming... https://m.media-amazon.com/images/m/mv5bmjaxmzy3njcxnf5bml5banbnxkftztcwnti5otm0mw@@._v1_sy1000_cr0,0,675,1 000_AL_.jpg
More informationDeep Learning Workshop. Nov. 20, 2015 Andrew Fishberg, Rowan Zellers
Deep Learning Workshop Nov. 20, 2015 Andrew Fishberg, Rowan Zellers Why deep learning? The ImageNet Challenge Goal: image classification with 1000 categories Top 5 error rate of 15%. Krizhevsky, Alex,
More informationResearch Faculty Summit Systems Fueling future disruptions
Research Faculty Summit 2018 Systems Fueling future disruptions Wolong: A Back-end Optimizer for Deep Learning Computation Jilong Xue Researcher, Microsoft Research Asia System Challenge in Deep Learning
More informationDeep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets)
Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets) Sayan D. Pathak, Ph.D., Principal ML Scientist, Microsoft Roland Fernandez, Senior Researcher, Microsoft Module Outline
More informationDeep Learning for Computer Vision II
IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L
More informationAccelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs
Accelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs Ritchie Zhao 1, Weinan Song 2, Wentao Zhang 2, Tianwei Xing 3, Jeng-Hau Lin 4, Mani Srivastava 3, Rajesh Gupta 4, Zhiru
More informationDECISION SUPPORT SYSTEM USING TENSOR FLOW
Volume 118 No. 24 2018 ISSN: 1314-3395 (on-line version) url: http://www.acadpubl.eu/hub/ http://www.acadpubl.eu/hub/ DECISION SUPPORT SYSTEM USING TENSOR FLOW D.Anji Reddy 1, G.Narasimha 2, K.Srinivas
More informationFei-Fei Li & Justin Johnson & Serena Yeung
Lecture 9-1 Administrative A2 due Wed May 2 Midterm: In-class Tue May 8. Covers material through Lecture 10 (Thu May 3). Sample midterm released on piazza. Midterm review session: Fri May 4 discussion
More informationtf-lenet Documentation
tf-lenet Documentation Release a1 Ragav Venkatesan Sep 27, 2017 Contents 1 Tutorial 3 1.1 Introduction............................................... 3 1.2 Dot-Product Layers...........................................
More informationMulti-Task Self-Supervised Visual Learning
Multi-Task Self-Supervised Visual Learning Sikai Zhong March 4, 2018 COMPUTER SCIENCE Table of contents 1. Introduction 2. Self-supervised Tasks 3. Architectures 4. Experiments 1 Introduction Self-supervised
More informationBinaryMatcher2. Preparing the Data. Preparing the Raw Data: 32 rows Binary and 1-Hot. D. Thiebaut. March 19, 2017
BinaryMatcher2 D. Thiebaut March 19, 2017 This Jupyter Notebook illustrates how to design a simple multi-layer Tensorflow Neural Net to recognize integers coded in binary and output them as 1-hot vector.
More informationIMPROVEMENT OF DEEP LEARNING MODELS ON CLASSIFICATION TASKS USING HAAR TRANSFORM AND MODEL ENSEMBLE
IMPROVEMENT OF DEEP LEARNING MODELS ON CLASSIFICATION TASKS USING HAAR TRANSFORM AND MODEL ENSEMBLE Bachelor s thesis Valkeakoski, Automation Engineering Spring 2017 Tung Son Nguyen 1 ABSTRACT Automation
More informationS8873 GBM INFERENCING ON GPU. Shankara Rao Thejaswi Nanditale, Vinay Deshpande
S8873 GBM INFERENCING ON GPU Shankara Rao Thejaswi Nanditale, Vinay Deshpande Introduction AGENDA Objective Experimental Results Implementation Details Conclusion 2 INTRODUCTION 3 BOOSTING What is it?
More informationFace Recognition A Deep Learning Approach
Face Recognition A Deep Learning Approach Lihi Shiloh Tal Perl Deep Learning Seminar 2 Outline What about Cat recognition? Classical face recognition Modern face recognition DeepFace FaceNet Comparison
More informationGPU FOR DEEP LEARNING. 周国峰 Wuhan University 2017/10/13
GPU FOR DEEP LEARNING chandlerz@nvidia.com 周国峰 Wuhan University 2017/10/13 Why Deep Learning Boost Today? Nvidia SDK for Deep Learning? Agenda CUDA 8.0 cudnn TensorRT (GIE) NCCL DIGITS 2 Why Deep Learning
More informationIndex. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,
A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,
More informationConvolutional Sequence to Sequence Learning. Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research
Convolutional Sequence to Sequence Learning Denis Yarats with Jonas Gehring, Michael Auli, David Grangier, Yann Dauphin Facebook AI Research Sequence generation Need to model a conditional distribution
More informationTraining Deep Neural Networks (in parallel)
Lecture 9: Training Deep Neural Networks (in parallel) Visual Computing Systems How would you describe this professor? Easy? Mean? Boring? Nerdy? Professor classification task Classifies professors as
More informationConvolutional Neural Network Layer Reordering for Acceleration
R1-15 SASIMI 2016 Proceedings Convolutional Neural Network Layer Reordering for Acceleration Vijay Daultani Subhajit Chaudhury Kazuhisa Ishizaka System Platform Labs Value Co-creation Center System Platform
More informationTensorflow Example: Fizzbuzz. Sargur N. Srihari
Tensorflow Example: Fizzbuzz Sargur N. srihari@cedar.buffalo.edu 1 Fizzbuzz in Tensorflow Fizzbuzz: Print the numbers from 1 to 100, except that if the number is divisible by 3 print "fizz", if it's divisible
More informationA Deep Learning Approach to Vehicle Speed Estimation
A Deep Learning Approach to Vehicle Speed Estimation Benjamin Penchas bpenchas@stanford.edu Tobin Bell tbell@stanford.edu Marco Monteiro marcorm@stanford.edu ABSTRACT Given car dashboard video footage,
More informationS8822 OPTIMIZING NMT WITH TENSORRT Micah Villmow Senior TensorRT Software Engineer
S8822 OPTIMIZING NMT WITH TENSORRT Micah Villmow Senior TensorRT Software Engineer 2 100 倍以上速く 本当に可能ですか? 2 DOUGLAS ADAMS BABEL FISH Neural Machine Translation Unit 3 4 OVER 100X FASTER, IS IT REALLY POSSIBLE?
More informationDeep Nets with. Keras
docs https://keras.io Deep Nets with Keras κέρας http://vem.quantumunlimited.org/the-gates-of-horn/ Professor Marie Roch These slides only cover enough to get started with feed-forward networks and do
More informationHello Edge: Keyword Spotting on Microcontrollers
Hello Edge: Keyword Spotting on Microcontrollers Yundong Zhang, Naveen Suda, Liangzhen Lai and Vikas Chandra ARM Research, Stanford University arxiv.org, 2017 Presented by Mohammad Mofrad University of
More informationApp Engine MapReduce. Mike Aizatsky 11 May Hashtags: #io2011 #AppEngine Feedback:
App Engine MapReduce Mike Aizatsky 11 May 2011 Hashtags: #io2011 #AppEngine Feedback: http://goo.gl/snv2i Agenda MapReduce Computational Model Mapper library Announcement Technical bits: Files API User-space
More informationConvolutional Neural Networks
NPFL114, Lecture 4 Convolutional Neural Networks Milan Straka March 25, 2019 Charles University in Prague Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics unless otherwise
More informationReal-time convolutional networks for sonar image classification in low-power embedded systems
Real-time convolutional networks for sonar image classification in low-power embedded systems Matias Valdenegro-Toro Ocean Systems Laboratory - School of Engineering & Physical Sciences Heriot-Watt University,
More informationBrainchip OCTOBER
Brainchip OCTOBER 2017 1 Agenda Neuromorphic computing background Akida Neuromorphic System-on-Chip (NSoC) Brainchip OCTOBER 2017 2 Neuromorphic Computing Background Brainchip OCTOBER 2017 3 A Brief History
More informationTENSORFLOW: LARGE-SCALE MACHINE LEARNING ON HETEROGENEOUS DISTRIBUTED SYSTEMS. By Sanjay Surendranath Girija
TENSORFLOW: LARGE-SCALE MACHINE LEARNING ON HETEROGENEOUS DISTRIBUTED SYSTEMS By Sanjay Surendranath Girija WHAT IS TENSORFLOW? TensorFlow is an interface for expressing machine learning algorithms, and
More informationDeep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.
Visualizing and Understanding Convolutional Networks Christopher Pennsylvania State University February 23, 2015 Some Slide Information taken from Pierre Sermanet (Google) presentation on and Computer
More informationDeep Learning for Embedded Security Evaluation
Deep Learning for Embedded Security Evaluation Emmanuel Prouff 1 1 Laboratoire de Sécurité des Composants, ANSSI, France April 2018, CISCO April 2018, CISCO E. Prouff 1/22 Contents 1. Context and Motivation
More informationDeep Learning Applications
October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning
More informationBasic Models in TensorFlow. CS 20SI: TensorFlow for Deep Learning Research Lecture 3 1/20/2017
Basic Models in TensorFlow CS 20SI: TensorFlow for Deep Learning Research Lecture 3 1/20/2017 1 2 Agenda Review Linear regression in TensorFlow Optimizers Logistic regression on MNIST Loss functions 3
More informationOnto Petaflops with Kubernetes
Onto Petaflops with Kubernetes Vishnu Kannan Google Inc. vishh@google.com Key Takeaways Kubernetes can manage hardware accelerators at Scale Kubernetes provides a playground for ML ML journey with Kubernetes
More informationNVIDIA DATA LOADING LIBRARY (DALI)
NVIDIA DATA LOADING LIBRARY (DALI) RN-09096-001 _v01 September 2018 Release Notes TABLE OF CONTENTS Chapter Chapter Chapter Chapter Chapter 1. 2. 3. 4. 5. DALI DALI DALI DALI DALI Overview...1 Release
More informationParallel Deep Network Training
Lecture 19: Parallel Deep Network Training Parallel Computer Architecture and Programming How would you describe this professor? Easy? Mean? Boring? Nerdy? Professor classification task Classifies professors
More informationLecture 3: Overview of Deep Learning System. CSE599W: Spring 2018
Lecture 3: Overview of Deep Learning System CSE599W: Spring 2018 The Deep Learning Systems Juggle We won t focus on a specific one, but will discuss the common and useful elements of these systems Typical
More informationMIXED PRECISION TRAINING OF NEURAL NETWORKS. Carl Case, Senior Architect, NVIDIA
MIXED PRECISION TRAINING OF NEURAL NETWORKS Carl Case, Senior Architect, NVIDIA OUTLINE 1. What is mixed precision training with FP16? 2. Considerations and methodology for mixed precision training 3.
More informationDEEP NEURAL NETWORKS AND GPUS. Julie Bernauer
DEEP NEURAL NETWORKS AND GPUS Julie Bernauer GPU Computing GPU Computing Run Computations on GPUs x86 CUDA Framework to Program NVIDIA GPUs A simple sum of two vectors (arrays) in C void vector_add(int
More informationDynamic Routing Between Capsules
Report Explainable Machine Learning Dynamic Routing Between Capsules Author: Michael Dorkenwald Supervisor: Dr. Ullrich Köthe 28. Juni 2018 Inhaltsverzeichnis 1 Introduction 2 2 Motivation 2 3 CapusleNet
More informationStreet Address Image Recognition (CNN) Machine Learning Engineer Nanodegree
Street Address Image Recognition (CNN) Machine Learning Engineer Nanodegree YOONSU PARK Oct 22 2016 Definition Project overview With the big advance in computing power with the enhanced deep learning technique,
More informationTraining Neural Networks with Mixed Precision MICHAEL CARILLI CHRISTIAN SAROFEEN MICHAEL RUBERRY BEN BARSDELL
Training Neural Networks with Mixed Precision MICHAEL CARILLI CHRISTIAN SAROFEEN MICHAEL RUBERRY BEN BARSDELL 1 THIS TALK Using mixed precision and Volta your networks can be: 1. 2-4x faster 2. half the
More informationTwo-Stream Convolutional Networks for Action Recognition in Videos
Two-Stream Convolutional Networks for Action Recognition in Videos Karen Simonyan Andrew Zisserman Cemil Zalluhoğlu Introduction Aim Extend deep Convolution Networks to action recognition in video. Motivation
More informationA performance comparison of Deep Learning frameworks on KNL
A performance comparison of Deep Learning frameworks on KNL R. Zanella, G. Fiameni, M. Rorro Middleware, Data Management - SCAI - CINECA IXPUG Bologna, March 5, 2018 Table of Contents 1. Problem description
More informationFrameworks in Python for Numeric Computation / ML
Frameworks in Python for Numeric Computation / ML Why use a framework? Why not use the built-in data structures? Why not write our own matrix multiplication function? Frameworks are needed not only because
More informationConvolutional Neural Networks and Supervised Learning
Convolutional Neural Networks and Supervised Learning Eilif Solberg August 30, 2018 Outline Convolutional Architectures Convolutional neural networks Training Loss Optimization Regularization Hyperparameter
More informationIST 597 Deep Learning Overfitting and Regularization. Sep. 27, 2018
IST 597 Deep Learning Overfitting and Regularization 1. Overfitting Sep. 27, 2018 Regression model y 1 3 x3 13 2 x2 36x10 import numpy as np import matplotlib.pyplot as plt from sklearn.linear_model import
More informationPouya Kousha Fall 2018 CSE 5194 Prof. DK Panda
Pouya Kousha Fall 2018 CSE 5194 Prof. DK Panda 1 Motivation And Intro Programming Model Spark Data Transformation Model Construction Model Training Model Inference Execution Model Data Parallel Training
More informationEmel: Deep Learning in One Line using Type Inference, Architecture Selection, and Hyperparameter Tuning
Emel: Deep Learning in One Line using Type Inference, Architecture Selection, and Hyperparameter Tuning BARAK OSHRI and NISHITH KHANDWALA We present Emel, a new framework for training baseline supervised
More informationMini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class
Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class Guidelines Submission. Submit a hardcopy of the report containing all the figures and printouts of code in class. For readability
More informationPointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space
PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space Sikai Zhong February 14, 2018 COMPUTER SCIENCE Table of contents 1. PointNet 2. PointNet++ 3. Experiments 1 PointNet Property
More informationQuo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018
Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Outline: Introduction Action classification architectures
More informationImageNet Classification with Deep Convolutional Neural Networks
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky Ilya Sutskever Geoffrey Hinton University of Toronto Canada Paper with same name to appear in NIPS 2012 Main idea Architecture
More informationAnalyzing rtweet data with kerasformula
Analyzing rtweet data with kerasformula Pete Mohanty BARUG March 2018 hosted by Google Overview Now on CRAN, kerasformula package offers a high-level interface for the R interface to Keras. install.packages("kerasformula")
More informationPolytechnic University of Tirana
1 Polytechnic University of Tirana Department of Computer Engineering SIBORA THEODHOR ELINDA KAJO M ECE 2 Computer Vision OCR AND BEYOND THE PRESENTATION IS ORGANISED IN 3 PARTS : 3 Introduction, previous
More informationDeep Model Compression
Deep Model Compression Xin Wang Oct.31.2016 Some of the contents are borrowed from Hinton s and Song s slides. Two papers Distilling the Knowledge in a Neural Network by Geoffrey Hinton et al What s the
More information