Fuzzy Set Theory in Computer Vision: Example 3

Similar documents
Fuzzy Set Theory in Computer Vision: Example 3, Part II

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn

Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets)

Inception Network Overview. David White CS793

Know your data - many types of networks

Convolutional Neural Networks

Spatial Localization and Detection. Lecture 8-1

Inception and Residual Networks. Hantao Zhang. Deep Learning with Python.

Fei-Fei Li & Justin Johnson & Serena Yeung

OBJECT DETECTION HYUNG IL KOO

Deep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia

Deep Learning with Tensorflow AlexNet

Deep Residual Learning

Smart Parking System using Deep Learning. Sheece Gardezi Supervised By: Anoop Cherian Peter Strazdins

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Deep Learning for Computer Vision with MATLAB By Jon Cherrie

CNNS FROM THE BASICS TO RECENT ADVANCES. Dmytro Mishkin Center for Machine Perception Czech Technical University in Prague

Supplementary material for Analyzing Filters Toward Efficient ConvNet

INTRODUCTION TO DEEP LEARNING

CNN Basics. Chongruo Wu

Computer Vision Lecture 16

Computer Vision Lecture 16

Encoder-Decoder Networks for Semantic Segmentation. Sachin Mehta

Convolutional Neural Networks: Applications and a short timeline. 7th Deep Learning Meetup Kornel Kis Vienna,

HENet: A Highly Efficient Convolutional Neural. Networks Optimized for Accuracy, Speed and Storage

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Convolutional Neural Networks + Neural Style Transfer. Justin Johnson 2/1/2017

Lecture: Deep Convolutional Neural Networks

Como funciona o Deep Learning

Real Time Monitoring of CCTV Camera Images Using Object Detectors and Scene Classification for Retail and Surveillance Applications

Fully Convolutional Network for Depth Estimation and Semantic Segmentation

Return of the Devil in the Details: Delving Deep into Convolutional Nets

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 11. Sinan Kalkan

Characterization and Benchmarking of Deep Learning. Natalia Vassilieva, PhD Sr. Research Manager

Elastic Neural Networks for Classification

Computer Vision Lecture 16

DL Tutorial. Xudong Cao

Learning Deep Features for Visual Recognition

Learning Deep Representations for Visual Recognition

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Generative Modeling with Convolutional Neural Networks. Denis Dus Data Scientist at InData Labs

POINT CLOUD DEEP LEARNING

Object Detection. Part1. Presenter: Dae-Yong

A performance comparison of Deep Learning frameworks on KNL

Deep Learning for Computer Vision II

Yelp Restaurant Photo Classification

Deep Learning and Its Applications

Advanced Video Analysis & Imaging

Deep learning for object detection. Slides from Svetlana Lazebnik and many others

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY

CS 2750: Machine Learning. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh April 13, 2016

Channel Locality Block: A Variant of Squeeze-and-Excitation

A HPX backend for TensorFlow

Structured Prediction using Convolutional Neural Networks

ECE 5470 Classification, Machine Learning, and Neural Network Review

Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018

Residual Networks And Attention Models. cs273b Recitation 11/11/2016. Anna Shcherbina

DeCAF: a Deep Convolutional Activation Feature for Generic Visual Recognition

MCMOT: Multi-Class Multi-Object Tracking using Changing Point Detection

Defense Data Generation in Distributed Deep Learning System Se-Yoon Oh / ADD-IDAR

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal

From Keyboards to Neural Networks 从键盘到神经网络

Convolu'onal Neural Networks

Content-Based Image Recovery

CSE 559A: Computer Vision

arxiv: v1 [cs.cv] 14 Jul 2017

Training Deep Neural Networks (in parallel)

Yiqi Yan. May 10, 2017

Study of Residual Networks for Image Recognition

Direct Multi-Scale Dual-Stream Network for Pedestrian Detection Sang-Il Jung and Ki-Sang Hong Image Information Processing Lab.

High Performance Computing

FUSION MODEL BASED ON CONVOLUTIONAL NEURAL NETWORKS WITH TWO FEATURES FOR ACOUSTIC SCENE CLASSIFICATION

Using Machine Learning for Classification of Cancer Cells

arxiv: v1 [cs.cv] 4 Dec 2014

A Novel Weight-Shared Multi-Stage Network Architecture of CNNs for Scale Invariance

Face Recognition A Deep Learning Approach

Mind the regularized GAP, for human action classification and semi-supervised localization based on visual saliency.

Object detection with CNNs

Cisco UCS C480 ML M5 Rack Server Performance Characterization

Rich feature hierarchies for accurate object detection and semantic segmentation

Kaggle Data Science Bowl 2017 Technical Report

Deep Learning Requirements for Autonomous Vehicles

Tiny ImageNet Challenge

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong

Convolutional Neural Networks

Deconvolutions in Convolutional Neural Networks

Restricted Connectivity in Deep Neural Networks

Exploration of the Effect of Residual Connection on top of SqueezeNet A Combination study of Inception Model and Bypass Layers

Self Driving. DNN * * Reinforcement * Unsupervised *

Yield Estimation using faster R-CNN

How to Estimate the Energy Consumption of Deep Neural Networks

HPE Deep Learning Cookbook: Recipes to Run Deep Learning Workloads. Natalia Vassilieva, Sergey Serebryakov

Binary Convolutional Neural Network on RRAM

Computer Vision. From traditional approaches to deep neural networks. Stanislav Frolov München,

DEPARTMENT INFORMATION ENGINEERING AND COMPUTER SCIENCE MASTER DEGREE IN TELECOMMUNICATIONS ENGINEERING THESIS. Title of the thesis

DEEP NEURAL NETWORKS FOR OBJECT DETECTION

Recurrent Convolutional Neural Networks for Scene Labeling

Deep Learning for Object detection & localization

Transcription:

Fuzzy Set Theory in Computer Vision: Example 3 Derek T. Anderson and James M. Keller FUZZ-IEEE, July 2017

Overview Purpose of these slides are to make you aware of a few of the different CNN architectures out there

Overview Purpose of these slides are to make you aware of a few of the different CNN architectures out there NOT comprehensive

Overview Purpose of these slides are to make you aware of a few of the different CNN architectures out there NOT comprehensive Many folks download these networks and use them

Overview Purpose of these slides are to make you aware of a few of the different CNN architectures out there NOT comprehensive Many folks download these networks and use them MatConvNet (and other libraries) support such functionality

GoogleNet Going deeper with convolutions, published in 2014

GoogleNet Going deeper with convolutions, published in 2014 Input data: ILSVRC 2014 challenge data set, it contains 1,000 categories. 1.2 million images for training, 50,000 for validation and 100,000 for testing

GoogleNet Going deeper with convolutions, published in 2014 Input data: ILSVRC 2014 challenge data set, it contains 1,000 categories. 1.2 million images for training, 50,000 for validation and 100,000 for testing Input image is 224x224x3

GoogleNet Going deeper with convolutions, published in 2014 Input data: ILSVRC 2014 challenge data set, it contains 1,000 categories. 1.2 million images for training, 50,000 for validation and 100,000 for testing Input image is 224x224x3 Zero mean

GoogleNet: inception module The goal of the inception module is to act as a multi-level feature extractor by computing 11, 33, and 55 convolutions within the same module of the network the output of these filters are then stacked along the channel dimension and before being fed into the next layer in the network

GoogleNet: inception module The goal of the inception module is to act as a multi-level feature extractor by computing 11, 33, and 55 convolutions within the same module of the network the output of these filters are then stacked along the channel dimension and before being fed into the next layer in the network The original incarnation of this architecture was called GoogLeNet, but subsequent manifestations have simply been called Inception vn where N refers to the version number put out by Google

GoogleNet: inception module The goal of the inception module is to act as a multi-level feature extractor by computing 11, 33, and 55 convolutions within the same module of the network the output of these filters are then stacked along the channel dimension and before being fed into the next layer in the network The original incarnation of this architecture was called GoogLeNet, but subsequent manifestations have simply been called Inception vn where N refers to the version number put out by Google

GoogleNet

AlexNet ImageNet Classification with Deep Convolutional Neural Networks, published on 2012

AlexNet ImageNet Classification with Deep Convolutional Neural Networks, published on 2012 ILSVRC-2010 uses subset of ImageNet data set with roughly 1,000 RGB images in each of 1,000 categories. Roughly 1.2 million training images, 50,000 validation images, and 150,000 testing images. Test set labels are available

AlexNet ImageNet Classification with Deep Convolutional Neural Networks, published on 2012 ILSVRC-2010 uses subset of ImageNet data set with roughly 1,000 RGB images in each of 1,000 categories. Roughly 1.2 million training images, 50,000 validation images, and 150,000 testing images. Test set labels are available The images were down sampled to 256x256 pixels

AlexNet ImageNet Classification with Deep Convolutional Neural Networks, published on 2012 ILSVRC-2010 uses subset of ImageNet data set with roughly 1,000 RGB images in each of 1,000 categories. Roughly 1.2 million training images, 50,000 validation images, and 150,000 testing images. Test set labels are available The images were down sampled to 256x256 pixels Zero mean

AlexNet ImageNet Classification with Deep Convolutional Neural Networks, published on 2012 ILSVRC-2010 uses subset of ImageNet data set with roughly 1,000 RGB images in each of 1,000 categories. Roughly 1.2 million training images, 50,000 validation images, and 150,000 testing images. Test set labels are available The images were down sampled to 256x256 pixels Zero mean

VGG network VGG network architecture was introduced by Simonyan and Zisserman in their 2014 paper, Very Deep Convolutional Networks for Large Scale Image Recognition

VGG network VGG network architecture was introduced by Simonyan and Zisserman in their 2014 paper, Very Deep Convolutional Networks for Large Scale Image Recognition Network is characterized by its simplicity, using only 33 convolutional layers stacked on top of each other in increasing depth. Reducing volume size is handled by max pooling. Two fully-connected layers, each with 4,096 nodes are then followed by a softmax classifier

VGG network VGG network architecture was introduced by Simonyan and Zisserman in their 2014 paper, Very Deep Convolutional Networks for Large Scale Image Recognition Network is characterized by its simplicity, using only 33 convolutional layers stacked on top of each other in increasing depth. Reducing volume size is handled by max pooling. Two fully-connected layers, each with 4,096 nodes are then followed by a softmax classifier The 16 and 19 stand for the number of weight layers in the network

VGG network

VGG network

ResNet He et al. in their 2015 paper, Deep Residual Learning for Image Recognition

ResNet He et al. in their 2015 paper, Deep Residual Learning for Image Recognition ResNet architecture has become a seminal work, demonstrating that extremely deep networks can be trained using standard SGD (and a reasonable initialization function) through the use of residual modules

ResNet He et al. in their 2015 paper, Deep Residual Learning for Image Recognition ResNet architecture has become a seminal work, demonstrating that extremely deep networks can be trained using standard SGD (and a reasonable initialization function) through the use of residual modules Unlike traditional sequential network architectures (AlexNet, OverFeat and VGG), ResNet relies on micro-architecture modules ( network-in-network architectures )

ResNet He et al. in their 2015 paper, Deep Residual Learning for Image Recognition ResNet architecture has become a seminal work, demonstrating that extremely deep networks can be trained using standard SGD (and a reasonable initialization function) through the use of residual modules Unlike traditional sequential network architectures (AlexNet, OverFeat and VGG), ResNet relies on micro-architecture modules ( network-in-network architectures ) Micro-architecture refers to the set of building blocks used to construct the network

ResNet He et al. in their 2015 paper, Deep Residual Learning for Image Recognition ResNet architecture has become a seminal work, demonstrating that extremely deep networks can be trained using standard SGD (and a reasonable initialization function) through the use of residual modules Unlike traditional sequential network architectures (AlexNet, OverFeat and VGG), ResNet relies on micro-architecture modules ( network-in-network architectures ) Micro-architecture refers to the set of building blocks used to construct the network A collection of micro-architecture building blocks (along with your standard CONV, POOL, etc. layers) leads to the macro-architecture (i.e,. the end network itself)

ResNet Further accuracy can be obtained by updating the residual module to use identity mappings, as demonstrated in their 2016 followup publication, Identity Mappings in Deep Residual Networks

ResNet Further accuracy can be obtained by updating the residual module to use identity mappings, as demonstrated in their 2016 followup publication, Identity Mappings in Deep Residual Networks Even though ResNet is much deeper than VGG16 and VGG19, the model size is actually substantially smaller due to the usage of global average pooling rather than fully-connected layers this reduces the model size down to 102MB for ResNet50

Summary... Field is changing daily almost

Summary... Field is changing daily almost NOT a solved problem

Summary... Field is changing daily almost NOT a solved problem DL has a high cost for entry

Summary... Field is changing daily almost NOT a solved problem DL has a high cost for entry Start reading...