MACHINE LEARNING CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS

Similar documents
New approaches to pattern recognition and automated learning

Deep Learning for Computer Vision II

CNN Basics. Chongruo Wu

Keras: Handwritten Digit Recognition using MNIST Dataset

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

ImageNet Classification with Deep Convolutional Neural Networks

Machine Learning 13. week

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

Polytechnic University of Tirana

Deep Learning with Tensorflow AlexNet

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Tutorial on Machine Learning Tools

Object Detection. Part1. Presenter: Dae-Yong

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

Introduction to Deep Learning

Fuzzy Set Theory in Computer Vision: Example 3, Part II

ECE 5470 Classification, Machine Learning, and Neural Network Review

CS 179 Lecture 16. Logistic Regression & Parallel SGD

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

M. Sc. (Artificial Intelligence and Machine Learning)

AttentionNet for Accurate Localization and Detection of Objects. (To appear in ICCV 2015)

Image Classification using Fast Learning Convolutional Neural Networks

Index. Springer Nature Switzerland AG 2019 B. Moons et al., Embedded Deep Learning,

EMBEDDED MACHINE VISION

Artificial Intelligence Introduction Handwriting Recognition Kadir Eren Unal ( ), Jakob Heyder ( )

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Perceptron: This is convolution!

Logical Rhythm - Class 3. August 27, 2018

CS6220: DATA MINING TECHNIQUES

Using Machine Learning for Classification of Cancer Cells

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Efficient Algorithms may not be those we think

Convolutional Neural Networks

Fuzzy Set Theory in Computer Vision: Example 3

ECG782: Multidimensional Digital Signal Processing

Deep Convolutional Neural Networks. Nov. 20th, 2015 Bruce Draper

Deep Learning Basic Lecture - Complex Systems & Artificial Intelligence 2017/18 (VO) Asan Agibetov, PhD.

Transfer Learning. Style Transfer in Deep Learning

Lecture #11: The Perceptron

Dynamic Routing Between Capsules

Deep Learning and Its Applications

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

Study of Residual Networks for Image Recognition

Capsule Networks. Eric Mintun

CSE 559A: Computer Vision

INTRODUCTION TO DEEP LEARNING

MoonRiver: Deep Neural Network in C++

Face Recognition A Deep Learning Approach

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Kaggle Data Science Bowl 2017 Technical Report

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Object detection using Region Proposals (RCNN) Ernest Cheung COMP Presentation

Deep Learning & Neural Networks

Predictive Analytics: Demystifying Current and Emerging Methodologies. Tom Kolde, FCAS, MAAA Linda Brobeck, FCAS, MAAA

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

A THREE LAYERED MODEL TO PERFORM CHARACTER RECOGNITION FOR NOISY IMAGES

Object Detection Lecture Introduction to deep learning (CNN) Idar Dyrdal

Recurrent Neural Network (RNN) Industrial AI Lab.

Keras: Handwritten Digit Recognition using MNIST Dataset

POINT CLOUD DEEP LEARNING

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Introduction to Deep Learning for Facial Understanding Part III: Regional CNNs

Convolution Neural Nets meet

Convolutional Neural Network for Image Classification

Report: Privacy-Preserving Classification on Deep Neural Network

On the Effectiveness of Neural Networks Classifying the MNIST Dataset

CS 229 Final Report: Artistic Style Transfer for Face Portraits

Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets)

Classification of objects from Video Data (Group 30)

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Graph Neural Network. learning algorithm and applications. Shujia Zhang

CS6501: Deep Learning for Visual Recognition. Object Detection I: RCNN, Fast-RCNN, Faster-RCNN

An Exploration of Computer Vision Techniques for Bird Species Classification

Hand Written Digit Recognition Using Tensorflow and Python

Deep Learning. Volker Tresp Summer 2014

DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

Motivation. Problem: With our linear methods, we can train the weights but not the basis functions: Activator Trainable weight. Fixed basis function

Convolutional Neural Network for Facial Expression Recognition

Bayes Risk. Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Discriminative vs Generative Models. Loss functions in classifiers

Lecture 12 Recognition

COMPUTATIONAL INTELLIGENCE

Encoder-Decoder Networks for Semantic Segmentation. Sachin Mehta

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Binary Convolutional Neural Network on RRAM

Smart Content Recognition from Images Using a Mixture of Convolutional Neural Networks *

Classifiers for Recognition Reading: Chapter 22 (skip 22.3)

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research

Mask R-CNN. By Kaiming He, Georgia Gkioxari, Piotr Dollar and Ross Girshick Presented By Aditya Sanghi

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

High Performance Computing

Real-time convolutional networks for sonar image classification in low-power embedded systems

Rich feature hierarchies for accurate object detection and semantic segmentation

Tutorial on Keras CAP ADVANCED COMPUTER VISION SPRING 2018 KISHAN S ATHREY

Data Mining. Neural Networks

The Mathematics Behind Neural Networks

Transcription:

MACHINE LEARNING CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS FRANK ORBEN, TECHNICAL SUPPORT / DEVELOPER IMAGE PROCESSING, STEMMER IMAGING

OUTLINE Introduction Task: Classification Theory Descriptors & feature space MRF/Ridge Regression (CVB Polimago) Conv. Neural Networks (TensorFlow ) Application Advantages & challenges Use cases Classification performance Excursion: Object search Summary SLIDE 2

TASK: CLASSIFICATION Classification assigning a class label to an object Objects image or ROI Features Shape Size Structure Gradient Edges Feature descriptor? Descriptor Classifier Classification SLIDE 3

DESCRIPTORS IN FEATURE SPACE Descriptor parametrical description of an object Feature space mapping of objects based on their descriptors 4 3,5 3 2,5 2 1,5 1 0,5 0 0 0,5 1 1,5 2 Descriptors Feature Space SLIDE 4

FEATURE SPACE OVER & UNDER FITTING Underfitting model is not accurate enough Low consistency High generalization Overfitting model is too accurate High consistency Weak generalization [WikiCommons1] Goal: compromise [WikiCommons2] SLIDE 5

TRAINING & TRAINING SET Training creating classifiers based on a training set A B C Classification Algorithm SVM KNN Regression ANN Classifier Training Set Descriptor Training SLIDE 6

MULTI RESOLUTION FILTER MRF feature extraction Multi scale analysis Convolution on different scale levels Various filters Abstract description Structure Gradients Local intensities Feature vector for every image of a class [WikiCommons3] SLIDE 7

REGULARIZATION& RIDGE REGRESSION Regression analysis Models relationships between variables Model description Regularization Solves ill-posed problems Adds additional information [WikiCommons4] Tikhonov Regularization (Andrey Tikhonov, 1960s) Regularization with weighted matrix Ridge Regression (special case of Tikhonov Reg.) Identity matrix [WikiCommons5] SLIDE 8

REGULARIZATION& RIDGE REGRESSION Training Descriptors > mapped as high-dimensional equation system Classifier calculation, Regularization with Ridge Regression Classifier Regressions predictor for every class pair Classifier: all regression predictors of a training set Classification Convolution of feature descriptor with classifier Scalar for every class pair Assigning class labels SLIDE 9

ARTIFICIAL NEURAL NETWORK - DEEP LEARNING Artificial neural network(1980s) Viable thanks to wide spread of GPGPU Imitation of biological nerve system Neuron/Unit multiple inputs > single output Net of neurons/units [WikiCommons7] Back propagating Hidden Layer Output Layer Output Layer [WikiCommons7/8] SLIDE 10

CNN CONVOLUTION LAYER Convolution Pooling Randomly initialized filters [WikiCommons9] Abstract description (structure, intensity, ) Feature extraction Self-optimizing (back propagating) SLIDE 11

CNN PUTTING IT ALL TOGETHER CNN = convolution stage + fully connected stage Training: repeated iterations for whole training set Result: optimized net after N iterations (epoch) Classification: image passes the net once Assigning class labels? convolution, pooling fully connected SLIDE 12

ADVANTAGES RR/MRF & CNN No pre-processing necessary (image size consistent for training set) Robust features for many use cases Low requirements regarding knowledge of feature selection CNN Self-optimization of feature extraction and classification stage Variable specialization via depth of net Robust for similar objects (suitable parameterization) RR/MRF Variable specialization via MRF & regularization parameters Robust with small training sets SLIDE 13

CHALLENGES General Higher number of classes > lengthy processing (training & classification) Image size is fixed per training set CNN (current parameterization/implementation) Overfitting (counter measures exist) Comprehensively large sets Dedicates hardware (GPU) for reasonable training times RR/MRF (current parameterization/implementation) Maximum number of classes is restricted Negligible acceleration through GPU SLIDE 14

USE CASES Identify/distinguish objects Quality control in manufacturing Distinguish good from bad parts Statistical analysis Counting people Counting vehicles Automated data processing Text recognition Batch processing Sorting SLIDE 15

CLASSIFICATION PERFORMANCE Vehicle Wafer OCR MNIST 240x240 px 8bit RGB 1,7k images 6 classes 63x68 px 8bit mono 14k images 10 classes 28x28 px binary 60k images 10 classes RR VEHICLE RR WAFER RR MNIST CNN VEHICLE CNN WAFER CNN MNIST Top-1-Error 5.18 2.02 1.89 5.05 5.75 1.95 Train. Time 0.12 min 8 min 14 h 7 min 23 min 20 min 62 min Class. Time 0.048 ms 0.045 ms 0.007 ms 1.722 ms 1.288 ms 1.791 ms SLIDE 16

EXCURSION: OBJECT SEARCH target Object search > reg. pred. for every degree of freedom (example, translation only) SLIDE 17

SUMMARY Traditional approach ridge regression (with multi resolution filter) Viable thanks to current technology CNN Many application areas Large number of classes > longer training times Hardware requirements Size of training set > constrains depend on use case SLIDE 18

THANK YOU FOR YOUR ATTENTION Your contact Frank Orben STEMMER IMAGING GmbH +49 89 80902-747 f.orben@stemmer-imaging.de www.stemmer-imaging.de Copyright STEMMER IMAGING. All texts, images, graphs, tone, video and animation files as well as their arrangements are subject to copyright law and other laws for the protection of intellectual property. They may not be copied or changed for any commercial use or for the purpose of being passed on nor used on other websites. Some of the pages of the STEMMER IMAGING presentation also contain images that are subject to the copyright belonging to those persons who have made them available

BIBLIOGRAPHY [WikiCommons1] https://commons.wikimedia.org/wiki/file:traintest.svg Web, Oct. 11th 2017 [WikiCommons2] https://commons.wikimedia.org/wiki/file:overfitting.svg Web, Oct. 11th 2017 [WikiCommons3] https://commons.wikimedia.org/wiki/file:image_pyramid.svg Web, Oct. 11th 2017 [WikiCommons4] https://commons.wikimedia.org/wiki/file:normdist_regression.png Web, Oct. 11th 2017 [WikiCommons5] https://commons.wikimedia.org/wiki/file:regularization.svg Web, Oct. 11th 2017 [WikiCommons6] https://commons.wikimedia.org/wiki/file:artificialneuronmodel_english.png Web, Oct. 11th 2017 [WikiCommons7] https://commons.wikimedia.org/wiki/file:single-layer_neural_network-vector-blank.svg Web, Oct. 11th 2017 [WikiCommons8] https://commons.wikimedia.org/wiki/file:multi-layer_neural_network-vector-blank.svg Web, Oct. 11th 2017 [WikiCommons9] https://commons.wikimedia.org/wiki/file:typical_cnn.png Web, Oct. 11th 2017 [LeCun] http://www.iro.umontreal.ca/~bengioy/talks/dl-tutorial-nips2015.pdf, Web, Oct. 11th 201 MACHINE VISION CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS SLIDE 20

APPENDIX MACHINE VISION CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS SLIDE 21

APPENDIX CNN FEATURE MAPS.[LeCun] MACHINE VISION CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS SLIDE 22

APPENDIX - TIKHONOV REGULARIZATION Know matrix and vector, looking for vector such that: Approach: ordinary least squares, minimize To give preference to a particular solution, introduce regularization term Tikhonov matrix Regularization improves conditioning, enabling a direct numerical solution MACHINE VISION CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS SLIDE 23