ECE 5470 Classification, Machine Learning, and Neural Network Review

Similar documents
Keras: Handwritten Digit Recognition using MNIST Dataset

Keras: Handwritten Digit Recognition using MNIST Dataset

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

DEEP LEARNING IN PYTHON. Creating a keras model

Tutorial on Machine Learning Tools

INTRODUCTION TO DEEP LEARNING

Evaluating Classifiers

Evaluating Classifiers

CIS680: Vision & Learning Assignment 2.b: RPN, Faster R-CNN and Mask R-CNN Due: Nov. 21, 2018 at 11:59 pm

Advanced Machine Learning

Deep Learning and Its Applications

CIS581: Computer Vision and Computational Photography Project 4, Part B: Convolutional Neural Networks (CNNs) Due: Dec.11, 2017 at 11:59 pm

List of Exercises: Data Mining 1 December 12th, 2015

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,

Inception and Residual Networks. Hantao Zhang. Deep Learning with Python.

Residual Networks And Attention Models. cs273b Recitation 11/11/2016. Anna Shcherbina

Deep Learning with Tensorflow AlexNet

Fuzzy Set Theory in Computer Vision: Example 3

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

Polytechnic University of Tirana

Know your data - many types of networks

Predict the Likelihood of Responding to Direct Mail Campaign in Consumer Lending Industry

Kaggle Data Science Bowl 2017 Technical Report

Evaluation Metrics. (Classifiers) CS229 Section Anand Avati

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Machine Learning 13. week

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Study of Residual Networks for Image Recognition

Deep Learning Explained Module 4: Convolution Neural Networks (CNN or Conv Nets)

Practical 8: Neural networks

Deep Learning for Computer Vision II

Weka ( )

CS 523: Multimedia Systems

LSTM: An Image Classification Model Based on Fashion-MNIST Dataset

Artificial Intelligence Introduction Handwriting Recognition Kadir Eren Unal ( ), Jakob Heyder ( )

The exam is closed book, closed notes except your one-page cheat sheet.

Deep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?

Pattern recognition (4)

Overall Description. Goal: to improve spatial invariance to the input data. Translation, Rotation, Scale, Clutter, Elastic

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

The Mathematics Behind Neural Networks

DEEP LEARNING IN PYTHON. Introduction to deep learning

Probabilistic Classifiers DWML, /27

Dynamic Routing Between Capsules

CS249: ADVANCED DATA MINING

Classification of objects from Video Data (Group 30)

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation

Introduction to Deep Learning

Programming Exercise 3: Multi-class Classification and Neural Networks

Convolutional Neural Networks: Applications and a short timeline. 7th Deep Learning Meetup Kornel Kis Vienna,

Data Mining Classification: Alternative Techniques. Imbalanced Class Problem

CS 2750: Machine Learning. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh April 13, 2016

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR

Programming Exercise 4: Neural Networks Learning

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research

CS145: INTRODUCTION TO DATA MINING

Multi-Task Self-Supervised Visual Learning

Classification Part 4

Mini-project 2 CMPSCI 689 Spring 2015 Due: Tuesday, April 07, in class

Deep Learning in NLP. Horacio Rodríguez. AHLT Deep Learning 2 1

Bayes Risk. Classifiers for Recognition Reading: Chapter 22 (skip 22.3) Discriminative vs Generative Models. Loss functions in classifiers

Classifying Depositional Environments in Satellite Images

Convolutional Neural Networks

CNN Basics. Chongruo Wu

Classifiers for Recognition Reading: Chapter 22 (skip 22.3)

CREDIT RISK MODELING IN R. Finding the right cut-off: the strategy curve

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Visual object classification by sparse convolutional neural networks

Evaluating Machine-Learning Methods. Goals for the lecture

An Introduction to NNs using Keras

Convolutional Neural Networks. CSC 4510/9010 Andrew Keenan

Mask R-CNN. By Kaiming He, Georgia Gkioxari, Piotr Dollar and Ross Girshick Presented By Aditya Sanghi

Convolution Neural Networks for Chinese Handwriting Recognition

Deep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 11. Sinan Kalkan

MACHINE LEARNING CLASSIFIERS ADVANTAGES AND CHALLENGES OF SELECTED METHODS

CS4491/CS 7265 BIG DATA ANALYTICS

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Ensemble methods in machine learning. Example. Neural networks. Neural networks

M. Sc. (Artificial Intelligence and Machine Learning)

Handwritten Hindi Numerals Recognition System

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn

Inception Network Overview. David White CS793

Practical Deep Learning

CS6220: DATA MINING TECHNIQUES

The exam is closed book, closed notes except your one-page (two-sided) cheat sheet.

CHAPTER 6 DETECTION OF MASS USING NOVEL SEGMENTATION, GLCM AND NEURAL NETWORKS

Advanced Video Content Analysis and Video Compression (5LSH0), Module 8B

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University

Convolutional Neural Network based Medical Imaging Segmentation: Recent Progress and Challenges. Jiaxing Tan

Perceptron: This is convolution!

Machine Learning for. Artem Lind & Aleskandr Tkachenko

SSD: Single Shot MultiBox Detector. Author: Wei Liu et al. Presenter: Siyu Jiang

Detecting Thoracic Diseases from Chest X-Ray Images Binit Topiwala, Mariam Alawadi, Hari Prasad { topbinit, malawadi, hprasad

Machine Learning for Physicists Lecture 6. Summer 2017 University of Erlangen-Nuremberg Florian Marquardt

Transcription:

ECE 5470 Classification, Machine Learning, and Neural Network Review Due December 1. Solution set Instructions: These questions are to be answered on this document which should be submitted to blackboard in pdf format. This document is made available in both word (.docx) and pdf formats. The answers may be added to the document in many ways including handwritten and scanned. Converting the final answers to a pdf format for submission is the responsibility of the student. For questions that ask for an explanation or definition it should be able provide a good response in one to three sentences in the space provided following the questions. For numerical questions shoe your working. 1. Suppose your input is 16x16 color (RGB) image, and you use a 3x3 convolutional layer with 30 neurons. What are the total number of weights in the hidden layer? Every neuron generates one pixel value in the output volume, it calculates the convolution value for one pixel point, so one neuron has 3x3 = 9 weights + a bias. The total weights in the hidden layer is 30x(9+1) = 300. 2. You have an input volume that is 63x63x16 and convolve it with 3 filters that are each 7x7, and stride of 1. You want the ( same convolution setting in TensorFlow). How large is the padding? Zero pad each input feature map (each depth of the input volume) by 3 ((filter size - 1)/2) pixels over all four directions. (Number of padding pixels: (63x3x4 + 4x3x3) x 16 = 12672) 3. Which of the following statement is true about k-nn algorithm? 1. k-nn performs much better if all of the data have the same scale 2. k-nn works well with a small number of input variables (p), but struggles when the number of inputs is very large 3. k-nn makes no assumptions about the functional form of the problem being solved a) 1 and 2 b) 1 and 3 c) Only 1 d) All of the above f) None of the above 1

4. A logistic regression classifier, which identifies an object as one of 4 classes (a-d), provides individual probability estimates for an image of a=0.4, b=0.2, c=0.9, and d=0.8 a. Which class is the object identified as? C This was not the best formed question; better wording would have been: If we use this classifier with a 1-out-of-n code output (i.e., as a multinomial logistic regression) b. what is the probability of correct classification? # For the individual outputs we have pi,j = - (,* #$ % &' (,* Thus e, (,* = #.- (,* e, (,/= 0.666, e, (,0= 0.25, e, (,1= 9, e, (,2 = 4 then the class probability is given by the softmax function Thus, the probability of correct classification is p(correct) = 9/(0.666 + 0.25 + 9 + 4). = 0.647 5. What is data augmentation and how does it improve classifier performance. The concept that the effective size of the training set may be increased by augmenting it with distorted or modified version of the original training images to improve classifier training (reduce overfitting) Data augmentation may add value to base data by adding information derived from internal and external sources (for instance, we could enlarge the image dataset by rotating, translating, shrinking original training images). It may make the dataset more noisy with the goal of achieving a more robust model (prevent overfitting). 6. Below is a CNN model specified in Keras input_shape = (32, 128,1) model = Sequential() model.add(conv2d(32, kernel_size=(3, 3), activation='relu', input_shape=input_shape)) model.add(conv2d(64, (3, 3), activation='relu')) model.add(maxpooling2d(pool_size=(2, 2))) model.add(dropout(0.25)) model.add(flatten()) model.add(dense(128, activation='relu')) model.add(dropout(0.5)) model.add(dense(num_classes, activation='softmax')) model.compile(loss=keras.losses.categorical_crossentropy, optimizer=keras.optimizers.adadelta(), metrics=['accuracy']) 2

Draw a block diagram of the model and clearly mark which block have trainable weights how they are organized; e.g., 12 x 16 x 3. See solution on next page Keras defaults: use_bias=true, padding= valid (none) 3

7. In a two-class task the results from logistic regression classifier are as follows: (sorted in order of probability of correct class) Index Confidence positive Correct class 1.97 + 2.93 + 3.72-4.51-5.45 + 6.33-7.25-8.20 - a. Carefully sketch or plot the ROC graph. Indicate the value of each point on the graph. Show working. b. Draw and label the confusion matrix. Supposed the threshold is 0.5: predicted class Positive Negative actual class Positive Negative (TP) 2 (FP) 2 (FN) 1 (TN) 3 4

c. Extra credit: carefully draw the Precision Recall Curve (PRC) 8. For the images used in in the Lab 7 tutorial, how are they preprocessed before being input to the classifier? List each preprocessing step and its purpose: (MNIST preprocessing) 1. Rescale binary images into a resampled 20 x 20 bounding box. Size normalization 2. Move the COM to the center of a 28 x 28 image. Location normalization 3. Rescale the data to have values between 0 and 1. Pixel scale normalization 4. Standardize the training and testing data to have mean 0 and standard deviation 1. --- Pixel scale normalization (makes step 3 unnecessary) 9. How is a ResNet convolutional layer different than a traditional CNN layer? From the paper Deep Residual Learning for Image Recognition, ResNet implements shortcut connections to perform identity mapping, and the identity mapping outputs are added to the outputs of the stacked layers. A traditional CNN simply stacks layers. That is, after a nonlinear stage the inputs are added to outputs before an activation function. 10. What is the advantage of a ResNet layer compared to a traditional CNN layer? As stated in the paper Deep Residual Learning for Image Recognition 1. ResNet is easier to optimize compared to traditional CNN layers. 2. ResNet gains accuracy from greatly increased depth, while traditional CNNs encounter degradation when networks depth increases. One motivation for skipping over layers in CNNs is to avoid the problem of vanishing gradients by reusing activations from a previous layer until the layer next to the current one has learned its weights. This may facilitate the more rapid practical training of deeper networks. 5