SELF SUPERVISED DEEP REPRESENTATION LEARNING FOR FINE-GRAINED BODY PART RECOGNITION

Similar documents
A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

Kaggle Data Science Bowl 2017 Technical Report

Real-time Object Detection CS 229 Course Project

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Content-Based Image Recovery

An Exploration of Computer Vision Techniques for Bird Species Classification

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

REGION AVERAGE POOLING FOR CONTEXT-AWARE OBJECT DETECTION

Computer Vision Lecture 16

Deep Learning with Tensorflow AlexNet

Cost-alleviative Learning for Deep Convolutional Neural Network-based Facial Part Labeling

Multi-Glance Attention Models For Image Classification

Deep Learning for Computer Vision II

Structured Prediction using Convolutional Neural Networks

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Volumetric and Multi-View CNNs for Object Classification on 3D Data Supplementary Material

Computer Vision Lecture 16

Know your data - many types of networks

Lecture 7: Semantic Segmentation

arxiv: v1 [cs.cv] 31 Mar 2016

CEA LIST s participation to the Scalable Concept Image Annotation task of ImageCLEF 2015

HENet: A Highly Efficient Convolutional Neural. Networks Optimized for Accuracy, Speed and Storage

Part Localization by Exploiting Deep Convolutional Networks

Human Action Recognition Using CNN and BoW Methods Stanford University CS229 Machine Learning Spring 2016

Computer Vision Lecture 16

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Semantic Segmentation

Iterative Multi-domain Regularized Deep Learning for Anatomical Structure Detection and Segmentation from Ultrasound Images

Channel Locality Block: A Variant of Squeeze-and-Excitation

Spatial Localization and Detection. Lecture 8-1

Unsupervised Deep Learning. James Hays slides from Carl Doersch and Richard Zhang

Comparison of Fine-tuning and Extension Strategies for Deep Convolutional Neural Networks

In Defense of Fully Connected Layers in Visual Representation Transfer

Detecting Bone Lesions in Multiple Myeloma Patients using Transfer Learning

Lung Nodule Detection using 3D Convolutional Neural Networks Trained on Weakly Labeled Data

Apparel Classifier and Recommender using Deep Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

ECCV Presented by: Boris Ivanovic and Yolanda Wang CS 331B - November 16, 2016

arxiv: v1 [cs.cv] 20 Dec 2016

Multi-Label Whole Heart Segmentation Using CNNs and Anatomical Label Configurations

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong

A Deep Learning Framework for Authorship Classification of Paintings

Fully Convolutional Networks for Semantic Segmentation

Object Detection Based on Deep Learning

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Classification of objects from Video Data (Group 30)

Regionlet Object Detector with Hand-crafted and CNN Feature

ImageNet Classification with Deep Convolutional Neural Networks

Recurrent Neural Networks and Transfer Learning for Action Recognition

Deconvolutions in Convolutional Neural Networks

Ryerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro

International Journal of Computer Engineering and Applications, Volume XII, Special Issue, September 18,

Deep learning for object detection. Slides from Svetlana Lazebnik and many others

Fine-tuning Pre-trained Large Scaled ImageNet model on smaller dataset for Detection task

Convolutional Neural Networks

Hide-and-Seek: Forcing a network to be Meticulous for Weakly-supervised Object and Action Localization

arxiv: v1 [cs.cv] 6 Jul 2016

Multi-Task Self-Supervised Visual Learning

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Perceptron: This is convolution!

Encoder-Decoder Networks for Semantic Segmentation. Sachin Mehta

Rich feature hierarchies for accurate object detection and semantic segmentation

3D model classification using convolutional neural network

MULTI-SCALE OBJECT DETECTION WITH FEATURE FUSION AND REGION OBJECTNESS NETWORK. Wenjie Guan, YueXian Zou*, Xiaoqun Zhou

3D Densely Convolutional Networks for Volumetric Segmentation. Toan Duc Bui, Jitae Shin, and Taesup Moon

CAP 6412 Advanced Computer Vision

arxiv: v1 [cs.cv] 26 Jun 2017

Where s Waldo? A Deep Learning approach to Template Matching

Rich feature hierarchies for accurate object detection and semant

Yelp Restaurant Photo Classification

Transfer Learning. Style Transfer in Deep Learning

arxiv: v1 [cs.cv] 11 Apr 2018

Machine Learning for Medical Image Analysis. A. Criminisi

Deep Neural Networks:

Real-Time Document Image Classification using Deep CNN and Extreme Learning Machines

CS 1674: Intro to Computer Vision. Neural Networks. Prof. Adriana Kovashka University of Pittsburgh November 16, 2016

Detection and Identification of Lung Tissue Pattern in Interstitial Lung Diseases using Convolutional Neural Network

FACIAL POINT DETECTION BASED ON A CONVOLUTIONAL NEURAL NETWORK WITH OPTIMAL MINI-BATCH PROCEDURE. Chubu University 1200, Matsumoto-cho, Kasugai, AICHI

Rich feature hierarchies for accurate object detection and semantic segmentation

Bilinear Models for Fine-Grained Visual Recognition

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

Supplementary material for Analyzing Filters Toward Efficient ConvNet

arxiv: v2 [cs.cv] 23 May 2016

Self-supervised Multi-level Face Model Learning for Monocular Reconstruction at over 250 Hz Supplemental Material

DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution and Fully Connected CRFs

Mimicking Very Efficient Network for Object Detection

SIIM 2017 Scientific Session Analytics & Deep Learning Part 2 Friday, June 2 8:00 am 9:30 am

Pull the Plug? Predicting If Computers or Humans Should Segment Images Supplementary Material

Deep Face Recognition. Nathan Sun

Smart Content Recognition from Images Using a Mixture of Convolutional Neural Networks *

Microscopy Cell Counting with Fully Convolutional Regression Networks

Temporal HeartNet: Towards Human-Level Automatic Analysis of Fetal Cardiac Screening Video

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Intro to Deep Learning. Slides Credit: Andrej Karapathy, Derek Hoiem, Marc Aurelio, Yann LeCunn

arxiv: v3 [cs.cv] 2 Jun 2017

Study of Residual Networks for Image Recognition

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601

Facial Expression Classification with Random Filters Feature Extraction

Latest development in image feature representation and extraction

Project 3 Q&A. Jonathan Krause

Transcription:

SELF SUPERVISED DEEP REPRESENTATION LEARNING FOR FINE-GRAINED BODY PART RECOGNITION Pengyue Zhang Fusheng Wang Yefeng Zheng Medical Imaging Technologies, Siemens Medical Solutions USA Inc., Princeton, NJ 08540, USA Stony Brook University, Stony Brook, NY 11794, USA ABSTRACT Difficulty on collecting annotated medical images leads to lack of enough supervision and makes discrimination tasks challenging. However, raw data, e.g., spatial context information from 3D CT images, even without annotation, may contain rich useful information. In this paper, we exploit spatial context information as a source of supervision to solve discrimination tasks for fine-grained body part recognition with conventional 3D CT and MR volumes. The proposed pipeline consists of two steps: 1) pre-train a convolutional network for an auxiliary task of 2D slices ordering in a self-supervised manner; 2) transfer and fine-tune the pre-trained network for fine-grained body part recognition. Without any use of human annotation in the first stage, the pre-trained network can still outperform CNN trained from scratch on CT as well as M- R data. Moreover, by comparing with pre-trained CNN from ImageNet, we discover that the distance between source and target tasks plays a crucial role in transfer learning. Our experiments demonstrate that our approach can achieve high accuracy with a slice location estimation error of only a few s- lices on CT and MR data. To the best of our knowledge, our work is the first attempt studying the problem of robust body part recognition at a continuous level. Index Terms Self Supervised Learning, Slice Ordering, Body Part Recognition 1. INTRODUCTION Deep learning techniques have received much attention in last decade. With the increasing computing power thanks to modern GPUs and large labeled datasets such as ImageNet and PASCAL VOC, deep architectures such as convolutional neural networks (CNNs) have improved the state-of-the-art of many computer vision problems, such as image categorization, object detection, image quality assessment and so on [1]. Recently, many efforts are exploited to apply deep learning techniques to medical imaging tasks with promising results [2]. Pengyue Zhang contributed to this work while he was an intern at Siemens. However, although CNNs have achieved impressive progress, the problem becomes much more complicated in medical domain. For CNNs, a large labeled image set is required for satisfying network training. However, collecting large-scale medical data and annotations is time-consuming and requires expertise from domain experts such as radiologists, which makes training a network from scratch infeasible. One natural intuition is to learn the network in an unsupervised manner. However, until now unsupervised learning methods have limited performance on learning meaningful representations for a discrimination task. In the literature, one successful solution to alleviate the lack of annotated training samples is to pre-train a network on large-scale natural image datasets (e.g., ImageNet) and then fine-tune the network parameters for specific tasks [3]. Such knowledge transfer based approach is proved to be not only feasible but also superior than training a CNN from scratch in terms of accuracy. N- evertheless, though most natural images and medical images share many low-level features, they still considerably differ in object-level structures. Thus transfer learning from natural image data to medical applications may bring substantial bias which can lead to limited accuracy. For unsupervised learning in natural language processing domain, the relationship between neighboring words or context, is widely used as a cost-effective yet effective source of supervision [4]. By learning vector representation of words and modeling their similarity and co-occurrence probability, context of a query word or sentence can be automatically predicted. In 2D natural images, spatial relationship between neighboring image patches is also proved to contain useful information for object detection and visual data mining [5, 6]. With one additional spatial dimension, a typical 3D CT or M- R volume contains much richer context information. Within a 3D volume, the transversal 2D slices can be easily indexed and the order of slices can be acquired for free as natural indicators of spatial positions. Following this idea, we are able to transform the unsupervised learning problem to a selfsupervised one, without the need of any manual efforts on annotation. CNN-based methods have reported promising results and become the state-of-the-art in body part recognition [7, 8]. However, previous CNN-based body part recognition remains

at a coarse level, while real-world applications require more precise recognition. In [7] and [8], the human body is classified into 5 or 12 parts covering different body ranges. However, human body is a coherent and continuous whole object instead of several unrelated object classes. Two neighboring slices across body part border are likely to have even more similar shapes and structures than two distant slices from the same body part, thus a hard division of the body parts is inappropriate. Once trained, the pre-defined body parts in [7, 8] are fixed, which prohibits re-definition of parts or splitting a large part into several fine-grained small parts. In this work we estimate the height (i.e., distance to the toe) of a transversal slice in a normalized body model. With a continuous output, our approach is more flexible and achieves a much finer level recognition compared to the approach based on several disjoint body parts. Moreover, both [7] and [8] consider only the body parts from nose to kidney instead of the full body, which limits the generalization ability to other body parts. In this paper, we provide an end-to-end convolutional network called P-CNN (for the paired CNN) to predict the spatial order of slices. We then use the network as a knowledge source for other tasks, e.g., fine-grained body part recognition. The idea behind the use of pre-trained slice ordering model is that correctly recognizing the relative position of s- lices requires good visual understanding of images. Indeed, as shown in [9], the initial several convolutional layers of a CN- N serve similarly like filters which automatically learn common low-level features such as edges, corners and textures. Therefore, weights from the pre-trained P-CNN can be used as a better initialization for body part recognition networks than randomized initialization. Compared to the network pre-trained on ImageNet, P-CNN is also a better knowledge source for medical image applications, closing the gap between pre-trained model and target medical tasks. In our experiments, we demonstrate that by pre-training P-CNN and fine-tuning, we can gain a boost for fine-grained body part recognition in terms of both resolution and accuracy, in CT as well as MR images. As far as we are aware, our work is the first attempt studying the problem of fine-grained body part recognition at a continuous level. 2. SLICE ORDERING WITH PAIRED CNN Conventional 3D medical data, such as CT or MR volumes are usually interpreted as sets of consecutive 2D slices for clinical diagnosis and treatment. Within high resolution 3D volumes, the difference between the slices within a small range may be very limited. The similar appearances of neighboring slices make it hard to differentiate them and estimate their spatial positions. On the other hand, the slices from distinct body parts may have distinct differences in structure. This makes it also difficult to design unified feature extractors in a handcrafted manner. Therefore, we employ a CNN in our method, which is known as an effective representation Fig. 1. Network architecture of the pair network for slice ordering. Every conv (convolution) and fc (fully connected) layers except fc8 are followed by a ReLU layer. The kernel size, output number and stride are listed after the layer names. learning technique without need of any feature designing expertise. We aim to learn a deep visual representation by pre-training a CNN for an auxiliary problem, i.e. predicting the spatial order of transversal slices. Then the CNN serves as a baseline with good generalization power to fine-grained body part recognition. For a complete 3D volume, a set of transversal slices as well as their order is available. Though every human body is unique in many details, the scanned CT or MR images share similar variation trend from head to toe. We can use their order as a free source of supervision and use a CNN to learn the intrinsic relationship among slices. Since the pre-training does not involve human annotation, we have more flexibility in choosing the size of training set. Using more data only means more computation time, but does not involve any extra human effort except data collection. We use a paired convolutional network called P-CNN to learn deep representation for slice ordering. The network architecture (Fig. 1) contains two stages: two AlexNet-like subnetworks for the lower six layers and the upper three global layers. We force the two sub-networks to share weights, leading to 50% fewer parameters in the lower six layers. Dropout layers with ratio 0.5 are used to prevent the network from overfitting. A final softmax loss layer is used for the binary classification. The training process is straightforward: 1) randomly sample a slice pair from the same volume; 2) feed the paired slices to the two subnetworks respectively; 3) fuse their outputs and compute the probabilities for two possible slice ordering outcomes (above vs. below); 4) predict the result as the class with higher probability. To solve this binary classification problem, the network must provide good visual understanding of objects and structures. Thus the network can serve as a universal low-level feature learner. Next, we demonstrate that the pre-trained network can be applied to body part recognition with fine-tuning.

Fig. 2. Transversal slices of eight different body parts from one full body CT volume. The scores at upper left corners indicate their normalized body heights in the volume. Six landmarks are used to build a normalized body height model. 3. FINE-GRAINED BODY PART RECOGNITION Transversal slice based body part recognition is an important pre-processing step which can help computer aided detection and diagnosis in several ways. First, precisely recognized body parts can serve as an ideal initialization for further analysis, e.g., detection, segmentation, and classification. Second, by labeling body part slices, organ/structure-based image retrieval can be easily implemented. Moreover, with fast recognition speed, automatic body part recognition can be integrated into current imaging techniques, such as CT and MR, to enable real-time planning and diagnosis. However, previous body part recognition methods remain at a coarse level, i.e., classifying a slice into one of a few region classes. In this paper, we model body part recognition as a regression problem instead, which is at a much finer recognition resolution. Our task is shown as in Fig. 2: given an slice from any part of the body, we aim to predict a real-number score in [0, 1] (with 0 for the bottom of a foot and 1 for the head top), indicating the normalized height in the body. To compensate the variation in body part ratio, six anatomical landmarks (head top, neck, lung top, spine, knee and foot) are manually annotated on the training set. Each landmark is mapped to its mean position in our normalized body height model and the height of other slices is obtained by linear interpolation using these six landmarks. This serves as ground truth for training and evaluation. We demonstrate that the learned P-CNN features can be successfully applied to fine-grained body part recognition for whole body images. We pre-train a P-CNN for slice ordering and copy one of the sub-networks of the lower six layers to the regression network (Fig. 3). The following two fully connected layers are randomly initialized from a Gaussian distribution with zero mean and small variation. An Euclidean loss layer replaces the softmax loss layer as the final layer [10]. We adopt the deep fine-tuning strategy [3] during training: we fine-tune all the layers and apply 1/10 original learning rate for the lower six layers, to preserve the power of the pretrained network and to boost learning speed for the following fully connected layers. Fig. 3. Network architecture for body part recognition. The lower six layers are initialized by copying a sub-network from the pre-trained P-CNN in Fig. 1, while Fc7 r and Fc8 r are trained with randomly initialized weights. 4.1. Slice ordering 4. EXPERIMENTS For slice ordering, we train a P-CNN to learn features from both slices and predict the spatial relative position. The slice pair is randomly sampled from the same volume to eliminate the effect of body shape variation between different people. With a large set of training samples covering every part of human body, the order of slices can be predicted no matter which part they belong to. Large variations in shape across slices require a good visual understanding of images, which makes the task challenging. A set of 370 CT volumes containing either full body or partial body is used for pre-training. The scanner table is removed from the image to eliminate the effect of non-body structures. The 2D transversal slices are extracted and resized to 256 256 pixels. Since the pre-trained AlexNet uses color images as input, we also transform the grayscale CT slices to color ones by duplicating the grayscale image to the three channels. Each pair of slices is randomly sampled from the same volume and labeled automatically with a binary pair label. Mirror images are also included for data augmentation. We subtract the mean from all images and shuffle the training pairs. Overall 83,000 pairs of slices are used for training and 32,400 pairs for testing. With P-CNN, we can achieve a prediction accuracy of 90% on the test set. The method is implemented based on the Caffe framework [10]. 4.2. Fine-grained body part recognition The proposed fine-grained body part recognition method is validated on two datasets (202 unseen CT volumes and 100 MR volumes). We compare the body part recognition results of different methods/settings: training from scratch (AlexNet-S), pre-trained AlexNet on ImageNet with finetuning (AlexNet-F) and pre-trained P-CNN on slice ordering with fine-tuning. The same preprocessing is performed to match the pre-training task. Hyperparameters are set as follows: momentum µ = 0.9; weight decay ω = 0.0005; the

Table 1. Euclidean loss of body part recognition on CT and MR data CT data MR data Portion of Training Set 100% 50% 25% 100% 50% 25% SIFT+BOW+SVR - - 0.0244 0.0096 0.0104 0.0130 SURF+BOW+SVR 0.0096 0.0194 0.0412 0.0108 0.0125 0.0134 AlexNet-S 0.00111 0.00655 0.02222 0.00046 0.00086 0.00209 AlexNet-F 0.00193 0.00867 0.02431 0.00061 0.00112 0.00242 Ours (P-CNN) 0.00137 0.00620 0.01973 0.00035 0.00073 0.00185 Table 2. Recognition error (in mm) of body part recognition on CT and MR data CT data MR data Portion of Training Set 100% 50% 25% 100% 50% 25% AlexNet-S 21.6 86.1 288.3 20.9 24.4 41.8 AlexNet-F 72.0 163.3 272.5 56.4 50.3 76.5 Ours (P-CNN) 25.3 81.2 229.8 20.1 24.3 34.9 (a) Fig. 4. Scatter plot of regression: our prediction (the x axis) versus ground truth (the y axis) on (a) CT data (b) MR data base learning rate α is set as high as possible such that the training still converges. Overall 102 CT volumes (18,700 slices) are used to train/fine-tune the network. The other 100 volumes (19,800 slices) are used as the test set. The CT data used for body part recognition are completely different to those used for pre-training. For MR images, we use 50 volumes (7,140 slices) for training and 50 volumes (7,245 slices) for testing. The test results without post processing on CT and MR image are illustrated in Fig. 4. The scatter plot is close to a diagonal line, which is the ideal case. If the input is a 3D volume, we can apply median filtering or line fitting to the raw regression scores of the slices to denoise and remove outliers for better performance. Quantitatively, 90.3% of CT test slices and 91.1% of MR test slices are predicted with an normalized error of less than 0.05. The recognition results are also compared with other methods. We train/fine-tune on both CT and MR volumes with three different sizes of training set: 100%, 50% and 25% of all training images. For all configurations, we evaluate on the same unseen CT/MR test set. We also provide results of several baseline methods including: 1) SIFT + Bag-of-Word + Support Vector Regression; 2) SURF + Bagof-Word + Support Vector Regression. As shown in Table 1, by pre-training on slice ordering task and fine-tuning, we can achieve the smallest Euclidean loss for most cases. We also demonstrate the average recognition error in millimeters in Table 2. With an average body height of 1809 mm and 1740 mm in CT and MR test set respectively, we can achieve a low recognition error of 25.3 mm and 20.1 mm, or 1.40% and 1.16% of full body height respectively. Note that a typical gap between neighboring transversal slices is 5 mm for CT (b) and 10 mm for MR, thus our body part recognition error is just within a few slices. The results also demonstrate that though P-CNN is pretrained only on CT data, it can generalize well to body part recognition on both CT and MR images. We also observe that the test performance is significantly affected by training set size. This is natural since more training data cover more cases with more information. With fewer training samples, the model tends to overfit after 10-200 epochs: the validation error begins to increase while the training error keeps decreasing. However, it is noticed that training from scratch overfits most easily and the models tend to overfit earlier with less training data. Without any human annotation, the pre-trained P-CNN outperforms pre-trained AlexNet by a large margin, though the training set of slice ordering is significantly smaller than ImageNet ILSVRC 2012 training set ( 89K vs 1.3M). Surprisingly, fine-tuning from pre-trained AlexNet performs not as well as training from scratch for this regression task. This indicates that the dissimilarity between natural images and medical images may pose considerable obstacle on the body part recognition task. 5. CONCLUSION In this paper, we propose a self supervised approach for deep representation learning for slice based body part recognition. Using only context information in 3D volumes, we can effectively solve the problem of slice ordering by P-CNN. The pre-trained P-CNN can also be transferred and fine-tuned for body part recognition. Our experimental results demonstrate that P-CNN can achieve promising results with minimal use of human annotations. 6. ACKNOWLEDGEMENT This research is supported in part by grants from National Science Foundation ACI 1443054 and IIS 1350885.

7. REFERENCES [1] Y. LeCun, Y. Bengio, and G. Hinton, Deep learning, Nature, vol. 521, no. 7553, pp. 436 444, 2015. [2] H. Greenspan, B. van Ginneken, and R. Summers, Guest editorial deep learning in medical imaging: Overview and future promise of an exciting new technique, IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1153 1159, 2016. [3] N. Tajbakhsh, J.Y. Shin, S.R. Gurudu, R.T. Hurst, C.B. Kendall, M.B. Gotway, and J. Liang, Convolutional neural networks for medical image analysis: Full training or fine tuning?, IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1299 1312, 2016. [4] T. Mikolov, I. Sutskever, K. Chen, G.S. Corrado, and J. Dean, Distributed representations of words and phrases and their compositionality, in Advances in Neural Information Processing Systems (NIPS), 2013, pp. 3111 3119. [5] C. Doersch, A. Gupta, and A. Efros, Unsupervised visual representation learning by context prediction, in IEEE International Conference on Computer Vision (ICCV), 2015, pp. 1422 1430. [6] X. Wang and A. Gupta, Unsupervised learning of visual representations using videos, in IEEE International Conference on Computer Vision (ICCV), 2015, pp. 2794 2802. [7] H.R. Roth, C.T. Lee, H. Shin, A. Seff, L. Kim, J. Yao, L. Lu, and R.M. Summers, Anatomy-specific classification of medical images using deep convolutional nets, in IEEE International Symposium on Biomedical Imaging (ISBI), 2015, pp. 101 104. [8] Z. Yan, Y. Zhan, Z. Peng, S. Liao, Y. Shinagawa, S. Zhang, D.N. Metaxas, and X.S. Zhou, Multiinstance deep learning: Discover discriminative local anatomies for bodypart recognition, IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1332 1343, 2016. [9] A. Krizhevsky, I. Sutskever, and G. Hinton, ImageNet classification with deep convolutional neural networks, in Advances in Neural Information Processing Systems (NIPS), 2012, pp. 1097 1105. [10] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell, Caffe: Convolutional architecture for fast feature embedding, in ACM International Conference on Multimedia, 2014, pp. 675 678.