Lung Nodule Detection using 3D Convolutional Neural Networks Trained on Weakly Labeled Data

Similar documents
LUNG NODULE DETECTION IN CT USING 3D CONVOLUTIONAL NEURAL NETWORKS. GE Global Research, Niskayuna, NY

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Kaggle Data Science Bowl 2017 Technical Report

Ryerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro

The MAGIC-5 CAD for nodule detection in low dose and thin slice lung CT. Piergiorgio Cerello - INFN

Deep Learning with Tensorflow AlexNet

Convolutional Neural Network based Medical Imaging Segmentation: Recent Progress and Challenges. Jiaxing Tan

arxiv: v1 [cs.cv] 20 Dec 2016

SELF SUPERVISED DEEP REPRESENTATION LEARNING FOR FINE-GRAINED BODY PART RECOGNITION

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Deep Face Recognition. Nathan Sun

Detecting Bone Lesions in Multiple Myeloma Patients using Transfer Learning

Graph based Image Segmentation using improved SLIC Superpixel algorithm

Robot localization method based on visual features and their geometric relationship

SIIM 2017 Scientific Session Analytics & Deep Learning Part 2 Friday, June 2 8:00 am 9:30 am

Classification of objects from Video Data (Group 30)

3D model classification using convolutional neural network

Study of Residual Networks for Image Recognition

Available Online through

End-to-end Lung Nodule Detection in Computed Tomography

Prostate Detection Using Principal Component Analysis

Leveraging Mid-Level Semantic Boundary Cues for Automated Lymph Node Detection

Machine Learning 13. week

Part Localization by Exploiting Deep Convolutional Networks

Using Machine Learning for Classification of Cancer Cells

Detection and Identification of Lung Tissue Pattern in Interstitial Lung Diseases using Convolutional Neural Network

3D-CNN and SVM for Multi-Drug Resistance Detection

arxiv: v1 [cs.cv] 11 Apr 2018

Alternatives to Direct Supervision

TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK

MEDICAL IMAGE COMPUTING (CAP 5937) LECTURE 19: Machine Learning in Medical Imaging (A Brief Introduction)

8/3/2017. Contour Assessment for Quality Assurance and Data Mining. Objective. Outline. Tom Purdie, PhD, MCCPM

The Automated Learning of Deep Features for Breast Mass Classification from Mammograms

Convolutional Neural Networks

Deep Learning in Visual Recognition. Thanks Da Zhang for the slides

Rotation Invariance Neural Network

Lung nodule detection by using. Deep Learning

Deep Learning for Computer Vision II

Microscopy Cell Counting with Fully Convolutional Regression Networks

Dynamic Routing Between Capsules

Unsupervised Learning

LIVER TUMOR SEGMENTATION CT DATA BASED ON ALEXNET-LIKE CONVOLUTION NEURAL NETS

3D Densely Convolutional Networks for Volumetric Segmentation. Toan Duc Bui, Jitae Shin, and Taesup Moon

Training Convolutional Neural Networks for Translational Invariance on SAR ATR

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR

Comparing Dropout Nets to Sum-Product Networks for Predicting Molecular Activity

Stochastic Function Norm Regularization of DNNs

An Exploration of Computer Vision Techniques for Bird Species Classification

AUTOMATED DETECTION AND CLASSIFICATION OF CANCER METASTASES IN WHOLE-SLIDE HISTOPATHOLOGY IMAGES USING DEEP LEARNING

Edge Detection Using Convolutional Neural Network

Deep Learning in Pulmonary Image Analysis with Incomplete Training Samples

Colonic Polyp Characterization and Detection based on Both Morphological and Texture Features

CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon

Automated Lesion Detection Methods for 2D and 3D Chest X-Ray Images

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

Hide-and-Seek: Forcing a network to be Meticulous for Weakly-supervised Object and Action Localization

Iterative fully convolutional neural networks for automatic vertebra segmentation

FEATURE DESCRIPTORS FOR NODULE TYPE CLASSIFICATION

Mass Classification Method in Mammogram Using Fuzzy K-Nearest Neighbour Equality

Real-time Object Detection CS 229 Course Project

Restricted Boltzmann Machines. Shallow vs. deep networks. Stacked RBMs. Boltzmann Machine learning: Unsupervised version

Artificial Intelligence. Programming Styles

Simultaneous Multiple Surface Segmentation Using Deep Learning

Learning-based Neuroimage Registration

Automatic Lymphocyte Detection in H&E Images with Deep Neural Networks

Medical images, segmentation and analysis

End-To-End Spam Classification With Neural Networks

Automatic Detection of Multiple Organs Using Convolutional Neural Networks

Hardware Processing Comparison in Convolutional Neural Networks Training

ECG782: Multidimensional Digital Signal Processing

Channel Locality Block: A Variant of Squeeze-and-Excitation

Machine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart

Automated segmentation methods for liver analysis in oncology applications

Stacked Denoising Autoencoders for Face Pose Normalization

Learning visual odometry with a convolutional network

Bioimage Informatics

Structured Prediction using Convolutional Neural Networks

Binary Image Processing. Introduction to Computer Vision CSE 152 Lecture 5

doi: /

Disguised Face Identification (DFI) with Facial KeyPoints using Spatial Fusion Convolutional Network. Nathan Sun CIS601

PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Su et al. Shape Descriptors - III

Multimodal Medical Image Retrieval based on Latent Topic Modeling

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Know your data - many types of networks

Journal of Asian Scientific Research FEATURES COMPOSITION FOR PROFICIENT AND REAL TIME RETRIEVAL IN CBIR SYSTEM. Tohid Sedghi

IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES

Engineering Problem and Goal

Object detection using Region Proposals (RCNN) Ernest Cheung COMP Presentation

Unsupervised learning in Vision

JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation

K-Means Clustering Using Localized Histogram Analysis

Object Localization, Segmentation, Classification, and Pose Estimation in 3D Images using Deep Learning

Copyright 2007 Society of Photo Optical Instrumentation Engineers. This paper was published in Proceedings of SPIE, volume 6514, Medical Imaging

Edge and corner detection

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

FACE DETECTION AND RECOGNITION OF DRAWN CHARACTERS HERMAN CHAU

Computer Vision Lecture 16

Transcription:

Lung Nodule Detection using 3D Convolutional Neural Networks Trained on Weakly Labeled Data Rushil Anirudh 1, Jayaraman J. Thiagarajan 2, Timo Bremer 2, and Hyojin Kim 2 1 School of Electrical, Computer and Energy Engineering, Arizona State University 2 Center for Applied Scientific Computing, Lawrence Livermore National Laboratory ABSTRACT Early detection of lung nodules is currently the one of the most effective ways to predict and treat lung cancer. As a result, the past decade has seen a lot of focus on computer aided diagnosis (CAD) of lung nodules, whose goal is to efficiently detect, segment lung nodules and classify them as being benign or malignant. Effective detection of such nodules remains a challenge due to their arbitrariness in shape, size and texture. In this paper, we propose to employ 3D convolutional neural networks (CNN) to learn highly discriminative features for nodule detection in lieu of hand-engineered ones such as geometric shape or texture. While 3D CNNs are promising tools to model the spatio-temporal statistics of data, they are limited by their need for detailed 3D labels, which can be prohibitively expensive when compared obtaining 2D labels. Existing CAD methods rely on obtaining detailed labels for lung nodules, to train models, which is also unrealistic and time consuming. To alleviate this challenge, we propose a solution wherein the expert needs to provide only a point label, i.e., the central pixel of of the nodule, and its largest expected size. We use unsupervised segmentation to grow out a 3D region, which is used to train the CNN. Using experiments on the SPIE-LUNGx dataset, we show that the network trained using these weak labels can produce reasonably low false positive rates with a high sensitivity, even in the absence of accurate 3D labels. 1. INTRODUCTION The last decade has seen significant advances in using machine learning for computer aided diagnosis (CAD), which can significantly improve efficiency and reduce costs. The continued success of CAD tools can be attributed to the development of feature representations that can work well under several different conditions with invariances to properties such as brightness, shape, size and geometric transformations. More recently, advances in representation learning (e.g. deep neural networks) have enabled inference of features from training data in lieu of hand-tuned feature design by an expert. 1 These have resulted in significant boosts in accuracy for tasks such as image recognition, natural language understanding, and speech recognition. 1 However, there are significant hurdles before such successes can be transferred to benefit the medical imaging community. A major limiting factor is the difficulty in obtaining annotated data, which is significantly more expensive than compared to traditional computer vision. In this paper, we consider the problem of detecting early stage lung nodules based on learned representations. This is a crucial problem in medical diagnosis since it is estimated that more people died due to lung and bronchus cancer than all other cancers combined in 2015. 2 Classical approaches typically segment the lung, extract features from the training data and train a classifier to detect potential nodules. 3 However, lung nodule detection is inherently more challenging due to the high variability of nodule shape, size, and texture. As a result, nodule detection techniques that employ classifiers learned using hand-engineered features often provide poor generalization to novel test data. More recent approaches that employ deep neural networks in their pipeline have achieved state-of-the-art detection performances. For example, Kumar et al. 4 use an autoencoder (an unsupervised learning network) to extract useful features from annotated nodules, these features are used to learn to classify nodules as being malignant or benign. Next, Ginneken et al. 5 have shown promising results using an off-the-shelf convolutional neural network (CNN), one that is pre-trained for an image recognition task. They use the network to obtain features which are used for classification. Two dimensional CNNs have been used in other CAD methods such as pancreas segmentation, lymph nodes and colonic polyp This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344. Corresponding author email: ranirudh@asu.edu

detection. 6 Most of these methods are trained individually on 2D images with 2D convolutional filters, whereas the data at hand is inherently 3 dimensional. Roth et. al. 7 have addressed this by considering a 2.5D representation that takes slices of the images from a point of interest in 3 orthogonal views. These slices are combined to be treated as a 3-channel image, which is used to train a deep network. In contrast, we propose to train a full 3D convolutional network, that can directly learn 3D convolutional filters from the data. Such filters are beneficial because it can capture the full range of variations expected from the lung nodules. However, there are two crucial challenges in generalizing 2D convolutional networks to 3D. The foremost challenge is the need for labeled training data, which can be prohibitively expensive to obtain for 3D images. In fact, most existing systems for detection rely on detailed nodule segmentations provided by an expert for model training - which is particularly unrealistic in the case convolutional neural nets since they require a much larger training dataset for learning features with high representative power. In several computer vision applications, this problem is addressed by outsourcing labeling using services such as Amazons MTurk, which cannot be readily adapted for medical image analysis since experts are required to effectively interpret the data. Consequently, the proposed detection system reduces the labeling effort on experts by working with point labels which are essentially single pixel locations potentially indicating the center of the nodules. By using unsupervised learning methods to estimate the true label from the weak information, we show that we can reduce the effort required on the expert to label, while being able to train 3D networks that can discriminate effectively. The second challenge is pertinent to the computational burden of 3D neural networks. While 3D convolutions are expensive, particularly for processing 3D scans (typically 512 512 200), building a single network that can potentially handle nodules of varying sizes in different regions of the scan is hard. To circumvent this, we propose to train our network on smaller 3D regions centered around the nodule instead of the whole image, and simultaneously build two networks with different context sizes, 41 41 7, and 25 25 7 respectively. The final detection is obtained as the consensus of the two networks. Our primary contributions can be summarized as follows: Contributions: 1. We present a modular system that leverages the robustness of 3D convolutional neural networks for the problem of lung nodule detection. Our system learns the most discriminative features for nodule detection instead of working with hand engineered features such as shape and texture. To the best of our knowledge, we are the first to explore lung nodule detection using 3D convolutional filters. 2. Our system works with point labels, which specify a single voxel location that indicates the presence of a nodule, and its largest cross sectional area. This is much more time efficient compared to the detailed annotations of a nodule in the training set, which is highly impractical since experts are needed to provide these. Using unsupervised learning methods, we estimate a final 3D label which is used to train our 3D CNN. 3. By learning two different networks with varying context sizes, our detection system achieves improved generalization. 4. We demonstrate promising results on the AAPM-SPIE-LungX nodule classification dataset. 2. METHODOLOGY In this section we outline different aspects of our system, that can make predictions on 3D volumes of CT scans. First we address the label estimation procedure for training data. Next we use these estimated labels to train our 3D convolutional network. 2.1 Estimating weak labels A limiting factor for using 3D CNNs is the cost of obtaining detailed 3D labels, which are significantly harder to obtain than 2D labels. This is exacerbated by the fact that experts such as radiologists are needed to label lung nodules, as opposed to crowd sourcing platforms such as Amazon MTurk, which have become a norm in computer vision. Therefore, we begin by using only a single voxel location or a point label, which indicates the presence of a nodule. Such point labels are a natural way for experts to annotate lung scans efficiently. We

Region of Interest Basic Thresholding Filtering Superpixels Final estimated label Original Slice Figure 1: Estimating the ground truth per slice from a point label given by the expert. These 3D labels are used to train the 3D-CNN. process the slices in 2D, and combine them using 3D Gaussian filtering. First, we obtain 2D SLIC superpixels 8 to oversegment each slice, as shown in the figure 1. These superpixels find contiguous regions in the image, which are used to eliminate obvious regions that are not nodules based on size and intensity. The 3D Gaussian filtering reduces noise and combines the 2D slices to form a coherent 3D nodule. We are able to do this accurately because we are looking at a local neighborhood around the nodule. The size of the local neighborhood is determined by the largest cross sectional area of the nodule, as given by the expert. The superpixels can effectively aid in capturing nodules that are hard to distinguish at times, such as when they are touching a lung wall. 2.2 Training The 3D CNN is trained to predict whether or not a single voxel is likely to be a nodule or not, based on the spatio-temporal statistics around it. For example, if the location of the nodule is at V (x, y, z), where V is the entire CT volume, we choose the input volume to be ˆv = V (x w : x+w, y w : y+w, z h : z+h), where h is the window size in X, Y planes and h in the Z plane. We used values in the range of w = 10 25 and h = 3, 5. The volume is thinner in the z plane because CT scans are typically sampled much more densely in X, Y planes than in Z. There are at the most 2 nodules per scan, but training a 3D CNN requires many examples to effectively learn the filters. Therefore in order to inflate our training set, we treat different voxels within the same nodule as different positive examples. A typical nodule can range from 3 28 pixels wide at its largest size, and spans 3 7 slices typically. We center our volume at several different randomly sampled voxels within the nodule and pick the resulting volume for a given w, h as a positive training example. Inflating training sets have been useful to train networks that can achieve robustness and avoid overfitting. 7, 9 The negative set is much harder to obtain than the positive set because its hard to define it. A negative class contains all examples from the lung that are not the nodule. A smart approach can provide a much better definition of what a negative sample should be. In the ideal case, we only need to choose negative samples to be those which are expected to be easily mistaken by our network. Therefore, we restrict the negative space to lie within the lung, since it is highly unlikely for a nodule to be found outside it. Next, we random sample locations which have an intensity above a threshold ( 400 500 on the Hounsfield scale). These sampling methods resulted in about 15K positive samples and 10, 11 around 20K negative samples from the AAPM-SPIE-LungX dataset. 2.3 Architecture of the Convolutional Neural Net We trained a 3D CNN using the MatConvNet toolbox for MATLAB. 12 The toolbox allows us to specify the kind of layers, and the number of filters needed. The network was designed to be similar to most of the popular models for image recognition. 9 As shown in figure 2, our network contained 5 convolutional layers which were followed by Rectified Linear Unit (ReLU) activation layers, 9 2 max-pooling layers, and a final 2-way softmax layer for classification. We also use dropout 13 to regularize the learning problem. Of the five convolutional layers, two are fully connected (FC) with convolution kernels of size 1 1. The generalization from 2D convolutional networks to 3D networks is trivial, in that the filters that are learned are 3 dimensional. Since we use a multiscale approach,

41x41x7 16x16x7x50 7x7x50x50 3x3x50x80 2x2x80x100 CT Volume Max pooling Layer Max pooling Layer FC Layers 2-way Softmax classifier P(nodule) Figure 2: Overall design of the 3D convolutional neural network trained for lung nodule detection. we train two different networks for each scale, the convolutional filter sizes for the larger 3D CNN are shown in 2. For the smaller scale, we use the same architecture, and modify the sizes of kernels accordingly. 2.4 Testing and candidate generation Our network is trained end-to-end, i.e. it is able to make predictions regarding the presence of a nodule directly from a CT volume of the appropriate size. However, a typical scan is 512 512 200 in size and searching the entire 3D volume to make predictions is highly impractical. Instead, we reduce the search space significantly by ruling out parts of the scan that are very unlikely to contain a lung nodule. Since the lung nodule is expected to be inside the lung, we perform lung segmentation using morphological operations on each 2D slice. The segmentation of the lung itself is a hard problem, and there are dedicated systems to perform effective segmentation in 2D and 3D. We also observed that most of the false positive detections on the system were because of the airways which are part of the lung but look a lot like nodules when observed locally. Therefore a robust 3D segmentation can significantly reduce the false positive rate, and improve speed of detection. Next, for each voxel we apply the dot enhancement filter using the 3D Hessian. The resulting dot score is high if the region around the current voxel is spherical. 6 The dot score map is thresholded in each local neighborhood to provide the final list of candidates. This method can be very effective when the nodules are expected to be approximately round in shape. The dot score is computed as λ 3 2 /λ 1, where λ 1, λ 3 are the first and third eigenvalues of the 3D Hessian. The dot score for a given volume essentially provides an estimation of its roundness such that a high score indicates a tendency towards roundness. We set a low threshold to eliminate obviously non nodule-like elements, and run 3D Gaussian smoothing filter to remove smaller stray particles within the volume. These steps significantly reduce the false positives, resulting in around 80-200 3D nodule like candidates per scan. After smoothing, these can be easily identified using a 3D connected component algorithm efficiently. Finally, we center a test volume at multiple locations inside each candidate and obtain a prediction from the deep network. This also allows us to perform voting in order to eliminate noisy predictions by running a smoothing filter on the predictions. 3. EXPERIMENTS In this section we describe the dataset, experimental conditions, and results obtained for lung nodule detection. SPIE-AAPM-LUNGx dataset: The dataset has been published for nodule classification, which requires labeling each nodule as benign or malignant. We use the dataset for detection, as it does not contain detailed labels for nodules, and hence a realistic test case. Of the 70 scans, we have used 20 for training and 47 for testing. Three scans were discarded because there was ambiguity regarding the presence of a nodule at the specified location. The label is provided as an (x, y, z) location along with information on the largest cross sectional area of the nodule. We did not use this information, however, it could be used to estimate better labels.

3.1 Evaluation settings For the test scans, we first generate ground truth labels in a similar fashion as described for the training data. These estimated labels are used to evaluate the performance of our system on the dataset. Multiscale CNN: The lung nodules vary significantly in size typically from around 3mm - 20mm. Many successful detection systems employ a multi-scale architecture. Since we are interested in 3D volumes, there are several ways to choose the scale. We chose two scales at 25 25 7, and 41 41 7 experimentally. We train them separately and obtain the predictions from each CNN to obtain the final result. The combination performed much better in terms of sensitivity and accuracy as expected. Finally, we generate the free receiver operating curves at various detection thresholds. At a particular threshold, we declare a match if there is a nodule around a small radius (typically 5 10) of the ground truth. This is done by first estimating the centroid of each 3D blob in the test prediction that is greater than the threshold. Next, we find the distances from each centroid to the ground truth. Only the one that is closest and within a distance threshold is considered a positive, the rest are considered false positives. For each threshold the total number of false positives divided by the number of scans gives the average false positive rate. Results: We compute the free receiver operating characteristic (FROC) for our system, which plots the sensitivity against the average number of false positives per scan. The results are shown in figure 3a. As it can be seen, even with a weak labeled system, we achieve sensitivities of 80% for 10 false positives per scan. Sample predictions are shown in figure 3. 0.9 0.8 Free Reciever Operating Characteristic 0.7 Sensitivity 0.6 0.5 0.4 scale 41x41 scale 25x25 2 scale CNN 0.3 0.2 0.1 10 20 30 40 Average False Positives Per Scan (a) ROC curve (b) Sample results Figure 3: Detection performance on the SPIE-AAPM LUNGx dataset. 4. CONCLUSION & FUTURE WORK: We have presented a system for lung nodule detection that works with 3D convolutional networks trained using weak label information. While the initial results look promising, there are areas to further improve the system. Our current system currently processes superpixels in 2D, this could be improved with a 3D superpixel system that clusters coherent spatio-temporal regions. A 3D lung segmentation approach could also eliminate the air tracts which are a primary cause for false positives, but these cannot be differentiated when observed in 2D. Next, the training set can be inflated even further using 3D transforms of existing labels, which has been done for 2D CNNs. Such a technique will also ensure there is little overlap between the original label and transformed label to avoid overfitting.

REFERENCES [1] LeCun, Y., Bengio, Y., and Hinton, G., Deep learning, Nature 521(7553), 436 444 (2015). [2] http://www.lung.org/lung-disease/lung-cancer/lung-cancer-screening-guidelines/ lung-cancer-screening-for-patients.pdf. [accessed 13-Aug-2015]. [3] Dhara, A. K., Mukhopadhyay, S., and Khandelwal, N., Computer-aided detection and analysis of pulmonary nodule from ct images: A survey, IETE Technical Review 29(4), 265 275 (2012). [4] Kumar, D., Wong, A., and Clausi, D. A., Lung nodule classification using deep features in ct images, in [Computer and Robot Vision (CRV), 2015 12th Conference on], 133 138, IEEE (2015). [5] van Ginneken, B., Setio, A. A., Jacobs, C., and Ciompi, F., Off-the-shelf convolutional neural network features for pulmonary nodule detection in computed tomography scans, in [Biomedical Imaging (ISBI), 2015 IEEE 12th International Symposium on], 286 289, IEEE (2015). [6] Choi, W.-J. and Choi, T.-S., Automated pulmonary nodule detection based on three-dimensional shapebased feature descriptor, Computer methods and programs in biomedicine 113(1), 37 54 (2014). [7] Roth, H. R., Lu, L., Seff, A., Cherry, K. M., Hoffman, J., Wang, S., Liu, J., Turkbey, E., and Summers, R. M., A new 2.5 d representation for lymph node detection using random sets of deep convolutional neural network observations, in [Medical Image Computing and Computer-Assisted Intervention MICCAI 2014], 520 527, Springer International Publishing (2014). [8] Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., and Susstrunk, S., Slic superpixels compared to state-of-the-art superpixel methods, Pattern Analysis and Machine Intelligence, IEEE Transactions on 34(11), 2274 2282 (2012). [9] Krizhevsky, A., Sutskever, I., and Hinton, G. E., Imagenet classification with deep convolutional neural networks, in [Advances in neural information processing systems], 1097 1105 (2012). [10] Spie-aapm-nci lung nodule classification challenge dataset. https://wiki.cancerimagingarchive.net/ display/doi/spie-aapm-nci+lung+nodule+classification+challenge+dataset. [accessed 13-Aug- 2015]. [11] Armato, III, S. G., Hadjiiski, L., Tourassi, G. D., Drukker, K., Giger, M. L., Li, F., Redmond, G., Farahani, K., Kirby, J. S., and Clarke, L. P., Guest editorial: Lungx challenge for computerized lung nodule classification: reflections and lessons learned, Journal of Medical Imaging 2(2), 020103 (2015). [12] Vedaldi, A. and Lenc, K., Matconvnet-convolutional neural networks for matlab, arxiv preprint arxiv:1412.4564 (2014). [13] Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R., Dropout: A simple way to prevent neural networks from overfitting, The Journal of Machine Learning Research 15(1), 1929 1958 (2014).