Lung Tumor Segmentation via Fully Convolutional Neural Networks

Size: px
Start display at page:

Download "Lung Tumor Segmentation via Fully Convolutional Neural Networks"

Transcription

1 Lung Tumor Segmentation via Fully Convolutional Neural Networks Austin Ray Stanford University CS 231N, Winter 2016 Abstract Recently, researchers have made great strides in extracting traits of lung tumors from Computerized Tomography (CT) scans in order to more accurately diagnose and better treat patients. Logically, these lung tumor features are extracted only from pixels that lie in and around the tumor. Currently, figuring out which voxels belong to the tumor (a problem which we will refer to as tumor segmentation ) is done manually by trained radiologists - an unfortunate inefficiency in the process. Algorithms based on traditional computer vision techniques have proven too inaccurate to be useful. Building on recent image segmentation breakthroughs at the Berkeley Vision and Learning Center [1], this paper presents a fully convolutional neural network that using an initial bounding box around the tumor can segment a tumor in two dimensions at a mean testing Dice Score [Figure 3] of 83%. 1. Introduction 1.1. CT Scans A CT scan of a patient s chest is usually one of the first tests a radiologist will do to see if a patient has a lung tumor. For reference, a single CT scan image (slice) is 512x512x1 pixels (grayscale) and there are usually ~200 slices per scan we say that these 200 slices compose the CT volume. If a lung tumor does appear on the CT scan, radiologists can then analyze the scan further to figure out the type, size, age, and danger of the tumor - in other words, the tumor s features. While trained radiologists can ascertain quite a bit about a tumor with their eyes alone, there exist a wealth of other features of the tumor that can only be gleaned through computational methods. pixels and pixels close (i.e. relevant) to those tumor pixels that is, we don t want to extract features from a rib on the left side of the lung when what we really care about is a tumor on the right side of the lung. This is the reason why accurate tumor segmentations are important to provide a pure set of pixels for the feature extractor to analyze Inputs and Outputs We will refer to the neural network discussed in this paper as the Tumor Segmentation Fully Convolutional Network, or TS-FCN. The input to the TS-FCN is a 100x100x3 image with a tumor somewhere in it. All inputs are assumed to have tumors in them (the focus here is not on detection). To form each 100x100 image, a liberal bounding box is drawn around the tumor on a 512x512 CT slice. Bounding boxes are 100x100 at minimum in order to retain spatial context from the lung. Next, the area inside this bounding box is cut out from the 512x512 image and resized down to 100x100 if necessary. For example, if a tumor is actually ~80x80, we might make a 160x160 liberal bounding box, cut out this box from the 512x512 source CT image, and then downsize to 100x100 to get an input to the TS-FCN. Since the original 512x512 slices have only 1 color channel, we copy this channel to 2 additional color channels for our 100x100 bounding box images in order to utilize pre-trained convolution layers that only accept 3- channel images. The output of the TS-FCN is a 100x100x1 image. A pixel in the image is 1 if the TS-FCN predicted that pixel to be a tumor, and 0 otherwise Importance of Segmentation The computational feature extractor is given some collection of 2d regions on various slices in a CT volume. The extractor only looks inside these regions for features. Ideally, these regions will be composed only of tumor 225

2 Figure 1: Output of previous state-of-the-art segmentation system (SDS) and of deep jet compared to the ground truth segmentation. 2. Related Work 2.1. Manual Segmentation Since radiology is a field where errors are extremely costly, it is vitally important for radiologists to use the most accurate tools available to them, no matter the cost or time. Consequently, since the human eye is currently the best segmentation algorithm out there, all lung tumor segmentation for the sake of feature extraction is done by hand. This can take as much as 10 minutes per patient, however, which is quite a lot of time for radiologists. Another issue with manual segmentation is that there is often no consensus about which regions in an image contain features worth extracting. Some radiologists are conservative with their segmentations and perfectly outline the solid portion of a tumor. Other radiologists are more liberal and tend to include surrounding regions, the idea being that there is important information contained in those regions as well. In fact, recent work by a graduate student in my lab shows that segmentations can differ by as much as 20% between radiologists (paper release pending), which can be disastrous for algorithms down the pipeline that rely on accurate segmentations. The reasons why it is important to develop computational segmentation algorithms are therefore (1) to save radiologists time, and (2) to provide precise segmentations Windowed Segmentation One way that researchers have approached segmentation is by using windowed neural network architectures. A group out of IDSIA recently implemented this architecture in order to perform two-class membrane segmentation on 2D slices from brain scans [2]. The neural network in that paper takes in a square image centered on a pixel and outputs a class for the center pixel. You can attain a full segmentation for an image by feeding in a series of windows centered on all pixels in the image. In other words, you would have to feed all 10,000 windows from a 100x100 image into the net in order to get a full segmentation for that image. One interesting issue this paper deals with is heavy class imbalance in a 2-class segmentation problem. Heavy class imbalance is a problem because the neural network can simply guess that a pixel belongs to the much-morefrequent class every time and automatically get extremely low loss a local minima that is hard to escape from. For my project, I have to face the same problem, since tumor pixels only compose ~4% of pixels in a 100x100 bounding-box input image. The IDSIA paper cleverly deals with class imbalance by down-sampling pixels from the more frequent class so that the neural network trains on an equal number of pixels from each class. After training, the IDSIA group places a calibration layer on top of their net s output for scaling probabilities to match the statistical reality of the two classes in the data. Unfortunately, since I ended up going the fully convolutional route, it was not possible for me to downsample inputs, since every pixel in an image was necessary to classify the other pixels in that image. However, the IDSIA group inspired me to implement a weighted loss function to counter class imbalance, the implementation of which I will talk about later. Although this paper reports extremely high accuracy and its methodology seems straightforward, its reported time of computation per pixel is over ten times that of the paper I will mention in section 2.3. Due to time constraints for this project, I chose not to pursue this methodology FCNs for Semantic Segmentation Most of the inspiration for this project comes from a recent paper out of the Berkeley Vision and Learning Center [1]. The authors of this paper outline a neural network architecture, which they call deep jet, that takes in a 2D image from the PASCAL VOC dataset [6] as input and outputs a 21-class semantic segmentation of that image (see examples in Figure 2). The architecture differs from past segmentation architectures in that it is fully convolutional, meaning it has no inner product (linear) layers only convolutional layers. This allows the network to take in 2D inputs of arbitrary size and return an output of the same size, where each pixel is a prediction for its corresponding pixel in the input image. Deep jet's ability to process all pixels in an image at once make it 1-2 orders of magnitude faster than windowed segmentation. Additionally, deep jet enjoys 226

3 larger receptive fields than the windowed segmentation architecture, as it is not limited by the chosen window size. This added context can be quite useful in segmentation. 3. Methods All neural networks used in this project were built and executed using the Caffe framework [5] Net Architecture The TS-FCN used in this paper is much like the deep jet architecture used in the BVLC paper [1], but slightly less complicated and smaller. The net architecture is outlined below: (1) Conv64 ReLU Conv64 ReLU MaxPool (2) Conv128 ReLU Conv128 ReLU MaxPool (3) Conv256 ReLU Conv256 ReLU MaxPool (4) Conv4096 ReLU Dropout0.5 (5) Conv4096 ReLU Dropout0.5 (6) Conv2 (7) Deconv8x Crop (8) Softmax WeightedMultiClassLoss All Conv64, Conv128, and Conv256 convolution layers use size 3 filters with a stride of 1. All use zero padding of 1 except for the first Conv64, which uses zero padding of 32 so that the image isn t too small after the 3 max-pools. The first Conv4096 uses a filter size of 7. The second Conv4096 and Conv2 both use filter size 1 these are the layers that make the net fully convolutional, as they mimic fully connected/linear layers applied to each pixel. Conv2 gives a score for each of the 2 classes for each pixel. At this point, the image has been max-pooled 3 times and is thus about 8x smaller than it started 2x15x15, to be exact. The deconvolution layer uses a filter size of 16 and a stride of 8, meaning the result is 2x128x128 (Solve [(H 16)/8 + 1 = 15] for H). This result is then cropped in the center to give a 2x100x100 image. The softmax function is then applied to each pixel s 2 class scores to give class probabilities. These probabilities are then fed into a weighted multi-class loss layer where each pixel is treated as a sample in a mini batch (100x100 image is a 10,000 member mini batch), with the total loss equaling the average loss over every pixel in the image. To account for the fact that tumor pixels are only 4% of all image pixels in my dataset, I weighted the contribution of each ground-truth tumor pixel by 0.96, and the contribution of each ground-truth non-tumor pixel by A regularization weight of 5e-4 was used in the loss equation in order to increase performance on the validation set. The final loss equation I used can be seen in Figure 2. Figure 2: Weighted softmax loss function. y_i is the ground truth value for pixel i - either 0 (not tumor) or 1 (tumor). f_j is the score for class j for the pixel (output from layer (7)), where j is 0 or 1. reg_image is a weight regularization term Learning Weights for layers (1), (2), and (3) were transferred from the corresponding layers in source [1], which were available at [3]. These weights can be interpreted as producing a feature set for each pixel from any given image, with the rest of the net learning to use these features to do segmentation. These pre-trained layers were trained at 1/10 th the learning rate of the other layers. The deconvolution layer was initialized with bilinear interpolation weights and its learning rate was set to 0 (it therefore just served as a basic image up-sizer). ADAM was used as an update rule, with learning rate 1e-6. The learning rate was multiplied by 0.9 four times per epoch. The net was trained for around 5 epochs. A batch size of 1 image was used Performance Metrics The main performance metric used was Dice Score, which is like intersection over union. To calculate the dice score for a given prediction/ground-truth pair, you count up all the pixels they both predicted were tumors, divide that by the combined sum of tumor pixels predicted by both, and multiply by 2. This gives a value between 0 (no agreement) and 1 (full agreement). The equation can be seen in Figure 3 below. This metric is a much better evaluator of performance than pure accuracy, as pure accuracy will be high for all images due to the prevalence of non-tumor pixels. 227

4 Figure 3: Dice score equation, where A is the set of predicted tumor pixel locations and B is the set of actual tumor pixel locations. One important thing to note is that 2D dice scores aren t that interesting to us here. What we re really interested in is the dice score over a patient s entire tumor, which spans multiple slices. Thus, all validation dice scores are 3D dice scores, where the numerators and denominators are each summed individually before being divided by each other. 4. Dataset and Features I have access to CT scans and corresponding radiologist segmentations for 107 patients through the Stanford Radiological Image and Information Processing Lab, where I work [4]. Each patient s scan has around 200 slices in it. Most of these slices have no tumor in them. Each slice is a 512x512 grayscale image. Not all scans were done by the same machine or technician a lot can vary from scan to scan. This is part of the challenge Preparing the Data First, I got rid of any slices that had no tumor in them (as seen when the segmentation for that layer had no 1 s (tumor pixels) in it). Next, I created square bounding boxes for each tumor that were about twice as big as the tumor s maximum width and height at any layer in the patient. Any bounding boxes under 100x100 were automatically set to 100x100 and centered on the tumor. Any bounding boxes over 100x100 were fine, but were resized to 100x100 after being cut out in order to standardize inputs. Labels (segmentations) were cut out and resized exactly like their corresponding slices. The data s grayscale values ranged from to I rescaled these values to 0 to 255 in order to be in line with what the first three pre-trained conv layers expect. The images color channels were then arbitrarily increased to 3 channels instead of 1, copying that 1 channel equally into 3 in order to make the images compatible with the first 3 pre-trained conv layers. Interestingly, using the color channel means from [3] which has a different mean for each channel worked better than using the actual mean for the data, which would also be the same across all 3 channels. The data was then further augmented through mirroring and rotation to produce 8x as much data, giving around 26,000 total images instead of the original ~3200 images. This turned out to help tremendously increasing the validation accuracy of the net. The data was split into 5 folds 4 folds with 21 patients and 1 folds with 23 patients. The data was split this way because the evaluation metric is not on a per-image basis but on a per-patient basis, so it is important to have an equal number of patients in each fold. Unfortunately, due to time constraints, cross-validation testing could only be carried out once. However, I intend to execute complete 5- fold cross-validation when I continue with this project. Figure 4: Example data before bounding boxes applied. Inputs on the left, labels on the right. 228

5 5. Results and Discussion 5.1. The Good After training for ~7 hours (~5 epochs), a 3D validation dice score of 0.86 was achieved. Figure 5 shows validation accuracy as a function of 1/10 th epochs (e.g. 20 corresponds to score after 2 epochs, 50 to score after 5 epochs apologies for the plot messiness). The overall accuracy on a per-image basis was ~99% (high, due to 96% of pixels being background). With a validation dice score of 0.86, these segmentations become well within the margin of the 20% variance between radiologists. As such, it is my opinion that this model is definitely worth pursuing further in order to develop a truly useful tool for radiologists going forward. Figure 5: Average validation 3D dice score over all validation patients as a function of number of iterations (10 = iterations, 20 = iterations, etc.) Some example outputs of the resulting model on the validation set can be seen in Figure 6. Although we can see a few mistakes here, the model seems to be highly accurate on the whole. It should be noted that these images are some of the higher-end 2D segmentations, with 2D dice scores in the low to mid 90s. Interestingly, we can see that the predicted segmentation is often more blob-y than the ground truth segmentation. My hypothesis is that this is a result of the high regularization used in order to prevent against overfitting in the training set. Another interesting point is that, while larger mini-batch sizes are usually better, they consistently performed worse for this model, with a mini batch size of 1 giving by far the best performance. I posit that this is because there is so much to learn from each image (a 10,000-pixel mini-batch itself) that grouping images together in a batch makes the net try to be too broad in its improvements. With just one image per batch, the net can make smaller adjustments to the net in order to compensate for specific cases, ultimately leading to a better result. Figure 6: Model evaluated on images from the validation set after 5 epochs of training. 229

6 Another interesting thing to look at is what an image looks like after the score, deconv+crop, and softmax layers (Figure 7). As said previously, the resolution of the image before deconvolution is only 15x15. The values for the yes tumor score for each pixel are displayed in figure 7. We can see that although the 15x15 image looks choppy, it actually carries enough information to be upscaled and converted to a very nice pre-threshold segmentation (the original image, true segmentation, and predicted segmentation after thresholding can be seen in the top left series of Figure 6). Figure 8: (top left) Original image, (top right) Image after softmax layer, (bottom left) Predicted segmentation after thresholding, (bottom right) Ground truth segmentation Figure 7: Data as it passes through last layers and gets transformed into a segmentation The Bad There are many areas where this framework fails. Besides the blob-iness mentioned previously, the network seems to lack confidence for some tumors (Figure 8). By this, I mean that the network sees the tumor and gives its pixels some positive probability of being tumor, but isn t convinced and ends up leaving them out of the final segmentation. This might be corrected by increasing the weight on the weighted loss in order to more harshly penalize for missing ground truth tumor pixels. The model also tends to sometimes think bones, tissues, and organs are tumors (Figure 9). This is a fair mistake, and one that I would make, but at the same time I would expect the model to recognize that that area s textures are not tumor-like and ignore that area. I think this is a symptom of a bigger problem: lack of data. With only 107 patients, I only have access to 107 tumors, and it s likely that some of these tumors are unique, given the genetic variety seen in tumors. Thus, while we might expect the training set to be representative of the validation set, this might not be the case, and some tumors in the validation set could be unlike anything seen in the training set. With more patients, this might not be as much of a problem. One other possible solution besides more data (which I m pretty sure is everyone s solution to every machine learning problem), is to allow different images belonging to the same patient to be in both the training and validation sets. Of course, we would need to change the performance metric from 3D dice score to 2D dice score, but we might see an overall increase in performance due to making the training set more representative. 230

7 Figure 9: (top left) Original image, (top right) Image after softmax layer, (bottom left) Predicted segmentation after thresholding, (bottom right) Ground truth segmentation. 6. Conclusions and Future Work Based on the results seen here, we can see that fully convolutional networks are fast, powerful systems for predicting segmentations. I plan to continue my work with FCNs for lung tumor segmentation next quarter for my lab, exploring the possibility of 3D convolutions across CT slices, attempting to segment a full 512x512 image instead of a window around the tumor, and finishing implementing the entirety of the Deep Jet architecture, including the combination of predictions from various resolutions in the net. 7. References [1] Fully Convolutional Networks for Semantic Segmentation; By Jonathan Long, Evan Shelhamer, and Trevor Darrell; BVLC, 2014 [2] Deep Neural Networks Segment Neural Membranes in Electron Microscopy Images; By Dan C. Ciresan, Luca M. Gambardella, Alessandro Giusti, and Jurgen Schmidhuber; IDSIA, 2012 [3] file-readme-md [4] [5] [6] 231

Fully Convolutional Networks for Semantic Segmentation

Fully Convolutional Networks for Semantic Segmentation Fully Convolutional Networks for Semantic Segmentation Jonathan Long* Evan Shelhamer* Trevor Darrell UC Berkeley Chaim Ginzburg for Deep Learning seminar 1 Semantic Segmentation Define a pixel-wise labeling

More information

DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution and Fully Connected CRFs

DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution and Fully Connected CRFs DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution and Fully Connected CRFs Zhipeng Yan, Moyuan Huang, Hao Jiang 5/1/2017 1 Outline Background semantic segmentation Objective,

More information

Additive Manufacturing Defect Detection using Neural Networks

Additive Manufacturing Defect Detection using Neural Networks Additive Manufacturing Defect Detection using Neural Networks James Ferguson Department of Electrical Engineering and Computer Science University of Tennessee Knoxville Knoxville, Tennessee 37996 Jfergu35@vols.utk.edu

More information

CSE 559A: Computer Vision

CSE 559A: Computer Vision CSE 559A: Computer Vision Fall 2018: T-R: 11:30-1pm @ Lopata 101 Instructor: Ayan Chakrabarti (ayan@wustl.edu). Course Staff: Zhihao Xia, Charlie Wu, Han Liu http://www.cse.wustl.edu/~ayan/courses/cse559a/

More information

Perceptron: This is convolution!

Perceptron: This is convolution! Perceptron: This is convolution! v v v Shared weights v Filter = local perceptron. Also called kernel. By pooling responses at different locations, we gain robustness to the exact spatial location of image

More information

Encoder-Decoder Networks for Semantic Segmentation. Sachin Mehta

Encoder-Decoder Networks for Semantic Segmentation. Sachin Mehta Encoder-Decoder Networks for Semantic Segmentation Sachin Mehta Outline > Overview of Semantic Segmentation > Encoder-Decoder Networks > Results What is Semantic Segmentation? Input: RGB Image Output:

More information

A Deep Learning Approach to Vehicle Speed Estimation

A Deep Learning Approach to Vehicle Speed Estimation A Deep Learning Approach to Vehicle Speed Estimation Benjamin Penchas bpenchas@stanford.edu Tobin Bell tbell@stanford.edu Marco Monteiro marcorm@stanford.edu ABSTRACT Given car dashboard video footage,

More information

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon Deep Learning For Video Classification Presented by Natalie Carlebach & Gil Sharon Overview Of Presentation Motivation Challenges of video classification Common datasets 4 different methods presented in

More information

SSD: Single Shot MultiBox Detector. Author: Wei Liu et al. Presenter: Siyu Jiang

SSD: Single Shot MultiBox Detector. Author: Wei Liu et al. Presenter: Siyu Jiang SSD: Single Shot MultiBox Detector Author: Wei Liu et al. Presenter: Siyu Jiang Outline 1. Motivations 2. Contributions 3. Methodology 4. Experiments 5. Conclusions 6. Extensions Motivation Motivation

More information

Kaggle Data Science Bowl 2017 Technical Report

Kaggle Data Science Bowl 2017 Technical Report Kaggle Data Science Bowl 2017 Technical Report qfpxfd Team May 11, 2017 1 Team Members Table 1: Team members Name E-Mail University Jia Ding dingjia@pku.edu.cn Peking University, Beijing, China Aoxue Li

More information

Martian lava field, NASA, Wikipedia

Martian lava field, NASA, Wikipedia Martian lava field, NASA, Wikipedia Old Man of the Mountain, Franconia, New Hampshire Pareidolia http://smrt.ccel.ca/203/2/6/pareidolia/ Reddit for more : ) https://www.reddit.com/r/pareidolia/top/ Pareidolia

More information

Semantic Segmentation

Semantic Segmentation Semantic Segmentation UCLA:https://goo.gl/images/I0VTi2 OUTLINE Semantic Segmentation Why? Paper to talk about: Fully Convolutional Networks for Semantic Segmentation. J. Long, E. Shelhamer, and T. Darrell,

More information

Know your data - many types of networks

Know your data - many types of networks Architectures Know your data - many types of networks Fixed length representation Variable length representation Online video sequences, or samples of different sizes Images Specific architectures for

More information

Fuzzy Set Theory in Computer Vision: Example 3, Part II

Fuzzy Set Theory in Computer Vision: Example 3, Part II Fuzzy Set Theory in Computer Vision: Example 3, Part II Derek T. Anderson and James M. Keller FUZZ-IEEE, July 2017 Overview Resource; CS231n: Convolutional Neural Networks for Visual Recognition https://github.com/tuanavu/stanford-

More information

CIS680: Vision & Learning Assignment 2.b: RPN, Faster R-CNN and Mask R-CNN Due: Nov. 21, 2018 at 11:59 pm

CIS680: Vision & Learning Assignment 2.b: RPN, Faster R-CNN and Mask R-CNN Due: Nov. 21, 2018 at 11:59 pm CIS680: Vision & Learning Assignment 2.b: RPN, Faster R-CNN and Mask R-CNN Due: Nov. 21, 2018 at 11:59 pm Instructions This is an individual assignment. Individual means each student must hand in their

More information

Machine Learning 13. week

Machine Learning 13. week Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of

More information

Additive Manufacturing Defect Detection using Neural Networks. James Ferguson May 16, 2016

Additive Manufacturing Defect Detection using Neural Networks. James Ferguson May 16, 2016 Additive Manufacturing Defect Detection using Neural Networks James Ferguson May 16, 2016 Outline Introduction Background Edge Detection Methods Results Porosity Detection Methods Results Conclusion /

More information

Dynamic Routing Between Capsules

Dynamic Routing Between Capsules Report Explainable Machine Learning Dynamic Routing Between Capsules Author: Michael Dorkenwald Supervisor: Dr. Ullrich Köthe 28. Juni 2018 Inhaltsverzeichnis 1 Introduction 2 2 Motivation 2 3 CapusleNet

More information

CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning

CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning CS231A Course Project Final Report Sign Language Recognition with Unsupervised Feature Learning Justin Chen Stanford University justinkchen@stanford.edu Abstract This paper focuses on experimenting with

More information

EE-559 Deep learning Networks for semantic segmentation

EE-559 Deep learning Networks for semantic segmentation EE-559 Deep learning 7.4. Networks for semantic segmentation François Fleuret https://fleuret.org/ee559/ Mon Feb 8 3:35:5 UTC 209 ÉCOLE POLYTECHNIQUE FÉDÉRALE DE LAUSANNE The historical approach to image

More information

turning data into dollars

turning data into dollars turning data into dollars Tom s Ten Data Tips November 2008 Neural Networks Neural Networks (NNs) are sometimes considered the epitome of data mining algorithms. Loosely modeled after the human brain (hence

More information

Finding Tiny Faces Supplementary Materials

Finding Tiny Faces Supplementary Materials Finding Tiny Faces Supplementary Materials Peiyun Hu, Deva Ramanan Robotics Institute Carnegie Mellon University {peiyunh,deva}@cs.cmu.edu 1. Error analysis Quantitative analysis We plot the distribution

More information

Spatial Localization and Detection. Lecture 8-1

Spatial Localization and Detection. Lecture 8-1 Lecture 8: Spatial Localization and Detection Lecture 8-1 Administrative - Project Proposals were due on Saturday Homework 2 due Friday 2/5 Homework 1 grades out this week Midterm will be in-class on Wednesday

More information

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro CMU 15-781 Lecture 18: Deep learning and Vision: Convolutional neural networks Teacher: Gianni A. Di Caro DEEP, SHALLOW, CONNECTED, SPARSE? Fully connected multi-layer feed-forward perceptrons: More powerful

More information

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin

More information

An Exploration of Computer Vision Techniques for Bird Species Classification

An Exploration of Computer Vision Techniques for Bird Species Classification An Exploration of Computer Vision Techniques for Bird Species Classification Anne L. Alter, Karen M. Wang December 15, 2017 Abstract Bird classification, a fine-grained categorization task, is a complex

More information

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah Reference Most of the slides are taken from the third chapter of the online book by Michael Nielson: neuralnetworksanddeeplearning.com

More information

Lecture 3: Linear Classification

Lecture 3: Linear Classification Lecture 3: Linear Classification Roger Grosse 1 Introduction Last week, we saw an example of a learning task called regression. There, the goal was to predict a scalar-valued target from a set of features.

More information

Lab 9. Julia Janicki. Introduction

Lab 9. Julia Janicki. Introduction Lab 9 Julia Janicki Introduction My goal for this project is to map a general land cover in the area of Alexandria in Egypt using supervised classification, specifically the Maximum Likelihood and Support

More information

Lecture 7: Semantic Segmentation

Lecture 7: Semantic Segmentation Semantic Segmentation CSED703R: Deep Learning for Visual Recognition (207F) Segmenting images based on its semantic notion Lecture 7: Semantic Segmentation Bohyung Han Computer Vision Lab. bhhanpostech.ac.kr

More information

Notes on Multilayer, Feedforward Neural Networks

Notes on Multilayer, Feedforward Neural Networks Notes on Multilayer, Feedforward Neural Networks CS425/528: Machine Learning Fall 2012 Prepared by: Lynne E. Parker [Material in these notes was gleaned from various sources, including E. Alpaydin s book

More information

Feature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule.

Feature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule. CS 188: Artificial Intelligence Fall 2008 Lecture 24: Perceptrons II 11/24/2008 Dan Klein UC Berkeley Feature Extractors A feature extractor maps inputs to feature vectors Dear Sir. First, I must solicit

More information

Ryerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro

Ryerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro Ryerson University CP8208 Soft Computing and Machine Intelligence Naive Road-Detection using CNNS Authors: Sarah Asiri - Domenic Curro April 24 2016 Contents 1 Abstract 2 2 Introduction 2 3 Motivation

More information

Object Detection on Self-Driving Cars in China. Lingyun Li

Object Detection on Self-Driving Cars in China. Lingyun Li Object Detection on Self-Driving Cars in China Lingyun Li Introduction Motivation: Perception is the key of self-driving cars Data set: 10000 images with annotation 2000 images without annotation (not

More information

Using Machine Learning for Classification of Cancer Cells

Using Machine Learning for Classification of Cancer Cells Using Machine Learning for Classification of Cancer Cells Camille Biscarrat University of California, Berkeley I Introduction Cell screening is a commonly used technique in the development of new drugs.

More information

Mapping of Hierarchical Activation in the Visual Cortex Suman Chakravartula, Denise Jones, Guillaume Leseur CS229 Final Project Report. Autumn 2008.

Mapping of Hierarchical Activation in the Visual Cortex Suman Chakravartula, Denise Jones, Guillaume Leseur CS229 Final Project Report. Autumn 2008. Mapping of Hierarchical Activation in the Visual Cortex Suman Chakravartula, Denise Jones, Guillaume Leseur CS229 Final Project Report. Autumn 2008. Introduction There is much that is unknown regarding

More information

Forecasting Lung Cancer Diagnoses with Deep Learning

Forecasting Lung Cancer Diagnoses with Deep Learning Forecasting Lung Cancer Diagnoses with Deep Learning Data Science Bowl 2017 Technical Report Daniel Hammack April 22, 2017 Abstract We describe our contribution to the 2nd place solution for the 2017 Data

More information

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 12: Deep Reinforcement Learning Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound Lecture 12: Deep Reinforcement Learning Types of Learning Supervised training Learning from the teacher Training data includes

More information

Channel Locality Block: A Variant of Squeeze-and-Excitation

Channel Locality Block: A Variant of Squeeze-and-Excitation Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan

More information

Real-time Object Detection CS 229 Course Project

Real-time Object Detection CS 229 Course Project Real-time Object Detection CS 229 Course Project Zibo Gong 1, Tianchang He 1, and Ziyi Yang 1 1 Department of Electrical Engineering, Stanford University December 17, 2016 Abstract Objection detection

More information

Animations involving numbers

Animations involving numbers 136 Chapter 8 Animations involving numbers 8.1 Model and view The examples of Chapter 6 all compute the next picture in the animation from the previous picture. This turns out to be a rather restrictive

More information

CEA LIST s participation to the Scalable Concept Image Annotation task of ImageCLEF 2015

CEA LIST s participation to the Scalable Concept Image Annotation task of ImageCLEF 2015 CEA LIST s participation to the Scalable Concept Image Annotation task of ImageCLEF 2015 Etienne Gadeski, Hervé Le Borgne, and Adrian Popescu CEA, LIST, Laboratory of Vision and Content Engineering, France

More information

3D model classification using convolutional neural network

3D model classification using convolutional neural network 3D model classification using convolutional neural network JunYoung Gwak Stanford jgwak@cs.stanford.edu Abstract Our goal is to classify 3D models directly using convolutional neural network. Most of existing

More information

MOTION ESTIMATION USING CONVOLUTIONAL NEURAL NETWORKS. Mustafa Ozan Tezcan

MOTION ESTIMATION USING CONVOLUTIONAL NEURAL NETWORKS. Mustafa Ozan Tezcan MOTION ESTIMATION USING CONVOLUTIONAL NEURAL NETWORKS Mustafa Ozan Tezcan Boston University Department of Electrical and Computer Engineering 8 Saint Mary s Street Boston, MA 2215 www.bu.edu/ece Dec. 19,

More information

Overall Description. Goal: to improve spatial invariance to the input data. Translation, Rotation, Scale, Clutter, Elastic

Overall Description. Goal: to improve spatial invariance to the input data. Translation, Rotation, Scale, Clutter, Elastic Philippe Giguère Overall Description Goal: to improve spatial invariance to the input data Translation, Rotation, Scale, Clutter, Elastic How: add a learnable module which explicitly manipulate spatially

More information

Automated Diagnosis of Vertebral Fractures using 2D and 3D Convolutional Networks

Automated Diagnosis of Vertebral Fractures using 2D and 3D Convolutional Networks Automated Diagnosis of Vertebral Fractures using 2D and 3D Convolutional Networks CS189 Final Project Naofumi Tomita Overview Automated diagnosis of osteoporosis-related vertebral fractures is a useful

More information

Deconvolutions in Convolutional Neural Networks

Deconvolutions in Convolutional Neural Networks Overview Deconvolutions in Convolutional Neural Networks Bohyung Han bhhan@postech.ac.kr Computer Vision Lab. Convolutional Neural Networks (CNNs) Deconvolutions in CNNs Applications Network visualization

More information

Weighted Convolutional Neural Network. Ensemble.

Weighted Convolutional Neural Network. Ensemble. Weighted Convolutional Neural Network Ensemble Xavier Frazão and Luís A. Alexandre Dept. of Informatics, Univ. Beira Interior and Instituto de Telecomunicações Covilhã, Portugal xavierfrazao@gmail.com

More information

Feature Extractors. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. The Perceptron Update Rule.

Feature Extractors. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. The Perceptron Update Rule. CS 188: Artificial Intelligence Fall 2007 Lecture 26: Kernels 11/29/2007 Dan Klein UC Berkeley Feature Extractors A feature extractor maps inputs to feature vectors Dear Sir. First, I must solicit your

More information

Multi-Glance Attention Models For Image Classification

Multi-Glance Attention Models For Image Classification Multi-Glance Attention Models For Image Classification Chinmay Duvedi Stanford University Stanford, CA cduvedi@stanford.edu Pararth Shah Stanford University Stanford, CA pararth@stanford.edu Abstract We

More information

DeepBIBX: Deep Learning for Image Based Bibliographic Data Extraction

DeepBIBX: Deep Learning for Image Based Bibliographic Data Extraction DeepBIBX: Deep Learning for Image Based Bibliographic Data Extraction Akansha Bhardwaj 1,2, Dominik Mercier 1, Sheraz Ahmed 1, Andreas Dengel 1 1 Smart Data and Services, DFKI Kaiserslautern, Germany firstname.lastname@dfki.de

More information

Efficient Segmentation-Aided Text Detection For Intelligent Robots

Efficient Segmentation-Aided Text Detection For Intelligent Robots Efficient Segmentation-Aided Text Detection For Intelligent Robots Junting Zhang, Yuewei Na, Siyang Li, C.-C. Jay Kuo University of Southern California Outline Problem Definition and Motivation Related

More information

Convolutional Networks in Scene Labelling

Convolutional Networks in Scene Labelling Convolutional Networks in Scene Labelling Ashwin Paranjape Stanford ashwinpp@stanford.edu Ayesha Mudassir Stanford aysh@stanford.edu Abstract This project tries to address a well known problem of multi-class

More information

Deep neural networks II

Deep neural networks II Deep neural networks II May 31 st, 2018 Yong Jae Lee UC Davis Many slides from Rob Fergus, Svetlana Lazebnik, Jia-Bin Huang, Derek Hoiem, Adriana Kovashka, Why (convolutional) neural networks? State of

More information

Rich feature hierarchies for accurate object detection and semant

Rich feature hierarchies for accurate object detection and semant Rich feature hierarchies for accurate object detection and semantic segmentation Speaker: Yucong Shen 4/5/2018 Develop of Object Detection 1 DPM (Deformable parts models) 2 R-CNN 3 Fast R-CNN 4 Faster

More information

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun Presented by Tushar Bansal Objective 1. Get bounding box for all objects

More information

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes CS 6501: Deep Learning for Computer Graphics Training Neural Networks II Connelly Barnes Overview Preprocessing Initialization Vanishing/exploding gradients problem Batch normalization Dropout Additional

More information

Deep Learning with Tensorflow AlexNet

Deep Learning with Tensorflow   AlexNet Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification

More information

PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D.

PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D. PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D. Rhodes 5/10/17 What is Machine Learning? Machine learning

More information

Structured Prediction using Convolutional Neural Networks

Structured Prediction using Convolutional Neural Networks Overview Structured Prediction using Convolutional Neural Networks Bohyung Han bhhan@postech.ac.kr Computer Vision Lab. Convolutional Neural Networks (CNNs) Structured predictions for low level computer

More information

Convolution Neural Networks for Chinese Handwriting Recognition

Convolution Neural Networks for Chinese Handwriting Recognition Convolution Neural Networks for Chinese Handwriting Recognition Xu Chen Stanford University 450 Serra Mall, Stanford, CA 94305 xchen91@stanford.edu Abstract Convolutional neural networks have been proven

More information

(Refer Slide Time 04:53)

(Refer Slide Time 04:53) Programming and Data Structure Dr.P.P.Chakraborty Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Lecture 26 Algorithm Design -1 Having made a preliminary study

More information

CS 523: Multimedia Systems

CS 523: Multimedia Systems CS 523: Multimedia Systems Angus Forbes creativecoding.evl.uic.edu/courses/cs523 Today - Convolutional Neural Networks - Work on Project 1 http://playground.tensorflow.org/ Convolutional Neural Networks

More information

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2016

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2016 CPSC 340: Machine Learning and Data Mining Deep Learning Fall 2016 Assignment 5: Due Friday. Assignment 6: Due next Friday. Final: Admin December 12 (8:30am HEBB 100) Covers Assignments 1-6. Final from

More information

Deep Learning for Computer Vision with MATLAB By Jon Cherrie

Deep Learning for Computer Vision with MATLAB By Jon Cherrie Deep Learning for Computer Vision with MATLAB By Jon Cherrie 2015 The MathWorks, Inc. 1 Deep learning is getting a lot of attention "Dahl and his colleagues won $22,000 with a deeplearning system. 'We

More information

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU, Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image

More information

Improving Positron Emission Tomography Imaging with Machine Learning David Fan-Chung Hsu CS 229 Fall

Improving Positron Emission Tomography Imaging with Machine Learning David Fan-Chung Hsu CS 229 Fall Improving Positron Emission Tomography Imaging with Machine Learning David Fan-Chung Hsu (fcdh@stanford.edu), CS 229 Fall 2014-15 1. Introduction and Motivation High- resolution Positron Emission Tomography

More information

DEFECT INSPECTION FROM SCRATCH TO PRODUCTION. Andrew Liu, Ryan Shen Deep Learning Solution Architect

DEFECT INSPECTION FROM SCRATCH TO PRODUCTION. Andrew Liu, Ryan Shen Deep Learning Solution Architect DEFECT INSPECTION FROM SCRATCH TO PRODUCTION Andrew Liu, Ryan Shen Deep Learning Solution Architect Defect Inspection and its challenges AGENDA NGC Docker images Model set up - Unet Data preparation -

More information

Final Report: Classification of Plankton Classes By Tae Ho Kim and Saaid Haseeb Arshad

Final Report: Classification of Plankton Classes By Tae Ho Kim and Saaid Haseeb Arshad Final Report: Classification of Plankton Classes By Tae Ho Kim and Saaid Haseeb Arshad Table of Contents 1. Project Overview a. Problem Statement b. Data c. Overview of the Two Stages of Implementation

More information

Classifying Depositional Environments in Satellite Images

Classifying Depositional Environments in Satellite Images Classifying Depositional Environments in Satellite Images Alex Miltenberger and Rayan Kanfar Department of Geophysics School of Earth, Energy, and Environmental Sciences Stanford University 1 Introduction

More information

COMPRESSING U-NET USING KNOWLEDGE DISTILLATION

COMPRESSING U-NET USING KNOWLEDGE DISTILLATION COMPRESSING U-NET USING KNOWLEDGE DISTILLATION Karttikeya Mangalam Master s Semester Project Computer Vision Lab EPFL, Switzerland Mentored By: Dr. Mathieu Salzmann 17 th January 2018 Introduction Objective

More information

Pose estimation using a variety of techniques

Pose estimation using a variety of techniques Pose estimation using a variety of techniques Keegan Go Stanford University keegango@stanford.edu Abstract Vision is an integral part robotic systems a component that is needed for robots to interact robustly

More information

Study of Residual Networks for Image Recognition

Study of Residual Networks for Image Recognition Study of Residual Networks for Image Recognition Mohammad Sadegh Ebrahimi Stanford University sadegh@stanford.edu Hossein Karkeh Abadi Stanford University hosseink@stanford.edu Abstract Deep neural networks

More information

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University. Visualizing and Understanding Convolutional Networks Christopher Pennsylvania State University February 23, 2015 Some Slide Information taken from Pierre Sermanet (Google) presentation on and Computer

More information

Measuring Aristic Similarity of Paintings

Measuring Aristic Similarity of Paintings Measuring Aristic Similarity of Paintings Jay Whang Stanford SCPD jaywhang@stanford.edu Buhuang Liu Stanford SCPD buhuang@stanford.edu Yancheng Xiao Stanford SCPD ycxiao@stanford.edu Abstract In this project,

More information

arxiv: v1 [cs.cv] 16 Nov 2018

arxiv: v1 [cs.cv] 16 Nov 2018 Improving Rotated Text Detection with Rotation Region Proposal Networks Jing Huang 1, Viswanath Sivakumar 1, Mher Mnatsakanyan 1,2 and Guan Pang 1 1 Facebook Inc. 2 University of California, Berkeley November

More information

Plankton Classification Using ConvNets

Plankton Classification Using ConvNets Plankton Classification Using ConvNets Abhinav Rastogi Stanford University Stanford, CA arastogi@stanford.edu Haichuan Yu Stanford University Stanford, CA haichuan@stanford.edu Abstract We present the

More information

Detecting and Recognizing Text in Natural Images using Convolutional Networks

Detecting and Recognizing Text in Natural Images using Convolutional Networks Detecting and Recognizing Text in Natural Images using Convolutional Networks Aditya Srinivas Timmaraju, Vikesh Khanna Stanford University Stanford, CA - 94305 adityast@stanford.edu, vikesh@stanford.edu

More information

Introduction to Deep Learning for Facial Understanding Part III: Regional CNNs

Introduction to Deep Learning for Facial Understanding Part III: Regional CNNs Introduction to Deep Learning for Facial Understanding Part III: Regional CNNs Raymond Ptucha, Rochester Institute of Technology, USA Tutorial-9 May 19, 218 www.nvidia.com/dli R. Ptucha 18 1 Fair Use Agreement

More information

Convolution Neural Network for Traditional Chinese Calligraphy Recognition

Convolution Neural Network for Traditional Chinese Calligraphy Recognition Convolution Neural Network for Traditional Chinese Calligraphy Recognition Boqi Li Mechanical Engineering Stanford University boqili@stanford.edu Abstract script. Fig. 1 shows examples of the same TCC

More information

Inception Network Overview. David White CS793

Inception Network Overview. David White CS793 Inception Network Overview David White CS793 So, Leonardo DiCaprio dreams about dreaming... https://m.media-amazon.com/images/m/mv5bmjaxmzy3njcxnf5bml5banbnxkftztcwnti5otm0mw@@._v1_sy1000_cr0,0,675,1 000_AL_.jpg

More information

Convolutional Neural Networks for Eye Tracking Algorithm

Convolutional Neural Networks for Eye Tracking Algorithm Convolutional Neural Networks for Eye Tracking Algorithm Jonathan Griffin Stanford University jgriffi2@stanford.edu Andrea Ramirez Stanford University aramire9@stanford.edu Abstract Eye tracking is an

More information

Toward Scale-Invariance and Position-Sensitive Region Proposal Networks

Toward Scale-Invariance and Position-Sensitive Region Proposal Networks Toward Scale-Invariance and Position-Sensitive Region Proposal Networks Hsueh-Fu Lu [0000 0003 1885 3805], Xiaofei Du [0000 0002 0071 8093], and Ping-Lin Chang [0000 0002 3341 8425] Umbo Computer Vision

More information

All You Want To Know About CNNs. Yukun Zhu

All You Want To Know About CNNs. Yukun Zhu All You Want To Know About CNNs Yukun Zhu Deep Learning Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image from http://imgur.com/ Deep Learning Image

More information

Segmenting Lesions in Multiple Sclerosis Patients James Chen, Jason Su

Segmenting Lesions in Multiple Sclerosis Patients James Chen, Jason Su Segmenting Lesions in Multiple Sclerosis Patients James Chen, Jason Su Radiologists and researchers spend countless hours tediously segmenting white matter lesions to diagnose and study brain diseases.

More information

YOLO 9000 TAEWAN KIM

YOLO 9000 TAEWAN KIM YOLO 9000 TAEWAN KIM DNN-based Object Detection R-CNN MultiBox SPP-Net DeepID-Net NoC Fast R-CNN DeepBox MR-CNN 2013.11 Faster R-CNN YOLO AttentionNet DenseBox SSD Inside-OutsideNet(ION) G-CNN NIPS 15

More information

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Tianyu Wang Australia National University, Colledge of Engineering and Computer Science u@anu.edu.au Abstract. Some tasks,

More information

Lecture note 7: Playing with convolutions in TensorFlow

Lecture note 7: Playing with convolutions in TensorFlow Lecture note 7: Playing with convolutions in TensorFlow CS 20SI: TensorFlow for Deep Learning Research (cs20si.stanford.edu) Prepared by Chip Huyen ( huyenn@stanford.edu ) This lecture note is an unfinished

More information

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR

3D Convolutional Neural Networks for Landing Zone Detection from LiDAR 3D Convolutional Neural Networks for Landing Zone Detection from LiDAR Daniel Mataruna and Sebastian Scherer Presented by: Sabin Kafle Outline Introduction Preliminaries Approach Volumetric Density Mapping

More information

Lung nodule detection by using. Deep Learning

Lung nodule detection by using. Deep Learning VRIJE UNIVERSITEIT AMSTERDAM RESEARCH PAPER Lung nodule detection by using Deep Learning Author: Thomas HEENEMAN Supervisor: Dr. Mark HOOGENDOORN Msc. Business Analytics Department of Mathematics Faculty

More information

CS231N Course Project Report Classifying Shadowgraph Images of Planktons Using Convolutional Neural Networks

CS231N Course Project Report Classifying Shadowgraph Images of Planktons Using Convolutional Neural Networks CS231N Course Project Report Classifying Shadowgraph Images of Planktons Using Convolutional Neural Networks Shane Soh Stanford University shanesoh@stanford.edu 1. Introduction In this final project report,

More information

Detecting Thoracic Diseases from Chest X-Ray Images Binit Topiwala, Mariam Alawadi, Hari Prasad { topbinit, malawadi, hprasad

Detecting Thoracic Diseases from Chest X-Ray Images Binit Topiwala, Mariam Alawadi, Hari Prasad { topbinit, malawadi, hprasad CS 229, Fall 2017 1 Detecting Thoracic Diseases from Chest X-Ray Images Binit Topiwala, Mariam Alawadi, Hari Prasad { topbinit, malawadi, hprasad }@stanford.edu Abstract Radiologists have to spend time

More information

A Novel Representation and Pipeline for Object Detection

A Novel Representation and Pipeline for Object Detection A Novel Representation and Pipeline for Object Detection Vishakh Hegde Stanford University vishakh@stanford.edu Manik Dhar Stanford University dmanik@stanford.edu Abstract Object detection is an important

More information

Convolutional Neural Network based Medical Imaging Segmentation: Recent Progress and Challenges. Jiaxing Tan

Convolutional Neural Network based Medical Imaging Segmentation: Recent Progress and Challenges. Jiaxing Tan Convolutional Neural Network based Medical Imaging Segmentation: Recent Progress and Challenges Jiaxing Tan Road Map Introduction CNN based Models Encoder-Decoder based Models GAN Based Models Some Challenges

More information

DB2 is a complex system, with a major impact upon your processing environment. There are substantial performance and instrumentation changes in

DB2 is a complex system, with a major impact upon your processing environment. There are substantial performance and instrumentation changes in DB2 is a complex system, with a major impact upon your processing environment. There are substantial performance and instrumentation changes in versions 8 and 9. that must be used to measure, evaluate,

More information

Lecture 37: ConvNets (Cont d) and Training

Lecture 37: ConvNets (Cont d) and Training Lecture 37: ConvNets (Cont d) and Training CS 4670/5670 Sean Bell [http://bbabenko.tumblr.com/post/83319141207/convolutional-learnings-things-i-learned-by] (Unrelated) Dog vs Food [Karen Zack, @teenybiscuit]

More information

Deep Learning for Computer Vision II

Deep Learning for Computer Vision II IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L

More information

Deconvolution Networks

Deconvolution Networks Deconvolution Networks Johan Brynolfsson Mathematical Statistics Centre for Mathematical Sciences Lund University December 6th 2016 1 / 27 Deconvolution Neural Networks 2 / 27 Image Deconvolution True

More information

Final Report: Smart Trash Net: Waste Localization and Classification

Final Report: Smart Trash Net: Waste Localization and Classification Final Report: Smart Trash Net: Waste Localization and Classification Oluwasanya Awe oawe@stanford.edu Robel Mengistu robel@stanford.edu December 15, 2017 Vikram Sreedhar vsreed@stanford.edu Abstract Given

More information

Character Recognition

Character Recognition Character Recognition 5.1 INTRODUCTION Recognition is one of the important steps in image processing. There are different methods such as Histogram method, Hough transformation, Neural computing approaches

More information