CAP 6412 Advanced Computer Vision
|
|
- Paul Scott
- 5 years ago
- Views:
Transcription
1 CAP 6412 Advanced Computer Vision Boqing Gong Feb 11, 2016
2 Today Administrivia Neural networks & Backpropagation (Part VII) Edge detection, by Goran
3 Next week: CNN & videos Tuesday (02/16) Abdullah Jamal Thursday (02/18) Amar Kelu Nair [Optical flow] Fischer, Philipp, Alexey Dosovitskiy, Eddy Ilg, Philip Häusser, Caner Hazırbaş, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, and Thomas Brox. "Flownet: Learning optical flow with convolutional networks." arxiv preprint arxiv: (2015). & Secondary papers [Pose estimation] Pfister, Tomas, James Charles, and Andrew Zisserman. Flowing convnets for human pose estimation in videos. In Proceedings of the IEEE International Conference on Computer Vision, pp & Secondary papers
4 Project 1: Due in two weeks (02/28) If you choose option 2, your own project Deadline for discussion & approval: 02/11/2016 (This Thursday) See instructions on how to prepare the slides for discussion
5 Upload slides before or after class See Paper Presentation on UCF webcourse Sharing your slides Refer to the originals sources of images, figures, etc. in your slides Convert them to a PDF file Upload the PDF file to Paper Presentation after your presentation
6 Today Administrivia Neural networks & Backpropagation (Part VI) Image super-resolution, by Jose
7 Training algorithm INPUT: BSD (training image Xn, edge annotations Yn1, Yn5 ), n=1,2, N Configurate network: Trim VGG-net (See right for network specification) Generate gold target Yn for each image Xn: Majority vote: 5 annotated edge maps Yn1, Yn5 à one edge map Yn Duplicate Yn to have gold side output Augment data: Rotate images by 16 angles Flip images Tune hyper-parameters: mini-batch size (10), learning rate (1e-6), loss weight (1), momentum (0.9), initialization filter (0), initialization fusion (1/5), weight decay (2e-4), #iteration 10,000 Train the neural network with those hyper-parameters and augmented data OUTPUT: A well-trained network
8 Test INPUT: A natural image, the trained network OUTPUT: Predicted edge maps
9 Holistically-Nested Edge Detection Authors: Saining Xie, Zhuowen Tu, UC San Diego In Proceedings of the IEEE International Conference on Computer Vision, December Presented at UCF Advance Computer Vision Class by: Goran Igic, February 11 th 2016.
10 About Paper and important links Paper was presented on ICCV 2015 conference which was held December th 2015 In Santiago Chile. It was announced on the second day (TUE DEC 15 th ) that paper got Marr Prize honorable mention. The author Xie s web site: The code used in the paper is open sourced and published at: The same page has links to other repositories including CAFFE, modified-caffe for HED. The paper used Piotr's Structured Forest matlab toolbox available here
11 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
12 Motivation of Research To address fundamental and important vision problem: The Edge Detection The problem was studied for years in Image Processing, Camera Vision, 3D Camera Vision and Robotics There are plenty of solutions : Early Pioneering Methods: Sobel, Zerro Crossing, and Canny detector; Information Theory Implementation: Statistical Edges, Pb, and gpb; Learning Based Methods: BEL, Multiscale, Sketch Tokens, and structured Edges The newest wave of CNN based methods: N4-fields, Deep-Contour, Deep Edge, and CSCNN
13 Edge Detection Classic Approach There are three basic types of intensity discontinuities in digital images: points, lines, and edges. The most common way to look for discontinuities is to run a mask through the image. The response R of the mask in any point of the image is given by: 9 R wizi, for the 3x3 mask; i1 The isolated point is detected when Line detection R T, where T is treshold; R R, for all j i; Points and lines are important in the image segmentation, edge detection is the most common approach. i j Convolution and Edge Detection : Computational Photography. Alexei Efros, CMU, Fall 2005.
14 Effects of noise, solution: smooth first Convolution and Edge Detection : Computational Photography. Alexei Efros, CMU, Fall 2005.
15 Derivative theorem of convolution Convolution and Edge Detection : Computational Photography. Alexei Efros, CMU, Fall 2005.
16 Laplacian of Gaussian, 2D edge detection filters Convolution and Edge Detection : Computational Photography. Alexei Efros, CMU, Fall 2005.
17 Canny Edge Detector (1) Authors compare their work to the 1986 edge detector CANNY First image needs to be converted to gray scale:
18 Canny Edge Detector (2) The threshold is auto-detected T=[0.05,0.125] Standard deviation of smoothing filter is =2.0 =4.0
19 Canny Edge Detector (3) =8.0
20 Classical Methods On Canny Edge Detector from Matlab Help: The Canny method, finds edges by looking for local maxima of the gradient of I. The gradient is calculated using the derivative of a Gaussian filter. The method uses two thresholds, to detect strong and weak edges, and includes the weak edges in the output only if they are connected to strong edges. For more information in regards with classical methods see chapter 10 of the book: E.C.Gonzalez, R.E.Woods, S.L.Eddins, Digital Image Processing Using Matlab, 2 nd edition 2010
21 Information Theory on the top of features Methods, statistical edges, Pb and gpb: 1. P. Arbelaez, M. Maire, C. Fowlkes, and J. Malik. Contour detection and hierarchical image segmentation. PAMI,2011, Contour detector combines multiple local cues into a globalization framework based on spectral clustering. It utilize segmentation algorithm for transforming the output of any contour detector into a hierarchical region tree. This paper is also important because introduces contour benchmark test.
22 Information Theory on the top of features
23 Learning Based Models Structured edges, SE: P. Doll ar and C. L. Zitnick. Fast edge detection using structured forests. PAMI, Patches of edges exhibit well-known forms of local structure, such as straight lines or T-junctions. This paper takes advantage of the structure present in local image patches to learn both an accurate and computationally efficient edge detector. The paper formulates the problem of predicting local edge masks in a structured learning framework applied to random decision forests.
24 Learning Based Models
25 Learning Based Models As input, the method takes an image that may contain multiple channels, such as an RGB or RGBD image. The task is to label each pixel with a binary variable indicating whether the pixel contains an edge or not. The labels within a small image patch are highly interdependent, providing a promising candidate problem for our structured forest approach. Method needs segmented training images, in which the boundaries between the segments correspond to contours. To train decision trees, the mapping is defined: :Y Z Ensemble model, Random forests achieve robust results by combining the output of multiple trees.
26 Models based on Convolutional Neural Networks Models with Integrated Automatic hierarchical feature learning, N4-fields, Deep Contour, Deep Edge, CSCNN Holistically-nested Edge Detection (HED) Holistic the system takes an image as input and directly produces the edge map as output. Nested the system produces edge maps as side outputs,
27 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
28 Problem Statement Paper needs to address two important issues of edge detection vision problem: 1. Holistic Image Training and Prediction, used in image to-image classification, CNN 2. Nested Multiscale and Multilevel Feature learning, deeply-supervised nets to guide early predictions, deep layer supervision to guide early classification results.
29 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
30 Main Contributions of the paper, Develops an end-to-end edge detection system, holistically-nested edge detection system (HED) Holistic it aims to train and predict edges in an image-toimage fashion; Nested the path along with each prediction is common to each of these edge maps; The system has integrated learning of hierarchical features
31 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
32 Approach Outline HED (Holistic Edge Detection) uses DSN (Deep Supervised Network) training to fine tune the VGG network (Visual Geometry Group Oxford University) for the task of boundary detection. The principle behind DSN is classifier stacking adapted to deep learning, where each layer is informed about the final objective. VGG Net has great depth (16 convolutional layers), great density (stride-1 convolutional kernels), and multiple stages (five 2-sttrade down sampling layers). J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In CVPR, K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. CoRR, abs/ , 2014.
33 Figure #2: Illustration of different multiscale-deep learning architecture configurations (a) (b) multi-stream architecture; Multiple parallel networks with different parameter numbers and receptive field sizes, corresponding to multiple scales. Same input data on all streams, concatenated feature responses are fed to the global output skip-layer net architecture; There is primary stream. The contributions from layers are added to this stream. There is only one output loss function. The edge prediction is better if there are multiple predictions to combine.
34 Figure #2: Illustration of different multiscale-deep learning architecture configurations (c) A single model running on multi-scale inputs; Used in non-deep learning based methods. Not very efficient prediction. (d) separate training of different networks; multiple independent networks with different depths; different output loss layers. Method has higher resource demanding.
35 Figure #2: Illustration of different multiscale-deep learning architecture configurations (e) Holistically-nested architectures, where multiple side outputs are added. This is a single stream deep network with multiple side outputs. The multiple scale prediction is possible.
36 Long at all DCN/VGG Network J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In CVPR, K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. CoRR, abs/ , 2014.
37 Approach Outline Authors of the paper changed net: Side output layer is connected to the last convolutional layer The last stage of VGGN is cut out including 5 th pooling layer J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In CVPR, K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. CoRR, abs/ , 2014.
38 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
39 Details of Proposed Approach Training Phase Input training data set: S X, Y, n 1,.., N n n n n n where raw image sample is X x, j 1,..., X ; coresponding ground truth binary edge map for image X : Y y, j 1,..., X, y 0,1 ; n j n j n j n Each image is treated holistically and independently, so sufix n can be omitted. The goal is to have a network that learns the features from witch it is possible to produce edge map approaching the ground truth. n
40 Details of Proposed Approach W is collection of all standard network layer parameters. Network has M side-output layers, each of them is associated with the classifier, with corresponding weights: (1) M w w,..., w The objective function of HED: M m W,w m L l W,w, where l is the image-level loss function for side outputs, side m side side m1 Loss function is computed over all pixels in training image. X = x j n, j = 1,..., X ; Y = y j n, j = 1,..., X n, y j n 0,1 ;
41 Details of Proposed Approach The typical natural image, has the distribution of edge / non-edge pixels as that 90% of pixels is non-edge. Simpler strategy (vs other papers) is used here to automatically balance loss between positive / negative classes, class balancing weight b on perpixel term basis. Class-balanced cross-entropy loss function: l W,w b log Pr y 1 X ; W,w (1 b) log Pr y 0 X ; W,w m m m m side j j jy jy Where b Y / Y ; and 1 b Y / Y ; Y m m y j X a j is edge ground truth label set; is non-edge ground truth label set; Pr 1 ; W,w 0,1 is computed using sigmoid function. on the activation value at pixel j ; Y
42 Details of Proposed Approach Edge map predictions on each side output layer: m m m ^ ^ ^ m Y, where, 1,..., side A side Aside a j j Y, are activations of the side output of layer m To directly utilize side output prediction, weighted-fusion layer is added. The fusion weight is leaned during the training. The fusion layer loss function: L fuse ^ ^ ^ m M W,w,h Dist Y, Y fuse ; Y fuse h side ; m 1 m A The fusion weight h h,..., h, 1 M
43 Details of Proposed Approach ^ Dist Y, Y fuse is the distance between the fused prediction and the ground truth label map; it is cross entropy loss; Minimized Objective Function: * W,w,h L L h arg min( W,w)+ W,w, ; side fuse
44 Details of Proposed Approach Testing phase: X is given image, edge-map prediction is obtain from both the side output layers and the weight fusion layer 1 M ^ ^ ^ * Y,,...,, W,w,h fuse Y side Y side CNN X, The final unified output: ^ ^ ^ Y AverageY, Y HED fuse side M,..., Y side. 1 ^
45 Figure #3 HED network architecture for edge detection. Error backpropagation path is highlighted. Side-output layers are inserted after convolutional layers. Deep supervision is in place at each side-output layer. The HED outputs are multiscale and multilevel. Side-output-plane size getting smaller and the repetitive field size become larger. The weight fusion layer is added to automatically learn how to combine outputs from multiple scales
46 Details of Proposed Approach, C1-C5 intermediate layer of DCNN, E1-E5 are side layers, loss function is red. Simplified presentation of HED/DCN architecture, from the work SURPASSING HUMANS IN BOUNDARY DETECTION USING DEEP LEARNING Iasonas Kokkinos, Center for Visual Computing CentraleSupelec and INRIA
47 Figure #2, results from HED
48 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
49 Experiments The work uses publically available Caffe Library, publically available implementation of FCN and DSN. The whole network is fine tuned from an initialization with the pre-trained VGG-16 Net model The hyper-parameters (mini-batch size, learning rate, loss-weight m for each side-output layer) are tuned. Data augmentation, different pooling functions, in-network bilinear interpolation is used with no much improvement. Running time, training takes about 7 hours on a single NVIDIA K40 GPU. From an image 320x480 pixels, HED produces the final edge map for about 400mS
50 Arbelaez at all.. P. Arbelaez, M. Maire, C. Fowlkes, and J. Malik. Contour detection and hierarchical image segmentation. PAMI, 33(5): , This paper proposed benchmark test: One option is to regard the segment boundaries as contours and evaluate them as such. One might argue that the boundary benchmark favors contour detectors over segmentation methods, since the former are not burdened with the constraint of producing closed curves. Leading contour detection approaches are ranked according to their maximum F-measure 2Precision Recall Precision+Recall with respect to human ground-truth boundaries. ODS fixed threshold contour - measure Boundary quality. Detector performance are measured in terms of precision, the fraction of true positives, and recall, the fraction of ground-truth boundary pixels detected. The global F measure, or harmonic mean of precision and recall at the optimal detector threshold, provides a summary score. ODS is for entire data set, OIS is per image, AP ia average precision.
51 BSDS500 dataset Results on BSDS500 data set, 200 training, 100validations, and 200 test images. The ground truth contour is manually annotated. Edge detection is evaluated using three standard measures: Fixed contour threshold (ODS) Per-image best threshold (OIS) Average precision (AP) HED performed better than CNNbased detectors
52 BSDS500 dataset - Fixed contour threshold (ODS), Per-image best threshold (OIS), Average precision (AP) Table 3. Results of single and averaged side output in HED on the BSDS 500 dataset.
53 NYUDv2 dataset Precision/recall curves on NYUD dataset. Holistically-nested edge detection (HED) trained with RGB and HHA features achieves the best result (ODS=.746). See
54 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
55 Related Work Lee, Chen-Yu, Xie, Saining, Gallagher, Patrick W., Zhang, Zhengyou, and Tu, Zhuowen. Deeply-supervised nets. In Proc. AISTATS, Iasonas Kokkinos "SURPASSING HUMANS IN BOUNDARY DETECTION USING DEEP LEARNING Liu, Ziwei, Li, Xiaoxiao, Luo, Ping, Loy, Chen Change, and Tang, Xiaoou. Semantic image segmentation via deep parsing network. arxiv preprint arxiv: , Yizhou, Yu, Chaowei, Fang, Zicheng, Liao, Piecewise Flat Embedding for Image Segmentation
56 Content Motivation of Research Problem Statement Main Contributions of the paper Conclusion Strengths and Weakness of the Paper Overall Rating Future Directions Approach Outline Details of Proposed Approach Experiments Related Work
57 Conclusion - Strength and Weakness The known problem is: Consensus sampling, the ground truth is duplicated at each side-output layer and side output is down sampled to its original scale. The mismatch exist and noise may cause convergence issue in the high-level side outputs even with the help of pretrained model. Develops an end-to-end edge detection system, holistically-nested edge detection system (HED) Holistic it aims to train and predict edges in an image-to-image fashion; Nested the path along with each prediction is common to each of these edge maps; The system has integrated learning of hierarchical features Experimental work done well
58 Conclusion Overall Rating The paper is very fresh and novel, it is the major part of the current wave of CNN implementations, and machine learning in Vision I have read the following works for ICCV 2016 The paper got Marr Prize honorable mention at the last ICCV 2015 conference which was held December th 2015 In Santiago Chile. The experimental results showing step forward The code is publically available Illustration of different multiscale-deep learning architecture configurations is given My rating is 0
59 Conclusion - Future Directions DCN can outperform humans (Iasonas Kokkinos said). There is certainly a great potential in usage of DCN. The system idea is not very complex. The training can be improved. There are works with including multiresolution architecture.
RSRN: Rich Side-output Residual Network for Medial Axis Detection
RSRN: Rich Side-output Residual Network for Medial Axis Detection Chang Liu, Wei Ke, Jianbin Jiao, and Qixiang Ye University of Chinese Academy of Sciences, Beijing, China {liuchang615, kewei11}@mails.ucas.ac.cn,
More informationA Deep Learning Approach to Vehicle Speed Estimation
A Deep Learning Approach to Vehicle Speed Estimation Benjamin Penchas bpenchas@stanford.edu Tobin Bell tbell@stanford.edu Marco Monteiro marcorm@stanford.edu ABSTRACT Given car dashboard video footage,
More informationCAP 6412 Advanced Computer Vision
CAP 6412 Advanced Computer Vision http://www.cs.ucf.edu/~bgong/cap6412.html Boqing Gong April 21st, 2016 Today Administrivia Free parameters in an approach, model, or algorithm? Egocentric videos by Aisha
More informationEdge Detection Using Convolutional Neural Network
Edge Detection Using Convolutional Neural Network Ruohui Wang (B) Department of Information Engineering, The Chinese University of Hong Kong, Hong Kong, China wr013@ie.cuhk.edu.hk Abstract. In this work,
More informationCAP 6412 Advanced Computer Vision
CAP 6412 Advanced Computer Vision http://www.cs.ucf.edu/~bgong/cap6412.html Boqing Gong Feb 04, 2016 Today Administrivia Attention Modeling in Image Captioning, by Karan Neural networks & Backpropagation
More informationHolistically-Nested Edge Detection
Noname manuscript No. (will be inserted by the editor) Holistically-Nested Edge Detection Saining Xie Zhuowen Tu Received: date / Accepted: date Abstract We develop a new edge detection algorithm that
More informationStructured Prediction using Convolutional Neural Networks
Overview Structured Prediction using Convolutional Neural Networks Bohyung Han bhhan@postech.ac.kr Computer Vision Lab. Convolutional Neural Networks (CNNs) Structured predictions for low level computer
More informationFast Edge Detection Using Structured Forests
Fast Edge Detection Using Structured Forests Piotr Dollár, C. Lawrence Zitnick [1] Zhihao Li (zhihaol@andrew.cmu.edu) Computer Science Department Carnegie Mellon University Table of contents 1. Introduction
More informationEdge Detection. Computer Vision Shiv Ram Dubey, IIIT Sri City
Edge Detection Computer Vision Shiv Ram Dubey, IIIT Sri City Previous two classes: Image Filtering Spatial domain Smoothing, sharpening, measuring texture * = FFT FFT Inverse FFT = Frequency domain Denoising,
More informationSupplementary Material for Zoom and Learn: Generalizing Deep Stereo Matching to Novel Domains
Supplementary Material for Zoom and Learn: Generalizing Deep Stereo Matching to Novel Domains Jiahao Pang 1 Wenxiu Sun 1 Chengxi Yang 1 Jimmy Ren 1 Ruichao Xiao 1 Jin Zeng 1 Liang Lin 1,2 1 SenseTime Research
More informationDense Volume-to-Volume Vascular Boundary Detection
Dense Volume-to-Volume Vascular Boundary Detection Jameson Merkow 1, David Kriegman 1, Alison Marsden 2, and Zhuowen Tu 1 1 University of California, San Diego. 2 Stanford University Abstract. In this
More informationEdge and Texture. CS 554 Computer Vision Pinar Duygulu Bilkent University
Edge and Texture CS 554 Computer Vision Pinar Duygulu Bilkent University Filters for features Previously, thinking of filtering as a way to remove or reduce noise Now, consider how filters will allow us
More informationPASCAL Boundaries: A Semantic Boundary Dataset with A Deep Semantic Boundary Detector
PASCAL Boundaries: A Semantic Boundary Dataset with A Deep Semantic Boundary Detector Vittal Premachandran 1, Boyan Bonev 2, Xiaochen Lian 2, and Alan Yuille 1 1 Department of Computer Science, Johns Hopkins
More informationEncoder-Decoder Networks for Semantic Segmentation. Sachin Mehta
Encoder-Decoder Networks for Semantic Segmentation Sachin Mehta Outline > Overview of Semantic Segmentation > Encoder-Decoder Networks > Results What is Semantic Segmentation? Input: RGB Image Output:
More informationDirect Multi-Scale Dual-Stream Network for Pedestrian Detection Sang-Il Jung and Ki-Sang Hong Image Information Processing Lab.
[ICIP 2017] Direct Multi-Scale Dual-Stream Network for Pedestrian Detection Sang-Il Jung and Ki-Sang Hong Image Information Processing Lab., POSTECH Pedestrian Detection Goal To draw bounding boxes that
More informationSemantic Segmentation
Semantic Segmentation UCLA:https://goo.gl/images/I0VTi2 OUTLINE Semantic Segmentation Why? Paper to talk about: Fully Convolutional Networks for Semantic Segmentation. J. Long, E. Shelhamer, and T. Darrell,
More information3 October, 2013 MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos
Machine Learning for Computer Vision 1 3 October, 2013 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Department of Applied Mathematics Ecole Centrale
More informationLecture 7: Semantic Segmentation
Semantic Segmentation CSED703R: Deep Learning for Visual Recognition (207F) Segmenting images based on its semantic notion Lecture 7: Semantic Segmentation Bohyung Han Computer Vision Lab. bhhanpostech.ac.kr
More informationEdge detection. Winter in Kraków photographed by Marcin Ryczek
Edge detection Winter in Kraków photographed by Marcin Ryczek Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, edges carry most of the semantic and shape information
More informationFinding Tiny Faces Supplementary Materials
Finding Tiny Faces Supplementary Materials Peiyun Hu, Deva Ramanan Robotics Institute Carnegie Mellon University {peiyunh,deva}@cs.cmu.edu 1. Error analysis Quantitative analysis We plot the distribution
More informationFully Convolutional Networks for Semantic Segmentation
Fully Convolutional Networks for Semantic Segmentation Jonathan Long* Evan Shelhamer* Trevor Darrell UC Berkeley Chaim Ginzburg for Deep Learning seminar 1 Semantic Segmentation Define a pixel-wise labeling
More informationMOTION ESTIMATION USING CONVOLUTIONAL NEURAL NETWORKS. Mustafa Ozan Tezcan
MOTION ESTIMATION USING CONVOLUTIONAL NEURAL NETWORKS Mustafa Ozan Tezcan Boston University Department of Electrical and Computer Engineering 8 Saint Mary s Street Boston, MA 2215 www.bu.edu/ece Dec. 19,
More informationDeep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks
Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin
More informationRecap from Monday. Visualizing Networks Caffe overview Slides are now online
Recap from Monday Visualizing Networks Caffe overview Slides are now online Today Edges and Regions, GPB Fast Edge Detection Using Structured Forests Zhihao Li Holistically-Nested Edge Detection Yuxin
More informationUsing RGB, Depth, and Thermal Data for Improved Hand Detection
Using RGB, Depth, and Thermal Data for Improved Hand Detection Rachel Luo, Gregory Luppescu Department of Electrical Engineering Stanford University {rsluo, gluppes}@stanford.edu Abstract Hand detection
More informationEfficient Segmentation-Aided Text Detection For Intelligent Robots
Efficient Segmentation-Aided Text Detection For Intelligent Robots Junting Zhang, Yuewei Na, Siyang Li, C.-C. Jay Kuo University of Southern California Outline Problem Definition and Motivation Related
More informationDeep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia
Deep learning for dense per-pixel prediction Chunhua Shen The University of Adelaide, Australia Image understanding Classification error Convolution Neural Networks 0.3 0.2 0.1 Image Classification [Krizhevsky
More informationDeep learning for object detection. Slides from Svetlana Lazebnik and many others
Deep learning for object detection Slides from Svetlana Lazebnik and many others Recent developments in object detection 80% PASCAL VOC mean0average0precision0(map) 70% 60% 50% 40% 30% 20% 10% Before deep
More informationDeconvolutions in Convolutional Neural Networks
Overview Deconvolutions in Convolutional Neural Networks Bohyung Han bhhan@postech.ac.kr Computer Vision Lab. Convolutional Neural Networks (CNNs) Deconvolutions in CNNs Applications Network visualization
More informationComputer Vision I. Announcement. Corners. Edges. Numerical Derivatives f(x) Edge and Corner Detection. CSE252A Lecture 11
Announcement Edge and Corner Detection Slides are posted HW due Friday CSE5A Lecture 11 Edges Corners Edge is Where Change Occurs: 1-D Change is measured by derivative in 1D Numerical Derivatives f(x)
More informationRicher Convolutional Features for Edge Detection
Richer Convolutional Features for Edge Detection Yun Liu 1 Ming-Ming Cheng 1 Xiaowei Hu 1 Kai Wang 1 Xiang Bai 2 1 Nankai University 2 HUST https://mmcheng.net/rcfedge/ Abstract In this paper, we propose
More informationMachine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,
Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image
More informationReview of Filtering. Filtering in frequency domain
Review of Filtering Filtering in frequency domain Can be faster than filtering in spatial domain (for large filters) Can help understand effect of filter Algorithm: 1. Convert image and filter to fft (fft2
More informationREGION AVERAGE POOLING FOR CONTEXT-AWARE OBJECT DETECTION
REGION AVERAGE POOLING FOR CONTEXT-AWARE OBJECT DETECTION Kingsley Kuan 1, Gaurav Manek 1, Jie Lin 1, Yuan Fang 1, Vijay Chandrasekhar 1,2 Institute for Infocomm Research, A*STAR, Singapore 1 Nanyang Technological
More informationarxiv: v2 [cs.cv] 9 Apr 2017
Richer Convolutional Features for Edge Detection Yun Liu 1 Ming-Ming Cheng 1 Xiaowei Hu 1 Kai Wang 1 Xiang Bai 2 1 Nankai University 2 HUST http://mmcheng.net/rcfedge/ arxiv:1612.02103v2 [cs.cv] 9 Apr
More informationLinear Span Network for Object Skeleton Detection
Linear Span Network for Object Skeleton Detection Chang Liu, Wei Ke, Fei Qin, and Qixiang Ye ( ) University of Chinese Academy of Sciences, Beijing, China {liuchang615, kewei11}@mails.ucas.ac.cn, {fqin1982,
More informationBoundaries and Sketches
Boundaries and Sketches Szeliski 4.2 Computer Vision James Hays Many slides from Michael Maire, Jitendra Malek Today s lecture Segmentation vs Boundary Detection Why boundaries / Grouping? Recap: Canny
More informationChannel Locality Block: A Variant of Squeeze-and-Excitation
Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan
More informationJOINT DETECTION AND SEGMENTATION WITH DEEP HIERARCHICAL NETWORKS. Zhao Chen Machine Learning Intern, NVIDIA
JOINT DETECTION AND SEGMENTATION WITH DEEP HIERARCHICAL NETWORKS Zhao Chen Machine Learning Intern, NVIDIA ABOUT ME 5th year PhD student in physics @ Stanford by day, deep learning computer vision scientist
More informationDeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution and Fully Connected CRFs
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution and Fully Connected CRFs Zhipeng Yan, Moyuan Huang, Hao Jiang 5/1/2017 1 Outline Background semantic segmentation Objective,
More informationTeam G-RMI: Google Research & Machine Intelligence
Team G-RMI: Google Research & Machine Intelligence Alireza Fathi (alirezafathi@google.com) Nori Kanazawa, Kai Yang, George Papandreou, Tyler Zhu, Jonathan Huang, Vivek Rathod, Chen Sun, Kevin Murphy, et
More informationarxiv: v1 [cs.cv] 31 Mar 2016
Object Boundary Guided Semantic Segmentation Qin Huang, Chunyang Xia, Wenchao Zheng, Yuhang Song, Hao Xu and C.-C. Jay Kuo arxiv:1603.09742v1 [cs.cv] 31 Mar 2016 University of Southern California Abstract.
More informationEdges and Binary Image Analysis April 12 th, 2018
4/2/208 Edges and Binary Image Analysis April 2 th, 208 Yong Jae Lee UC Davis Previously Filters allow local image neighborhood to influence our description and features Smoothing to reduce noise Derivatives
More informationConvolutional Networks in Scene Labelling
Convolutional Networks in Scene Labelling Ashwin Paranjape Stanford ashwinpp@stanford.edu Ayesha Mudassir Stanford aysh@stanford.edu Abstract This project tries to address a well known problem of multi-class
More informationMask R-CNN. By Kaiming He, Georgia Gkioxari, Piotr Dollar and Ross Girshick Presented By Aditya Sanghi
Mask R-CNN By Kaiming He, Georgia Gkioxari, Piotr Dollar and Ross Girshick Presented By Aditya Sanghi Types of Computer Vision Tasks http://cs231n.stanford.edu/ Semantic vs Instance Segmentation Image
More informationQuo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018
Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Outline: Introduction Action classification architectures
More informationConvolution Neural Networks for Chinese Handwriting Recognition
Convolution Neural Networks for Chinese Handwriting Recognition Xu Chen Stanford University 450 Serra Mall, Stanford, CA 94305 xchen91@stanford.edu Abstract Convolutional neural networks have been proven
More informationFaster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun Presented by Tushar Bansal Objective 1. Get bounding box for all objects
More informationObject Detection Based on Deep Learning
Object Detection Based on Deep Learning Yurii Pashchenko AI Ukraine 2016, Kharkiv, 2016 Image classification (mostly what you ve seen) http://tutorial.caffe.berkeleyvision.org/caffe-cvpr15-detection.pdf
More informationReconstructive Sparse Code Transfer for Contour Detection and Semantic Labeling
Reconstructive Sparse Code Transfer for Contour Detection and Semantic Labeling Michael Maire 1,2 Stella X. Yu 3 Pietro Perona 2 1 TTI Chicago 2 California Institute of Technology 3 University of California
More informationDeepEdge: A Multi-Scale Bifurcated Deep Network for Top-Down Contour Detection
DeepEdge: A Multi-Scale Bifurcated Deep Network for Top-Down Contour Detection Gedas Bertasius University of Pennsylvania gberta@seas.upenn.edu Jianbo Shi University of Pennsylvania jshi@seas.upenn.edu
More informationDIGITAL IMAGE PROCESSING
The image part with relationship ID rid2 was not found in the file. DIGITAL IMAGE PROCESSING Lecture 6 Wavelets (cont), Lines and edges Tammy Riklin Raviv Electrical and Computer Engineering Ben-Gurion
More informationDeep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon
Deep Learning For Video Classification Presented by Natalie Carlebach & Gil Sharon Overview Of Presentation Motivation Challenges of video classification Common datasets 4 different methods presented in
More informationEE-559 Deep learning Networks for semantic segmentation
EE-559 Deep learning 7.4. Networks for semantic segmentation François Fleuret https://fleuret.org/ee559/ Mon Feb 8 3:35:5 UTC 209 ÉCOLE POLYTECHNIQUE FÉDÉRALE DE LAUSANNE The historical approach to image
More informationObject Recognition II
Object Recognition II Linda Shapiro EE/CSE 576 with CNN slides from Ross Girshick 1 Outline Object detection the task, evaluation, datasets Convolutional Neural Networks (CNNs) overview and history Region-based
More informationObject detection using Region Proposals (RCNN) Ernest Cheung COMP Presentation
Object detection using Region Proposals (RCNN) Ernest Cheung COMP790-125 Presentation 1 2 Problem to solve Object detection Input: Image Output: Bounding box of the object 3 Object detection using CNN
More informationPredicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture David Eigen, Rob Fergus
Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture David Eigen, Rob Fergus Presented by: Rex Ying and Charles Qi Input: A Single RGB Image Estimate
More informationSupplemental Material for End-to-End Learning of Video Super-Resolution with Motion Compensation
Supplemental Material for End-to-End Learning of Video Super-Resolution with Motion Compensation Osama Makansi, Eddy Ilg, and Thomas Brox Department of Computer Science, University of Freiburg 1 Computation
More informationSupplementary Material: Unconstrained Salient Object Detection via Proposal Subset Optimization
Supplementary Material: Unconstrained Salient Object via Proposal Subset Optimization 1. Proof of the Submodularity According to Eqns. 10-12 in our paper, the objective function of the proposed optimization
More informationReal-time Object Detection CS 229 Course Project
Real-time Object Detection CS 229 Course Project Zibo Gong 1, Tianchang He 1, and Ziyi Yang 1 1 Department of Electrical Engineering, Stanford University December 17, 2016 Abstract Objection detection
More informationTRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK
TRANSPARENT OBJECT DETECTION USING REGIONS WITH CONVOLUTIONAL NEURAL NETWORK 1 Po-Jen Lai ( 賴柏任 ), 2 Chiou-Shann Fuh ( 傅楸善 ) 1 Dept. of Electrical Engineering, National Taiwan University, Taiwan 2 Dept.
More informationWhat is an edge? Paint. Depth discontinuity. Material change. Texture boundary
EDGES AND TEXTURES The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their
More informationCAP 6412 Advanced Computer Vision
CAP 6412 Advanced Computer Vision http://www.cs.ucf.edu/~bgong/cap6412.html Boqing Gong April 7th, 2016 Today Administrivia A guest lecture by David Hill on LSTM Attribute in computer vision, by Abdullah
More informationComputer Vision I. Announcements. Fourier Tansform. Efficient Implementation. Edge and Corner Detection. CSE252A Lecture 13.
Announcements Edge and Corner Detection HW3 assigned CSE252A Lecture 13 Efficient Implementation Both, the Box filter and the Gaussian filter are separable: First convolve each row of input image I with
More informationProceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong
, March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong TABLE I CLASSIFICATION ACCURACY OF DIFFERENT PRE-TRAINED MODELS ON THE TEST DATA
More information3D Shape Analysis with Multi-view Convolutional Networks. Evangelos Kalogerakis
3D Shape Analysis with Multi-view Convolutional Networks Evangelos Kalogerakis 3D model repositories [3D Warehouse - video] 3D geometry acquisition [KinectFusion - video] 3D shapes come in various flavors
More informationSpatial Localization and Detection. Lecture 8-1
Lecture 8: Spatial Localization and Detection Lecture 8-1 Administrative - Project Proposals were due on Saturday Homework 2 due Friday 2/5 Homework 1 grades out this week Midterm will be in-class on Wednesday
More informationEdge detection. Winter in Kraków photographed by Marcin Ryczek
Edge detection Winter in Kraków photographed by Marcin Ryczek Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the image
More informationarxiv: v1 [cs.cv] 29 Sep 2016
arxiv:1609.09545v1 [cs.cv] 29 Sep 2016 Two-stage Convolutional Part Heatmap Regression for the 1st 3D Face Alignment in the Wild (3DFAW) Challenge Adrian Bulat and Georgios Tzimiropoulos Computer Vision
More informationDeep Crisp Boundaries
Deep Crisp Boundaries Yupei Wang 1,2, Xin Zhao 1,2, Kaiqi Huang 1,2,3 1 CRIPAC & NLPR, CASIA 2 University of Chinese Academy of Sciences 3 CAS Center for Excellence in Brain Science and Intelligence Technology
More informationMULTI-SCALE OBJECT DETECTION WITH FEATURE FUSION AND REGION OBJECTNESS NETWORK. Wenjie Guan, YueXian Zou*, Xiaoqun Zhou
MULTI-SCALE OBJECT DETECTION WITH FEATURE FUSION AND REGION OBJECTNESS NETWORK Wenjie Guan, YueXian Zou*, Xiaoqun Zhou ADSPLAB/Intelligent Lab, School of ECE, Peking University, Shenzhen,518055, China
More informationSEMANTIC SEGMENTATION AVIRAM BAR HAIM & IRIS TAL
SEMANTIC SEGMENTATION AVIRAM BAR HAIM & IRIS TAL IMAGE DESCRIPTIONS IN THE WILD (IDW-CNN) LARGE KERNEL MATTERS (GCN) DEEP LEARNING SEMINAR, TAU NOVEMBER 2017 TOPICS IDW-CNN: Improving Semantic Segmentation
More informationGlobal Probability of Boundary
Global Probability of Boundary Learning to Detect Natural Image Boundaries Using Local Brightness, Color, and Texture Cues Martin, Fowlkes, Malik Using Contours to Detect and Localize Junctions in Natural
More informationCS 4495 Computer Vision. Linear Filtering 2: Templates, Edges. Aaron Bobick. School of Interactive Computing. Templates/Edges
CS 4495 Computer Vision Linear Filtering 2: Templates, Edges Aaron Bobick School of Interactive Computing Last time: Convolution Convolution: Flip the filter in both dimensions (right to left, bottom to
More informationMask R-CNN. presented by Jiageng Zhang, Jingyao Zhan, Yunhan Ma
Mask R-CNN presented by Jiageng Zhang, Jingyao Zhan, Yunhan Ma Mask R-CNN Background Related Work Architecture Experiment Mask R-CNN Background Related Work Architecture Experiment Background From left
More information(Deep) Learning for Robot Perception and Navigation. Wolfram Burgard
(Deep) Learning for Robot Perception and Navigation Wolfram Burgard Deep Learning for Robot Perception (and Navigation) Lifeng Bo, Claas Bollen, Thomas Brox, Andreas Eitel, Dieter Fox, Gabriel L. Oliveira,
More informationUnFlow: Unsupervised Learning of Optical Flow with a Bidirectional Census Loss
UnFlow: Unsupervised Learning of Optical Flow with a Bidirectional Census Loss AAAI 2018, New Orleans, USA Simon Meister, Junhwa Hur, and Stefan Roth Department of Computer Science, TU Darmstadt 2 Deep
More informationYOLO9000: Better, Faster, Stronger
YOLO9000: Better, Faster, Stronger Date: January 24, 2018 Prepared by Haris Khan (University of Toronto) Haris Khan CSC2548: Machine Learning in Computer Vision 1 Overview 1. Motivation for one-shot object
More informationarxiv: v2 [cs.cv] 29 Nov 2016 Abstract
Object Detection Free Instance Segmentation With Labeling Transformations Long Jin 1, Zeyu Chen 1, Zhuowen Tu 2,1 1 Dept. of CSE and 2 Dept. of CogSci, University of California, San Diego 9500 Gilman Drive,
More informationCamera-based Vehicle Velocity Estimation using Spatiotemporal Depth and Motion Features
Camera-based Vehicle Velocity Estimation using Spatiotemporal Depth and Motion Features Moritz Kampelmuehler* kampelmuehler@student.tugraz.at Michael Mueller* michael.g.mueller@student.tugraz.at Christoph
More informationProject 3 Q&A. Jonathan Krause
Project 3 Q&A Jonathan Krause 1 Outline R-CNN Review Error metrics Code Overview Project 3 Report Project 3 Presentations 2 Outline R-CNN Review Error metrics Code Overview Project 3 Report Project 3 Presentations
More informationEdge Detection CSC 767
Edge Detection CSC 767 Edge detection Goal: Identify sudden changes (discontinuities) in an image Most semantic and shape information from the image can be encoded in the edges More compact than pixels
More informationDigital Image Processing COSC 6380/4393
Digital Image Processing COSC 6380/4393 Lecture 21 Nov 16 th, 2017 Pranav Mantini Ack: Shah. M Image Processing Geometric Transformation Point Operations Filtering (spatial, Frequency) Input Restoration/
More informationCEA LIST s participation to the Scalable Concept Image Annotation task of ImageCLEF 2015
CEA LIST s participation to the Scalable Concept Image Annotation task of ImageCLEF 2015 Etienne Gadeski, Hervé Le Borgne, and Adrian Popescu CEA, LIST, Laboratory of Vision and Content Engineering, France
More informationCS231N Section. Video Understanding 6/1/2018
CS231N Section Video Understanding 6/1/2018 Outline Background / Motivation / History Video Datasets Models Pre-deep learning CNN + RNN 3D convolution Two-stream What we ve seen in class so far... Image
More informationMULTI-LEVEL 3D CONVOLUTIONAL NEURAL NETWORK FOR OBJECT RECOGNITION SAMBIT GHADAI XIAN LEE ADITYA BALU SOUMIK SARKAR ADARSH KRISHNAMURTHY
MULTI-LEVEL 3D CONVOLUTIONAL NEURAL NETWORK FOR OBJECT RECOGNITION SAMBIT GHADAI XIAN LEE ADITYA BALU SOUMIK SARKAR ADARSH KRISHNAMURTHY Outline Object Recognition Multi-Level Volumetric Representations
More informationComputer Vision Lecture 16
Computer Vision Lecture 16 Deep Learning for Object Categorization 14.01.2016 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar registration period
More informationFinal Report: Smart Trash Net: Waste Localization and Classification
Final Report: Smart Trash Net: Waste Localization and Classification Oluwasanya Awe oawe@stanford.edu Robel Mengistu robel@stanford.edu December 15, 2017 Vikram Sreedhar vsreed@stanford.edu Abstract Given
More informationKaggle Data Science Bowl 2017 Technical Report
Kaggle Data Science Bowl 2017 Technical Report qfpxfd Team May 11, 2017 1 Team Members Table 1: Team members Name E-Mail University Jia Ding dingjia@pku.edu.cn Peking University, Beijing, China Aoxue Li
More informationSSD: Single Shot MultiBox Detector. Author: Wei Liu et al. Presenter: Siyu Jiang
SSD: Single Shot MultiBox Detector Author: Wei Liu et al. Presenter: Siyu Jiang Outline 1. Motivations 2. Contributions 3. Methodology 4. Experiments 5. Conclusions 6. Extensions Motivation Motivation
More informationCS 2770: Computer Vision. Edges and Segments. Prof. Adriana Kovashka University of Pittsburgh February 21, 2017
CS 2770: Computer Vision Edges and Segments Prof. Adriana Kovashka University of Pittsburgh February 21, 2017 Edges vs Segments Figure adapted from J. Hays Edges vs Segments Edges More low-level Don t
More informationAn Exploration of Computer Vision Techniques for Bird Species Classification
An Exploration of Computer Vision Techniques for Bird Species Classification Anne L. Alter, Karen M. Wang December 15, 2017 Abstract Bird classification, a fine-grained categorization task, is a complex
More informationEdge Detection. Today s reading. Cipolla & Gee on edge detection (available online) From Sandlot Science
Edge Detection From Sandlot Science Today s reading Cipolla & Gee on edge detection (available online) Project 1a assigned last Friday due this Friday Last time: Cross-correlation Let be the image, be
More informationMulti-View 3D Object Detection Network for Autonomous Driving
Multi-View 3D Object Detection Network for Autonomous Driving Xiaozhi Chen, Huimin Ma, Ji Wan, Bo Li, Tian Xia CVPR 2017 (Spotlight) Presented By: Jason Ku Overview Motivation Dataset Network Architecture
More informationDetection of a Single Hand Shape in the Foreground of Still Images
CS229 Project Final Report Detection of a Single Hand Shape in the Foreground of Still Images Toan Tran (dtoan@stanford.edu) 1. Introduction This paper is about an image detection system that can detect
More informationVisual features detection based on deep neural network in autonomous driving tasks
430 Fomin I., Gromoshinskii D., Stepanov D. Visual features detection based on deep neural network in autonomous driving tasks Ivan Fomin, Dmitrii Gromoshinskii, Dmitry Stepanov Computer vision lab Russian
More informationMicroscopy Cell Counting with Fully Convolutional Regression Networks
Microscopy Cell Counting with Fully Convolutional Regression Networks Weidi Xie, J. Alison Noble, Andrew Zisserman Department of Engineering Science, University of Oxford,UK Abstract. This paper concerns
More informationMartian lava field, NASA, Wikipedia
Martian lava field, NASA, Wikipedia Old Man of the Mountain, Franconia, New Hampshire Pareidolia http://smrt.ccel.ca/203/2/6/pareidolia/ Reddit for more : ) https://www.reddit.com/r/pareidolia/top/ Pareidolia
More informationRyerson University CP8208. Soft Computing and Machine Intelligence. Naive Road-Detection using CNNS. Authors: Sarah Asiri - Domenic Curro
Ryerson University CP8208 Soft Computing and Machine Intelligence Naive Road-Detection using CNNS Authors: Sarah Asiri - Domenic Curro April 24 2016 Contents 1 Abstract 2 2 Introduction 2 3 Motivation
More informationDeeply Cascaded Networks
Deeply Cascaded Networks Eunbyung Park Department of Computer Science University of North Carolina at Chapel Hill eunbyung@cs.unc.edu 1 Introduction After the seminal work of Viola-Jones[15] fast object
More information