3D Photography. Marc Pollefeys, Torsten Sattler. Spring 2015
|
|
- Clinton Morgan
- 5 years ago
- Views:
Transcription
1 3D Photography Marc Pollefeys, Torsten Sattler Spring 2015
2 Schedule (tentative) Feb 16 Feb 23 Mar 2 Mar 9 Mar 16 Mar 23 Mar 30 Apr 6 Apr 13 Apr 20 Apr 27 May 4 May 11 May 18 May 25 Introduction Geometry, Camera Model, Calibration Features, Tracking / Matching Project Proposals by Students Structure from Motion (SfM) + 2 papers Dense Correspondence (stereo / optical flow) + 2 papers Bundle Adjustment & SLAM + 2 papers Easter Project Updates Multi-View Stereo & Volumetric Modeling + 2 papers 3D Modeling with Depth Sensors + 2 papers 3D Scene Understanding + 2 papers 4D Video & Dynamic Scenes + 2 papers Final Demos Pentecost
3 3D Photography Class 2 Features & Correspondences feature extraction, image descriptors, feature matching, feature tracking Chapters 4, 8 in Szeliski Book
4 Overview Local Features Intro Invariant Feature Extraction & Matching Feature Tracking
5 Today: Features & Correspondences M 2D-3D 2D-3D m i 2D-2D m i+1 Correspondences are at the heart of 3D reconstruction from images
6
7 Feature Matching vs. Tracking Image-to-image correspondences are key to passive triangulation-based 3D reconstruction Extract features independently and then match by comparing descriptors Extract features in first images and then try to find same feature back in next view What is a good feature?
8 Comparing Image Regions Compare intensities pixel-by-pixel I(x,y) I (x,y) Dissimilarity measures Sum of Square Differences SSD x y I (x,y) I(x,y) 2
9 Feature Points / Keypoints Required properties: Well-localized Stable across views, repeatable (i.e. same 3D point should be extracted as feature for neighboring viewpoints)
10 Feature Point Extraction Find points (local image patches) that differ as much as possible from all neighboring points homogeneous edge corner
11 Finding Stable Features Measure uniqueness of candidate Approximate SSD for small displacement Δ SSD w(x i ) I(x i ) I(x i ) 2 i w(x i ) I(x i ) I I x y i 2 I w(x i ) T x I x I y 2 I x I y I i y possible weights 2 I(x i ) T M I x I x
12 Finding Stable Features homogeneous edge corner Suitable feature positions should maximize i.e. maximize smallest eigenvalue of M
13 Harris Corner Detector Use small local window: Maximize cornerness : Only use local maxima, subpixel accuracy through second order surface fitting Select strongest features over whole image and over each tile (e.g. 1000/image, 2/tile)
14 Simple Matching For each corner in image 1 find the corner in image 2 that is most similar and vice-versa Only compare geometrically compatible points (epipolar constraints, ) Keep mutual best matches What transformations does this work for?
15 Comparing Image Regions Compare intensities pixel-by-pixel I(x,y) I (x,y) Dissimilarity measures Sum of Square Differences SSD x y I (x,y) I(x,y) 2
16 Comparing Image Regions Compare intensities pixel-by-pixel I(x,y) I (x,y) Similarity measures Zero-mean Normalized Cross Correlation NCC N I,I N I',I N I',I' N I,I I(x,y) I x y I (x,y) I
17 Feature Matching Example What transformations does this work for? What level of transformation do we need?
18 Overview Local Features Intro Invariant Feature Extraction & Matching Feature Tracking
19 Wide Baseline Matching Requirement to cope with larger variations between images Translation, rotation, scaling Perspective foreshortening Non-diffuse reflections Illumination geometric photometric changes transformations
20 Invariant Detectors Rotation invariant Scale invariant Affine invariant (approximately invariant w.r.t. perspective/viewpoint)
21 2D Transformations of a Local Patch e.g. VIP e.g. Harris corners & NCC e.g. MSER e.g. SIFT
22 David Lowe s SIFT Features (Lowe, ICCV99) Detector + descriptor (later) Recover features with position, orientation and scale
23 Position Look for strong responses of DOG filter (Difference-Of-Gaussian) Only consider local maxima
24 Scale Look for strong responses of DOG filter (Difference-Of-Gaussian) over scale space Only consider local maxima in both position and scale Fit quadratic around maxima for subpixel
25 Minimum Contrast and Cornerness
26 Invariants So Far Additive and multiplicative changes in brightness Difference-Of-Gaussian! Translation Scale Rotation is missing!
27 Orientation Create histogram of local gradient directions computed at selected scale Assign canonical orientation at peak of smoothed histogram Each key specifies stable 2D coordinates (x, y, scale, orientation)
28 2D Transformations of a Local Patch e.g. VIP e.g. Harris corners & NCC e.g. MSER e.g. SIFT
29 Affine Invariant Features Perspective effects can locally be approximated by affine transformation
30 Extreme Wide Baseline Matching Find Correspondences between these images using the MSER Detector [Matas 02]
31 Maximally Stable Extremal Regions Local regions, not points!
32 Maximally Stable Extremal Regions Extremal regions: Much brighter than surrounding Use intensity threshold
33 Maximally Stable Extremal Regions Extremal regions: OR: Much darker than surrounding Use intensity threshold
34 Maximally Stable Extremal Regions Regions: Connected pixels at some threshold Region size = #pixels Maximally stable: Region constant near some threshold
35 A Sample Feature
36 A Sample Feature T is maximally stable with respect to surrounding
37 From Regions To Ellipses Compute center of gravity Compute Scatter (PCA / Ellipsoid)
38 From Regions To Ellipses Ellipse abstracts from pixels! Geometric representation: position/size/shape
39 Comparing Features Idea: Normalize to default position/size/shape e.g. circle of radius 16 pixels
40 Normalize ellipse to circle 2D rotation still unresolved
41 Same approach as for SIFT: Compute histogram of local gradients Find dominant orientation in histogram Rotate local patch into dominant orientation
42 Each normalized patch obtained from single image!
43 MSER Features Detect sets of pixels brighter/darker than surr. Fit elliptical shape to pixel set Warp image so that ellipse becomes circle Rotate to dominant gradient direction [other constructions possible as well]
44 MSER Features - Invariants Constant brightness changes (additive and multiplicative) Rotation, translation, scale Affine transformations Affine normalization of feature leads to similar patches in different views!
45 Local Patch Descriptors Small misalignments Brightness change More tolerant comparison needed!
46 Lowe s SIFT Descriptor Gradient Orientation/Magnitude Gradient Magnitude Ignore pixel values, use only local gradients Gradient direction more important than positions Partition into sectors to retain spatial information
47 Lowe s SIFT Descriptor Thresholded image gradients are sampled over 16x16 array of locations in scale space Create array of orientation histograms 8 orientations x 4x4 histogram array = 128D
48 Descriptor Computation Gradient Orientation/Magnitude Orientation Histogram per Sector Quantize gradient orientations in 45 steps Bin gradients into histogram Weight of gradient = gradient magnitude Concatenate histograms
49 Descriptor Computation Quantization errors: Small differences can lead to different bins! 22 quantized/rounded to 0 23 quantized/rounded to 45 Can be caused by Small errors in feature position, size, shape, or orientation Image noise Descriptor must be robust against this!
50 Hard Binning Classical (closest bin) If orientation is 3 different, all measurements go to second bin! Sudden change in histogram from ( ) to ( )
51 Soft Binning Soft weights: Bin Correctness Soft Binning If orientation is 3 different, descriptor changes only gradually!
52 Descriptor Invariance? Translation, scale, affine deformations: Inherited from detector Rotation: Align bins / histograms with dominant orientation of patch Uniform intensity / illumination changes? Adding constant value does not affect gradient Normalize vector to handle multiplicative changes Robustness to non-uniform changes Idea: Change affects gradient magnitude but not direction After normalization: Clip descriptor entries to be 0.2 Renormalize! But no invariance!
53 Memory Consumption After normalization: floating point values Need 128 x 4 = 512 bytes per descriptor Reduce memory consumption by quantization Use only 1 byte per descriptor entry 128 bytes Simple quantization: Multiply with 512 and round to integer value Little impact on descriptor quality For comparison: Storing raw grey values for 11x11 patch requires 121 bytes
54 For each region: 128-dim. descriptor How to find correspondences?
55 Descriptor Matching - Scenario I Two images in a dense image sequence: Think about maximum movement d (e.g. 50 pixel) Search in a window +/- d of old position Compare descriptors (Euclidean distance), choose nearest neighbor
56 Descriptor Matching - Scenario II Two arbitrary images / wide baseline Brute force search (e.g. GPU) OR: Approximate nearest neighbor search in descriptor space (kd-tree) OR: Find small set of matches, predict others
57 kd-tree-based Matching Descriptor Space kd-tree Iteratively split dimension with largest variance Matching: Traverse tree based on splits Depth 30 1B descriptors (~122MB) Curse of dimensionality: Need to visit all leaves to guarantee finding nearest neighbor Approximate search: Visit N leaf nodes
58 Searching Descriptor Space Learn important dimensions of 128D space for a given scene, e.g. PCA or LDA Project descriptors to important dimensions, use kd-tree
59 Descriptor Matching Spatial Search Window: Requires/exploits good prediction Can avoid far away similar-looking features Good for sequences Descriptor Space: Initial tree setup Fast lookup for huge amounts of features More sophisticated outlier detection required Good for asymmetric (offline/online) problems, registration, initialization, object recognition, wide baseline matching
60 Correspondence Verification Features have only very local view Mismatches How to detect? Discard matches with low similarity Discard non-distinctive matches through Lowe s ratio test / SIFT ratio test Check for bi-directional consistency Geometric verification, e.g., RANSAC
61 Object Detection / Pose Estimation using single MSER Affine feature: position(2x), shape(3x), orientation(1x) 6 DOF, more than 2D simple point!
62 2D Transformations of a Local Patch e.g. VIP e.g. Harris corners & NCC In practice hardly observable for small patches! e.g. MSER e.g. SIFT
63 Viewpoint Invariant Patches (VIP) (Wu, CVPR 08) Use known planar geometry to remove perspective distortion Alternative: Use vanishing points to rectify patch
64 Feature Detectors And Descriptors Detector Find interesting regions (position/size/shape) Assign dominant gradient orientation Normalize regions Descriptor Compute signature upon normalized region Behave smoothly in presence of distortions: brightness changes / normalization inaccuracies
65 Recent Variants and Accelerations (much faster, but this usually comes at a price) BRIEF[Calonder10]: binary descriptor (tests=position a darker than b), compare descriptors by XOR (Hamming) + POPCNT RIFF[Takacs10]: CENSURE + gradients tangential/radial ORB[Rublee11] FAST+orientation BRISK[Leutenegger11] FAST+scale+BRIEF FREAK[Alahi12] FAST + daisy -BRIEF Lucid[Ziegler12]: sort intensities D-BRIEF[Trzcinski12]:Box-Filter+learned projection+brief LDA HASH[Strecha12]: binary tests on descriptor
66 Some Feature Resources Affine feature evaluation + binaries: SIFT, MSER & much more (mostly Matlab): SURF: GPU-SIFT: DAISY (dense descriptors) FAST[er] corner detector (simple but ) OpenCV (MSER, binary descriptors, matching, )
67 Require Features: Local Normalization + Robust Descriptors (Nearly) planarity accross feature in 3D Detectable regions, enough structure / texture Allow us to handle Changing perspective/illumination/scale Scenarios with many occlusions Different reasoning (descriptor vectors)
68 Features: Local Normalization + Robust Descriptors But... Not all is perfect No true invariance against viewpoint changes SIFT: ~30 degrees Affine covariant, e.g., MSER, up to maybe 60 degrees Higher invariance = slower computation, fewer features Invariance vs. Descriptiveness: Similar descriptors even when regions are not! What level of invariance / speed / properties do YOU really need?
69 Overview Local Features Intro Invariant Feature Extraction & Matching Feature Tracking
70 Feature Tracking Identify features and track them over video Small difference between frames potential large difference overall Standard approach: KLT (Kanade-Lukas-Tomasi)
71 Tracking Corners Through Video
72 Good Features to Track Use same window in feature selection as for tracking itself (see first part of lecture) Compute motion assuming it is small differentiate: linear system Affine is also possible, but a bit harder (6x6 instead of 2x2)
73 Example affine translation Simple displacement is sufficient between consecutive frames, but not to compare to reference template
74 Example translation affine
75 Good features to keep tracking Perform affine alignment between first and last frame Stop tracking features with too large errors
76 Intensity Linearization Brightness constancy assumption (small motion) 1D example possibility for iterative refinement
77 Intensity Linearization Brightness constancy assumption (small motion) 2D example (1 constraint) (2 unknowns)? the aperture problem isophote I(t+1)=I isophote I(t)=I Barberpole illusion (image source: Wikipedia)
78 Intensity Linearization How to deal with aperture problem? (3 constraints if color gradients are different) Assume neighbors have same displacement
79 Lucas-Kanade Assume neighbors have same displacement least-squares:
80 Revisiting the Small Motion Assumption Is this motion small enough? Probably not it s much larger than one pixel (1 st order Taylor not sufficient) How might we solve this problem? * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003
81 Reduce the Resolution! * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003
82 Gaussian pyramid of image I t-1 Gaussian pyramid of image I Coarse-to-Fine Optical Flow Estimation slides from Bradsky and Thrun u=1.25 pixels u=2.5 pixels u=5 pixels image I t-1 u=10 pixels image I
83 Coarse-to-Fine Optical Flow Estimation slides from Bradsky and Thrun run iterative L-K run iterative L-K warp & upsample... image J I t-1 Gaussian pyramid of image I t-1 image I Gaussian pyramid of image I
84 Next week: Project Proposals Reminder: Submit your proposal until Friday!
3D Vision. Viktor Larsson. Spring 2019
3D Vision Viktor Larsson Spring 2019 Schedule Feb 18 Feb 25 Mar 4 Mar 11 Mar 18 Mar 25 Apr 1 Apr 8 Apr 15 Apr 22 Apr 29 May 6 May 13 May 20 May 27 Introduction Geometry, Camera Model, Calibration Features,
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationFeature Tracking and Optical Flow
Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve Seitz, Rick Szeliski,
More informationPeripheral drift illusion
Peripheral drift illusion Does it work on other animals? Computer Vision Motion and Optical Flow Many slides adapted from J. Hays, S. Seitz, R. Szeliski, M. Pollefeys, K. Grauman and others Video A video
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationFeature Based Registration - Image Alignment
Feature Based Registration - Image Alignment Image Registration Image registration is the process of estimating an optimal transformation between two or more images. Many slides from Alexei Efros http://graphics.cs.cmu.edu/courses/15-463/2007_fall/463.html
More informationOutline 7/2/201011/6/
Outline Pattern recognition in computer vision Background on the development of SIFT SIFT algorithm and some of its variations Computational considerations (SURF) Potential improvement Summary 01 2 Pattern
More informationSIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014
SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image
More informationImage processing and features
Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry
More informationECE Digital Image Processing and Introduction to Computer Vision
ECE592-064 Digital Image Processing and Introduction to Computer Vision Depart. of ECE, NC State University Instructor: Tianfu (Matt) Wu Spring 2017 Recap, SIFT Motion Tracking Change Detection Feature
More informationComputer Vision for HCI. Topics of This Lecture
Computer Vision for HCI Interest Points Topics of This Lecture Local Invariant Features Motivation Requirements, Invariances Keypoint Localization Features from Accelerated Segment Test (FAST) Harris Shi-Tomasi
More informationAugmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit
Augmented Reality VU Computer Vision 3D Registration (2) Prof. Vincent Lepetit Feature Point-Based 3D Tracking Feature Points for 3D Tracking Much less ambiguous than edges; Point-to-point reprojection
More informationMulti-stable Perception. Necker Cube
Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix
More informationNinio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,
Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29, 1209-1217. CS 4495 Computer Vision A. Bobick Sparse to Dense Correspodence Building Rome in
More informationOptical flow and tracking
EECS 442 Computer vision Optical flow and tracking Intro Optical flow and feature tracking Lucas-Kanade algorithm Motion segmentation Segments of this lectures are courtesy of Profs S. Lazebnik S. Seitz,
More informationCS 4495 Computer Vision Motion and Optic Flow
CS 4495 Computer Vision Aaron Bobick School of Interactive Computing Administrivia PS4 is out, due Sunday Oct 27 th. All relevant lectures posted Details about Problem Set: You may *not* use built in Harris
More information3D Photography: Active Ranging, Structured Light, ICP
3D Photography: Active Ranging, Structured Light, ICP Kalin Kolev, Marc Pollefeys Spring 2013 http://cvg.ethz.ch/teaching/2013spring/3dphoto/ Schedule (tentative) Feb 18 Feb 25 Mar 4 Mar 11 Mar 18 Mar
More informationLecture 16: Computer Vision
CS4442/9542b: Artificial Intelligence II Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field
More informationVisual Tracking (1) Tracking of Feature Points and Planar Rigid Objects
Intelligent Control Systems Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects Shingo Kagami Graduate School of Information Sciences, Tohoku University swk(at)ic.is.tohoku.ac.jp http://www.ic.is.tohoku.ac.jp/ja/swk/
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 09 130219 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Feature Descriptors Feature Matching Feature
More informationLecture 16: Computer Vision
CS442/542b: Artificial ntelligence Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field Methods
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationLocal Features: Detection, Description & Matching
Local Features: Detection, Description & Matching Lecture 08 Computer Vision Material Citations Dr George Stockman Professor Emeritus, Michigan State University Dr David Lowe Professor, University of British
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical
More informationCS 4495 Computer Vision A. Bobick. CS 4495 Computer Vision. Features 2 SIFT descriptor. Aaron Bobick School of Interactive Computing
CS 4495 Computer Vision Features 2 SIFT descriptor Aaron Bobick School of Interactive Computing Administrivia PS 3: Out due Oct 6 th. Features recap: Goal is to find corresponding locations in two images.
More informationMotion illusion, rotating snakes
Motion illusion, rotating snakes Local features: main components 1) Detection: Find a set of distinctive key points. 2) Description: Extract feature descriptor around each interest point as vector. x 1
More informationBSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy
BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving
More informationScale Invariant Feature Transform
Why do we care about matching features? Scale Invariant Feature Transform Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Automatic
More informationScale Invariant Feature Transform
Scale Invariant Feature Transform Why do we care about matching features? Camera calibration Stereo Tracking/SFM Image moiaicing Object/activity Recognition Objection representation and recognition Image
More informationBuilding a Panorama. Matching features. Matching with Features. How do we build a panorama? Computational Photography, 6.882
Matching features Building a Panorama Computational Photography, 6.88 Prof. Bill Freeman April 11, 006 Image and shape descriptors: Harris corner detectors and SIFT features. Suggested readings: Mikolajczyk
More informationFeature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1
Feature Detection Raul Queiroz Feitosa 3/30/2017 Feature Detection 1 Objetive This chapter discusses the correspondence problem and presents approaches to solve it. 3/30/2017 Feature Detection 2 Outline
More informationSIFT - scale-invariant feature transform Konrad Schindler
SIFT - scale-invariant feature transform Konrad Schindler Institute of Geodesy and Photogrammetry Invariant interest points Goal match points between images with very different scale, orientation, projective
More informationFundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F
Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix F Fundamental
More informationMatching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.
Matching Compare region of image to region of image. We talked about this for stereo. Important for motion. Epipolar constraint unknown. But motion small. Recognition Find object in image. Recognize object.
More informationCS231A Section 6: Problem Set 3
CS231A Section 6: Problem Set 3 Kevin Wong Review 6 -! 1 11/09/2012 Announcements PS3 Due 2:15pm Tuesday, Nov 13 Extra Office Hours: Friday 6 8pm Huang Common Area, Basement Level. Review 6 -! 2 Topics
More informationLocal Image Features
Local Image Features Computer Vision CS 143, Brown Read Szeliski 4.1 James Hays Acknowledgment: Many slides from Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial This section: correspondence and alignment
More informationFeatures Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)
Features Points Andrea Torsello DAIS Università Ca Foscari via Torino 155, 30172 Mestre (VE) Finding Corners Edge detectors perform poorly at corners. Corners provide repeatable points for matching, so
More informationSUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS
SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS Cognitive Robotics Original: David G. Lowe, 004 Summary: Coen van Leeuwen, s1460919 Abstract: This article presents a method to extract
More informationMotion Estimation and Optical Flow Tracking
Image Matching Image Retrieval Object Recognition Motion Estimation and Optical Flow Tracking Example: Mosiacing (Panorama) M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 Example 3D Reconstruction
More informationCorner Detection. GV12/3072 Image Processing.
Corner Detection 1 Last Week 2 Outline Corners and point features Moravec operator Image structure tensor Harris corner detector Sub-pixel accuracy SUSAN FAST Example descriptor: SIFT 3 Point Features
More informationImage Features: Local Descriptors. Sanja Fidler CSC420: Intro to Image Understanding 1/ 58
Image Features: Local Descriptors Sanja Fidler CSC420: Intro to Image Understanding 1/ 58 [Source: K. Grauman] Sanja Fidler CSC420: Intro to Image Understanding 2/ 58 Local Features Detection: Identify
More informationAK Computer Vision Feature Point Detectors and Descriptors
AK Computer Vision Feature Point Detectors and Descriptors 1 Feature Point Detectors and Descriptors: Motivation 2 Step 1: Detect local features should be invariant to scale and rotation, or perspective
More informationCAP 5415 Computer Vision Fall 2012
CAP 5415 Computer Vision Fall 01 Dr. Mubarak Shah Univ. of Central Florida Office 47-F HEC Lecture-5 SIFT: David Lowe, UBC SIFT - Key Point Extraction Stands for scale invariant feature transform Patented
More informationApplication questions. Theoretical questions
The oral exam will last 30 minutes and will consist of one application question followed by two theoretical questions. Please find below a non exhaustive list of possible application questions. The list
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationMotion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi
Motion and Optical Flow Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi We live in a moving world Perceiving, understanding and predicting motion is an important part of our daily lives Motion
More informationLocal Feature Detectors
Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,
More informationFeature Detection and Matching
and Matching CS4243 Computer Vision and Pattern Recognition Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (CS4243) Camera Models 1 /
More informationIntroduction. Introduction. Related Research. SIFT method. SIFT method. Distinctive Image Features from Scale-Invariant. Scale.
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe presented by, Sudheendra Invariance Intensity Scale Rotation Affine View point Introduction Introduction SIFT (Scale Invariant Feature
More informationFeature descriptors. Alain Pagani Prof. Didier Stricker. Computer Vision: Object and People Tracking
Feature descriptors Alain Pagani Prof. Didier Stricker Computer Vision: Object and People Tracking 1 Overview Previous lectures: Feature extraction Today: Gradiant/edge Points (Kanade-Tomasi + Harris)
More informationThe SIFT (Scale Invariant Feature
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical
More informationLocal features and image matching. Prof. Xin Yang HUST
Local features and image matching Prof. Xin Yang HUST Last time RANSAC for robust geometric transformation estimation Translation, Affine, Homography Image warping Given a 2D transformation T and a source
More informationFeature Descriptors. CS 510 Lecture #21 April 29 th, 2013
Feature Descriptors CS 510 Lecture #21 April 29 th, 2013 Programming Assignment #4 Due two weeks from today Any questions? How is it going? Where are we? We have two umbrella schemes for object recognition
More informationLocal Features Tutorial: Nov. 8, 04
Local Features Tutorial: Nov. 8, 04 Local Features Tutorial References: Matlab SIFT tutorial (from course webpage) Lowe, David G. Distinctive Image Features from Scale Invariant Features, International
More informationAutomatic Image Alignment (feature-based)
Automatic Image Alignment (feature-based) Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 Today s lecture Feature
More informationComputer Vision Lecture 20
Computer Perceptual Vision and Sensory WS 16/17 Augmented Computing Computer Perceptual Vision and Sensory WS 16/17 Augmented Computing Computer Perceptual Vision and Sensory WS 16/17 Augmented Computing
More informationLocal features: detection and description. Local invariant features
Local features: detection and description Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms
More informationComputer Vision Lecture 20
Computer Perceptual Vision and Sensory WS 16/76 Augmented Computing Many slides adapted from K. Grauman, S. Seitz, R. Szeliski, M. Pollefeys, S. Lazebnik Computer Vision Lecture 20 Motion and Optical Flow
More information3D Photography: Stereo
3D Photography: Stereo Marc Pollefeys, Torsten Sattler Spring 2016 http://www.cvg.ethz.ch/teaching/3dvision/ 3D Modeling with Depth Sensors Today s class Obtaining depth maps / range images unstructured
More informationLocal Image Features
Local Image Features Computer Vision Read Szeliski 4.1 James Hays Acknowledgment: Many slides from Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial Flashed Face Distortion 2nd Place in the 8th Annual Best
More informationComputational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --
Computational Optical Imaging - Optique Numerique -- Single and Multiple View Geometry, Stereo matching -- Autumn 2015 Ivo Ihrke with slides by Thorsten Thormaehlen Reminder: Feature Detection and Matching
More informationSchool of Computing University of Utah
School of Computing University of Utah Presentation Outline 1 2 3 4 Main paper to be discussed David G. Lowe, Distinctive Image Features from Scale-Invariant Keypoints, IJCV, 2004. How to find useful keypoints?
More informationHarder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford
Image matching Harder case by Diva Sian by Diva Sian by scgbt by swashford Even harder case Harder still? How the Afghan Girl was Identified by Her Iris Patterns Read the story NASA Mars Rover images Answer
More informationCEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.
CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt. Section 10 - Detectors part II Descriptors Mani Golparvar-Fard Department of Civil and Environmental Engineering 3129D, Newmark Civil Engineering
More informationAutomatic Image Alignment (direct) with a lot of slides stolen from Steve Seitz and Rick Szeliski
Automatic Image Alignment (direct) with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2005 Today Go over Midterm Go over Project #3
More informationImage Features: Detection, Description, and Matching and their Applications
Image Features: Detection, Description, and Matching and their Applications Image Representation: Global Versus Local Features Features/ keypoints/ interset points are interesting locations in the image.
More informationObject Recognition with Invariant Features
Object Recognition with Invariant Features Definition: Identify objects or scenes and determine their pose and model parameters Applications Industrial automation and inspection Mobile robots, toys, user
More informationLocal features: detection and description May 12 th, 2015
Local features: detection and description May 12 th, 2015 Yong Jae Lee UC Davis Announcements PS1 grades up on SmartSite PS1 stats: Mean: 83.26 Standard Dev: 28.51 PS2 deadline extended to Saturday, 11:59
More informationAutonomous Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 7.1: 2D Motion Estimation in Images Jürgen Sturm Technische Universität München 3D to 2D Perspective Projections
More informationAutomatic Image Alignment
Automatic Image Alignment Mike Nese with a lot of slides stolen from Steve Seitz and Rick Szeliski 15-463: Computational Photography Alexei Efros, CMU, Fall 2010 Live Homography DEMO Check out panoramio.com
More informationDiscovering Visual Hierarchy through Unsupervised Learning Haider Razvi
Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi hrazvi@stanford.edu 1 Introduction: We present a method for discovering visual hierarchy in a set of images. Automatically grouping
More informationAutomatic Image Alignment
Automatic Image Alignment with a lot of slides stolen from Steve Seitz and Rick Szeliski Mike Nese CS194: Image Manipulation & Computational Photography Alexei Efros, UC Berkeley, Fall 2018 Live Homography
More informationDense Image-based Motion Estimation Algorithms & Optical Flow
Dense mage-based Motion Estimation Algorithms & Optical Flow Video A video is a sequence of frames captured at different times The video data is a function of v time (t) v space (x,y) ntroduction to motion
More informationImage matching. Announcements. Harder case. Even harder case. Project 1 Out today Help session at the end of class. by Diva Sian.
Announcements Project 1 Out today Help session at the end of class Image matching by Diva Sian by swashford Harder case Even harder case How the Afghan Girl was Identified by Her Iris Patterns Read the
More informationComparison between Motion Analysis and Stereo
MOTION ESTIMATION The slides are from several sources through James Hays (Brown); Silvio Savarese (U. of Michigan); Octavia Camps (Northeastern); including their own slides. Comparison between Motion Analysis
More information3D from Photographs: Automatic Matching of Images. Dr Francesco Banterle
3D from Photographs: Automatic Matching of Images Dr Francesco Banterle francesco.banterle@isti.cnr.it 3D from Photographs Automatic Matching of Images Camera Calibration Photographs Surface Reconstruction
More informationCOMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE
COMPUTER VISION 2017-2018 > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE OUTLINE Optical flow Lucas-Kanade Horn-Schunck Applications of optical flow Optical flow tracking Histograms of oriented flow Assignment
More informationFeature descriptors and matching
Feature descriptors and matching Detections at multiple scales Invariance of MOPS Intensity Scale Rotation Color and Lighting Out-of-plane rotation Out-of-plane rotation Better representation than color:
More information2D Image Processing Feature Descriptors
2D Image Processing Feature Descriptors Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Overview
More informationMidterm Wed. Local features: detection and description. Today. Last time. Local features: main components. Goal: interest operator repeatability
Midterm Wed. Local features: detection and description Monday March 7 Prof. UT Austin Covers material up until 3/1 Solutions to practice eam handed out today Bring a 8.5 11 sheet of notes if you want Review
More informationStructure from Motion. Introduction to Computer Vision CSE 152 Lecture 10
Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley
More informationLocal invariant features
Local invariant features Tuesday, Oct 28 Kristen Grauman UT-Austin Today Some more Pset 2 results Pset 2 returned, pick up solutions Pset 3 is posted, due 11/11 Local invariant features Detection of interest
More informationLecture 10 Detectors and descriptors
Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of detectors SIFT Shape context Silvio Savarese Lecture 10-26-Feb-14 From the 3D to 2D & vice versa P =
More informationVC 11/12 T11 Optical Flow
VC 11/12 T11 Optical Flow Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos Miguel Tavares Coimbra Outline Optical Flow Constraint Equation Aperture
More informationHarder case. Image matching. Even harder case. Harder still? by Diva Sian. by swashford
Image matching Harder case by Diva Sian by Diva Sian by scgbt by swashford Even harder case Harder still? How the Afghan Girl was Identified by Her Iris Patterns Read the story NASA Mars Rover images Answer
More informationVisual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.
Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania 1 What is visual tracking? estimation of the target location over time 2 applications Six main areas:
More informationStep-by-Step Model Buidling
Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure
More informationLeow Wee Kheng CS4243 Computer Vision and Pattern Recognition. Motion Tracking. CS4243 Motion Tracking 1
Leow Wee Kheng CS4243 Computer Vision and Pattern Recognition Motion Tracking CS4243 Motion Tracking 1 Changes are everywhere! CS4243 Motion Tracking 2 Illumination change CS4243 Motion Tracking 3 Shape
More informationComputational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --
Computational Optical Imaging - Optique Numerique -- Multiple View Geometry and Stereo -- Winter 2013 Ivo Ihrke with slides by Thorsten Thormaehlen Feature Detection and Matching Wide-Baseline-Matching
More informationCS 558: Computer Vision 4 th Set of Notes
1 CS 558: Computer Vision 4 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview Keypoint matching Hessian
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 19: Optical flow http://en.wikipedia.org/wiki/barberpole_illusion Readings Szeliski, Chapter 8.4-8.5 Announcements Project 2b due Tuesday, Nov 2 Please sign
More informationVisual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The
More informationProf. Feng Liu. Spring /26/2017
Prof. Feng Liu Spring 2017 http://www.cs.pdx.edu/~fliu/courses/cs510/ 04/26/2017 Last Time Re-lighting HDR 2 Today Panorama Overview Feature detection Mid-term project presentation Not real mid-term 6
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationCS201: Computer Vision Introduction to Tracking
CS201: Computer Vision Introduction to Tracking John Magee 18 November 2014 Slides courtesy of: Diane H. Theriault Question of the Day How can we represent and use motion in images? 1 What is Motion? Change
More informationComputer Vision Lecture 20
Computer Vision Lecture 2 Motion and Optical Flow Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de 28.1.216 Man slides adapted from K. Grauman, S. Seitz, R. Szeliski,
More informationEppur si muove ( And yet it moves )
Eppur si muove ( And yet it moves ) - Galileo Galilei University of Texas at Arlington Tracking of Image Features CSE 4392-5369 Vision-based Robot Sensing, Localization and Control Dr. Gian Luca Mariottini,
More informationFinal Exam Study Guide
Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.
More informationCS5670: Computer Vision
CS5670: Computer Vision Noah Snavely Lecture 4: Harris corner detection Szeliski: 4.1 Reading Announcements Project 1 (Hybrid Images) code due next Wednesday, Feb 14, by 11:59pm Artifacts due Friday, Feb
More informationCapturing, Modeling, Rendering 3D Structures
Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights
More information