Feature Tracking and Optical Flow

Similar documents
Feature Tracking and Optical Flow

Peripheral drift illusion

ECE Digital Image Processing and Introduction to Computer Vision

Motion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi

Optical flow and tracking

CS 4495 Computer Vision Motion and Optic Flow

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

Multi-stable Perception. Necker Cube

Computer Vision Lecture 20

Computer Vision Lecture 20

Visual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys

Lecture 16: Computer Vision

Lecture 16: Computer Vision

EECS 556 Image Processing W 09

Comparison between Motion Analysis and Stereo

Lucas-Kanade Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

EE795: Computer Vision and Intelligent Systems

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Lecture 19: Motion. Effect of window size 11/20/2007. Sources of error in correspondences. Review Problem set 3. Tuesday, Nov 20

Computer Vision Lecture 20

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

CS6670: Computer Vision

VC 11/12 T11 Optical Flow

EE795: Computer Vision and Intelligent Systems

Overview. Video. Overview 4/7/2008. Optical flow. Why estimate motion? Motion estimation: Optical flow. Motion Magnification Colorization.

Automatic Image Alignment (direct) with a lot of slides stolen from Steve Seitz and Rick Szeliski

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

Lecture 14: Tracking mo3on features op3cal flow

Dense Image-based Motion Estimation Algorithms & Optical Flow

Motion Estimation. There are three main types (or applications) of motion estimation:

Optical flow. Cordelia Schmid

Optical flow. Cordelia Schmid

Lecture 13: Tracking mo3on features op3cal flow

Structure from Motion

Capturing, Modeling, Rendering 3D Structures

Motion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Autonomous Navigation for Flying Robots

Lecture 20: Tracking. Tuesday, Nov 27

CS201: Computer Vision Introduction to Tracking

Leow Wee Kheng CS4243 Computer Vision and Pattern Recognition. Motion Tracking. CS4243 Motion Tracking 1

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Wikipedia - Mysid

Computer Vision II Lecture 4

3D Photography. Marc Pollefeys, Torsten Sattler. Spring 2015

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.

CS664 Lecture #18: Motion

Visual Tracking (1) Feature Point Tracking and Block Matching

Particle Tracking. For Bulk Material Handling Systems Using DEM Models. By: Jordan Pease

Edge and corner detection

CSCI 1290: Comp Photo

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

(More) Algorithms for Cameras: Edge Detec8on Modeling Cameras/Objects. Connelly Barnes

Image processing and features

Visual Tracking. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

CS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow

3D Vision. Viktor Larsson. Spring 2019

Displacement estimation

Computer Vision I. Announcements. Fourier Tansform. Efficient Implementation. Edge and Corner Detection. CSE252A Lecture 13.

Straight Lines and Hough

Computer Vision Lecture 18

CS231A Section 6: Problem Set 3

Local Feature Detectors

Computer Vision II Lecture 4

Fi#ng & Matching Region Representa3on Image Alignment, Op3cal Flow

Visual Tracking (1) Pixel-intensity-based methods

Optic Flow and Motion Detection

Spatial track: motion modeling

Spatial track: motion modeling

Using temporal seeding to constrain the disparity search range in stereo matching

Kanade Lucas Tomasi Tracking (KLT tracker)

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

CPSC 425: Computer Vision

Face Tracking : An implementation of the Kanade-Lucas-Tomasi Tracking algorithm

Computer Vision I - Filtering and Feature detection

Towards the completion of assignment 1

Motion. CS 554 Computer Vision Pinar Duygulu Bilkent University

Feature Based Registration - Image Alignment

Final Exam Study Guide

Computer Vision I - Basics of Image Processing Part 2

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

16720 Computer Vision: Homework 3 Template Tracking and Layered Motion.

Computer Vision I. Announcement. Corners. Edges. Numerical Derivatives f(x) Edge and Corner Detection. CSE252A Lecture 11

SE 263 R. Venkatesh Babu. Object Tracking. R. Venkatesh Babu

Automatic Image Alignment (feature-based)

Mariya Zhariy. Uttendorf Introduction to Optical Flow. Mariya Zhariy. Introduction. Determining. Optical Flow. Results. Motivation Definition

Marcel Worring Intelligent Sensory Information Systems

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Tracking Computer Vision Spring 2018, Lecture 24

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Optical Flow CS 637. Fuxin Li. With materials from Kristen Grauman, Richard Szeliski, S. Narasimhan, Deqing Sun

Targil 12 : Image Segmentation. Image segmentation. Why do we need it? Image segmentation

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Computer Vision Lecture 17

Computer Vision Lecture 17

Single-view 3D Reconstruction

SUMMARY: DISTINCTIVE IMAGE FEATURES FROM SCALE- INVARIANT KEYPOINTS

Transcription:

Feature Tracking and Optical Flow Prof. D. Stricker Doz. G. Bleser Many slides adapted from James Hays, Derek Hoeim, Lana Lazebnik, Silvio Saverse, who 1 in turn adapted slides from Steve Seitz, Rick Szeliski, Martial Hebert, Mark Pollefeys, and others

Outline of the lecture Feature-tracking Extract visual features (corners, textured areas) and track them over multiple frames Optical flow Recover image motion at each pixel from spatio-temporal image brightness variations (optical flow) Two problems, one registration method B. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the International Joint Conference on Artificial Intelligence, pp. 674 679, 1981. 2

Some examples Video examples (external viewer) 3

Feature tracking Many problems, such as object tracking or structure from motion require tracking points. If motion is small, tracking is an easy way to get them. 4

Feature tracking Challenges Figure out which features can be tracked Efficiently track across frames Some points may change appearance over time (e.g., due to rotation, moving into shadows, etc.) Drift: small errors can accumulate as appearance model is updated Points may appear or disappear: need to be able to add/delete tracked points 5

Three assumptions Brightness consistency Spatial coherence Temporal persistence 6

Brightness consistency Image measurement (e.g. brightness) in a small region remain the same although their location may change. 7

Spatial coherence Neighboring points in the scene typically belong to the same surface and hence typically have similar motions. Since they also project to nearby pixels in the image, we expect spatial coherence in image flow. 8

Temporal persistence The image motion of a surface patch changes gradually over time. 9

Feature tracking I(x,y,t) I(x,y,t+1) Given two subsequent frames, estimate the point translation Key assumptions of Kanade-Lucas Tracker (KLT) Brightness constancy: projection of the same point looks the same in every frame Temporal persistence ( small motions ): points do not move very far Spatial coherence: points move like their neighbors 10

t y x I v I u I t y x I t v y u x I + + + + + + ),, ( 1),, ( The brightness constancy constraint Brightness Constancy Equation: ), ( ),, ( 1, + + + = t v y u x I t y x I Take Taylor expansion of I(x+u, y+v, t+1) at (x,y,t) to linearize the right side: I(x,y,t) I(x,y,t+1) Image derivative along x t y x I v I u I t y x I t v y u x I + + = + + + + ),, ( 1),, ( Difference over frames 11 0 + + t y x I v I u I Hence, [ ] 0 I v u I t T = +

The brightness constancy constraint Can we use this equation to recover image motion (u,v) at each pixel? I [ ] T u v + I = 0 How many equations and unknowns per pixel? One equation (this is a scalar equation!), two unknowns (u,v) The component of the motion perpendicular to the gradient (i.e., parallel to the edge) cannot be measured t If (u, v) satisfies the equation, so does (u+u, v+v ) if I [ u' v' ] T = 0 gradient (u,v) (u+u,v+v ) (u,v ) edge 12

The aperture problem Actual motion 13

The aperture problem aperture problem: In the absence of additional information the visual system prefers the slowest possible motion: i.e., motion orthogonal to the moving line http://en.wikipedia.org/wiki/barberpole_illusion Perceived motion 14

Solving the ambiguity B. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the International Joint Conference on Artificial Intelligence, pp. 674 679, 1981. How to get more equations for a pixel? Spatial coherence constraint Assume the pixel s neighbors have the same (u,v) If we use a 5x5 window, that gives us 25 equations per pixel 15

RGB version How to get more equations for a pixel? Basic idea: impose additional constraints If we use a 5x5 window, that gives us 25*3 equations per pixel! 16 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Lucas-Kanade flow Prob: we have more equations than unknowns Solution: solve least squares problem Recall: Least squares solution: Set derivative to zero: Also called pseudo-inverse Matlab operator: x=a\b 17

Lukas-Kanade flow Prob: we have more equations than unknowns Solution: solve least squares problem minimum least squares solution given by solution (in d) of: The summations are over all pixels in the K x K window This technique was first proposed by Lukas & Kanade (1981) 18 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Conditions for solvability Optimal (u, v) satisfies Lucas-Kanade equation When is this solvable? I.e., what are good points to track? A T A should be invertible A T A should not be too small due to noise eigenvalues λ 1 and λ 2 of A T A should not be too small A T A should be well-conditioned λ 1 / λ 2 should not be too large (λ 1 = larger eigenvalue) Does this remind you of anything? 19 Criteria for Corner Detector (See also: Harris)

Eigenvectors of A T A Suppose (x,y) is on an edge. What is A T A? gradients along edge all point the same direction gradients away from edge have small magnitude is an eigenvector with eigenvalue What s the other eigenvector of A T A? let N be perpendicular to N is the second eigenvector with eigenvalue 0 The eigenvectors of A T A relate to edge direction and magnitude 20 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

M = A T A is the second moment matrix! (Harris corner detector ) Eigenvectors and eigenvalues of A T A relate to edge direction and magnitude The eigenvector associated with the larger eigenvalue points in the direction of fastest intensity change The other eigenvector is orthogonal to it 21

Edge large gradients, all the same large λ 1, small λ 2 22 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Low texture region gradients have small magnitude small λ 1, small λ 2 23 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

High textured region gradients are different, large magnitudes large λ 1, large λ 2 24 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Observation This is a two image problem BUT Can measure sensitivity by just looking at one of the images! This tells us which pixels are easy to track, which are hard very useful later on when we do feature tracking... 25 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

The aperture problem resolved Actual motion 26

The aperture problem resolved Perceived motion 27

Errors in Lucas-Kanade What are the potential causes of errors in this procedure? Suppose A T A is easily invertible Suppose there is not much noise in the image When our assumptions are violated Brightness constancy is not satisfied The motion is not small A point does not move like its neighbors window size is too large what is the ideal window size? 28 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Revisiting the small motion assumption Is this motion small enough? Probably not it s much larger than one pixel (2 nd order terms dominate) How might we solve this problem? 29 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Reduce the resolution! 30 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Coarse-to-fine optical flow estimation u=1.25 pixels u=2.5 pixels u=5 pixels image I t-1 u=10 pixels image I Gaussian pyramid of image I t-1 31 Gaussian pyramid of image I

Dealing with larger movements: coarse-to-fine registration run iterative L-K upsample run iterative L-K... image J 1 image I 2 Gaussian pyramid of image 1 (t) Gaussian pyramid of image 2 (t+1) 32

Dealing with larger movements: Iterative refinement 1. Initialize (x,y ) = (x,y) 2. Compute (u,v) by Original (x,y) position I t = I(x, y, t+1) - I(x, y, t) 2 nd moment matrix for feature patch in first image displacement 3. Shift window by (u, v): x =x +u; y =y +v; 4. Recalculate I t 5. Repeat steps 2-4 until small change Use interpolation for subpixel values 33

Shi-Tomasi feature tracker Find good features using eigenvalues of second-moment matrix (e.g., Harris detector or threshold on the smallest eigenvalue) Key idea: good features to track are the ones whose motion can be estimated reliably Track from frame to frame with Lucas-Kanade This amounts to assuming a translation model for frame-to-frame feature movement Check consistency of tracks by affine registration to the first observed instance of the feature Affine model is more accurate for larger displacements Comparing to the first frame helps to minimize drift J. Shi and C. Tomasi. Good Features to Track. CVPR 1994. 34

Tracking example J. Shi and C. Tomasi. Good Features to Track. CVPR 1994. 35

Summary of KLT tracking Find a good point to track (harris corner) Use intensity second moment matrix and difference across frames to find displacement Iterate and use coarse-to-fine search to deal with larger movements When creating long tracks, check appearance of registered patch against appearance of initial patch to find points that have drifted 36

Implementation issues Window size Small window more sensitive to noise and may miss larger motions (without pyramid) Large window more likely to cross an occlusion boundary (and it s slower) 15x15 to 31x31 seems typical Weighting the window Common to apply weights so that center matters more (e.g., with Gaussian) 37

Tracking over many frames Select features in first frame For each frame: Update positions of tracked features Discrete search (e.g. cross-correlation) or Lucas-Kanade (or a combination of the two) Terminate inconsistent tracks Compute similarity with corresponding feature in the previous frame or in the first frame where it s visible Find more features to track 38

Optical flow Vector field function of the spatio-temporal image brightness variations Picture courtesy of Selim Temizer - Learning and Intelligent Systems (LIS) Group, MIT 39

Uses of motion Estimating 3D structure Segmenting objects based on motion cues Learning and tracking dynamical models Recognizing events and activities Improving video quality (motion stabilization) 40

Motion field The motion field is the projection of the 3D scene motion into the image 41

Optical flow Definition: optical flow is the apparent motion of brightness patterns in the image Ideally, optical flow would be the same as the motion field Have to be careful: apparent motion can be caused by lighting changes without any actual motion Think of a uniform rotating sphere under fixed lighting vs. a stationary sphere under moving illumination 42

Lucas-Kanade Optical Flow Same as Lucas-Kanade feature tracking, but for each pixel As we saw, works better for textured pixels Operations can be done one frame at a time, rather than pixel by pixel Efficient 43

Iterative Refinement Iterative Lukas-Kanade Algorithm 1. Estimate displacement at each pixel by solving Lucas-Kanade equations 2. Warp I(t) towards I(t+1) using the estimated flow field - Basically, just interpolation 3. Repeat until convergence 44 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Coarse-to-fine optical flow estimation run iterative L-K run iterative L-K warp & upsample... image J 1 image I 2 Gaussian pyramid of image 1 (t) Gaussian pyramid of image 2 (t+1) 45

Example 46 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Multi-resolution registration 47 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Optical Flow Results 48 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Optical Flow Results 49 * From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

Summary Major contributions from Kanade, Lucas, Tomasi, We speak about the KLT-Tracker Tracking feature points Optical flow Stereo Structure from motion Key ideas By assuming brightness constancy, truncated Taylor expansion leads to simple and fast patch matching across frames Coarse-to-fine registration 50

Dense Non-Rigid Structure from Motion Obtained 3D shape (first results) Future work: Fusion with non-rigid 3D reconstruction from Kinect data 02.06.14 Loop closure to obtain a closed Copyright 3D shape 2013 Augmented Vision - DFKI

Thank you! 52