Visual motion. Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys

Similar documents
Computer Vision Lecture 20

EECS 556 Image Processing W 09

Computer Vision Lecture 20

Computer Vision Lecture 20

Peripheral drift illusion

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow

Computer Vision Lecture 18

Optical flow and tracking

CS 4495 Computer Vision Motion and Optic Flow

Motion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi

Optical flow. Cordelia Schmid

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

Comparison between Motion Analysis and Stereo

Lecture 19: Motion. Effect of window size 11/20/2007. Sources of error in correspondences. Review Problem set 3. Tuesday, Nov 20

Multi-stable Perception. Necker Cube

ECE Digital Image Processing and Introduction to Computer Vision

Optical flow. Cordelia Schmid

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Lecture 16: Computer Vision

Lecture 16: Computer Vision

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Motion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

EE 264: Image Processing and Reconstruction. Image Motion Estimation I. EE 264: Image Processing and Reconstruction. Outline

EE795: Computer Vision and Intelligent Systems

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Lucas-Kanade Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

Visual Tracking. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

CS6670: Computer Vision

Automatic Image Alignment (direct) with a lot of slides stolen from Steve Seitz and Rick Szeliski

VC 11/12 T11 Optical Flow

Visual Tracking (1) Feature Point Tracking and Block Matching

Particle Tracking. For Bulk Material Handling Systems Using DEM Models. By: Jordan Pease

Capturing, Modeling, Rendering 3D Structures

Spatial track: motion modeling

CS201: Computer Vision Introduction to Tracking

Lecture 14: Tracking mo3on features op3cal flow

Overview. Video. Overview 4/7/2008. Optical flow. Why estimate motion? Motion estimation: Optical flow. Motion Magnification Colorization.

Optic Flow and Basics Towards Horn-Schunck 1

Spatial track: motion modeling

Optical Flow CS 637. Fuxin Li. With materials from Kristen Grauman, Richard Szeliski, S. Narasimhan, Deqing Sun

Visual Tracking (1) Tracking of Feature Points and Planar Rigid Objects

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Dense Image-based Motion Estimation Algorithms & Optical Flow

CS231A Section 6: Problem Set 3

EE795: Computer Vision and Intelligent Systems

Computer Vision II Lecture 4

Autonomous Navigation for Flying Robots

CS664 Lecture #18: Motion

CSCI 1290: Comp Photo

Lecture 13: Tracking mo3on features op3cal flow

Leow Wee Kheng CS4243 Computer Vision and Pattern Recognition. Motion Tracking. CS4243 Motion Tracking 1

Marcel Worring Intelligent Sensory Information Systems

Displacement estimation

Mariya Zhariy. Uttendorf Introduction to Optical Flow. Mariya Zhariy. Introduction. Determining. Optical Flow. Results. Motivation Definition

CS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow

Optical Flow-Based Motion Estimation. Thanks to Steve Seitz, Simon Baker, Takeo Kanade, and anyone else who helped develop these slides.

Wikipedia - Mysid

Lecture 20: Tracking. Tuesday, Nov 27

Visual Tracking (1) Pixel-intensity-based methods

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Motion Estimation. There are three main types (or applications) of motion estimation:

Computer Vision II Lecture 4

Local Feature Detectors

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Image Features. Work on project 1. All is Vanity, by C. Allan Gilbert,

Image processing and features

Feature Detectors and Descriptors: Corners, Lines, etc.

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Feature Based Registration - Image Alignment

Motion estimation. Lihi Zelnik-Manor

Computer Vision I. Announcements. Fourier Tansform. Efficient Implementation. Edge and Corner Detection. CSE252A Lecture 13.

CAP5415-Computer Vision Lecture 8-Mo8on Models, Feature Tracking, and Alignment. Ulas Bagci

CPSC 425: Computer Vision

Computer Vision I - Filtering and Feature detection

Feature Detection. Raul Queiroz Feitosa. 3/30/2017 Feature Detection 1

Edge and corner detection

Comparison Between The Optical Flow Computational Techniques

Announcements. Motion. Structure-from-Motion (SFM) Motion. Discrete Motion: Some Counting

Horn-Schunck and Lucas Kanade 1

Kanade Lucas Tomasi Tracking (KLT tracker)

Automatic Image Alignment (feature-based)

Final Exam Study Guide

SURVEY OF LOCAL AND GLOBAL OPTICAL FLOW WITH COARSE TO FINE METHOD

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Corner Detection. Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology

Computer Vision for HCI. Topics of This Lecture

Robert Collins CSE598G. Intro to Template Matching and the Lucas-Kanade Method

Tracking Computer Vision Spring 2018, Lecture 24

Tracking in image sequences

Towards the completion of assignment 1

Computer Vision. Computer Science Tripos Part II. Dr Christopher Town. Texture. Includes: more regular patterns. Includes: more random patterns

3D Photography. Marc Pollefeys, Torsten Sattler. Spring 2015

Features Points. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Ruch (Motion) Rozpoznawanie Obrazów Krzysztof Krawiec Instytut Informatyki, Politechnika Poznańska. Krzysztof Krawiec IDSS

Segmentation and Tracking of Partial Planar Templates

Introduction to Computer Vision

Using temporal seeding to constrain the disparity search range in stereo matching

Transcription:

Visual motion Man slides adapted from S. Seitz, R. Szeliski, M. Pollefes

Motion and perceptual organization Sometimes, motion is the onl cue

Motion and perceptual organization Sometimes, motion is the onl cue

Motion and perceptual organization Even impoverished motion data can evoke a strong percept G. Johansson, Visual Perception of Biological Motion and a Model For Its Analsis", Perception and Pschophsics 14, 201-211, 1973.

Motion and perceptual organization Even impoverished motion data can evoke a strong percept G. Johansson, Visual Perception of Biological Motion and a Model For Its Analsis", Perception and Pschophsics 14, 201-211, 1973.

Motion and perceptual organization Even impoverished motion data can evoke a strong percept G. Johansson, Visual Perception of Biological Motion and a Model For Its Analsis", Perception and Pschophsics 14, 201-211, 1973.

Uses of motion Estimating 3D structure Segmenting objects based on motion cues Learning and tracking dnamical models Recognizing events and activities

Motion field The motion field is the projection of the 3D scene motion into the image

Motion field and parallax X(t) is a moving 3D point Velocit of scene point: V =dx/dt x(t) = (x(t),(t)) is the projection of X in the image Apparent velocit v in the image: given b components v x = dx/dt and v =d/dt These components are known as the motion field of the image X(t) V x(t) v X(t+dt) x(t+dt)

Motion field and parallax To find image velocit v, differentiate x=(x,) ( with respect to t (using quotient rule): X(t) V X(t+dt) x = f X Z v x ZVx V = f 2 Z f V x V z x = Z z X v x(t+dt) = f Y Z v = f V V Z z x(t) Image motion is a function of both the 3D motion (V) and the depth of the 3D point (Z)

Motion field and parallax Pure translation: V is constant everwhere v x = v = f V f V V Z V x Z z z x v 1 = ( v 0 Vzx), v ( ) 0 = f Vx, f V Z

Motion field and parallax Pure translation: V is constant everwhere 1 v = ( v 0 V x), ) v = ( ) z 0 f Vx, f V Z The length of the motion vectors is inversel proportional to the depth Z V z is nonzero: Ever motion vector points toward (or awa from) the vanishing point of the translation direction

Motion field and parallax Pure translation: V is constant everwhere 1 v = ( v 0 V x), ) v = ( ) z 0 f Vx, f V Z The length of the motion vectors is inversel proportional to the depth Z V z is nonzero: Ever motion vector points toward (or awa from) the vanishing point of the translation direction V z is zero: Motion is parallel to the image plane, all the motion vectors are parallel

Optical flow Definition: optical flow is the apparent motion of brightness patterns in the image Ideall, optical flow would be the same as the motion field Have to be careful: apparent motion can be caused b lighting changes without an actual motion Think of a uniform rotating sphere under fixed lighting vs. a stationar sphere under moving illumination

Estimating optical flow I(x,,t 1) I(x,,t) Given two subsequent frames, estimate the apparent motion field u(x,) and v(x,) between them Ke assumptions Brightness constanc: projection of the same point looks the same in ever frame Small motion: points do not move ver far Spatial coherence: points move like their neighbors

The brightness constanc constraint I(x,,t 1) I(x,,t) Brightness Constanc Equation: ), ( 1),, ( ),, ( ), ( t x x v u x I t x I + + = ), ( ),, ( ),, ( ), ( Linearizing the right side using Talor expansion: ), ( ), ( ),, ( 1),, ( x v I x u I t x I t x I x + + 0 + + t x I v I u I Hence,

The brightness constanc constraint I x u + I v + = 0 How man equations and unknowns per pixel? One equation, two unknowns Intuitivel, what does this constraint mean? I The component of the flow perpendicular p to the gradient (i.e., parallel to the edge) is unknown I ( u, v) + I = t t 0

The brightness constanc constraint I x u + I v + = 0 How man equations and unknowns per pixel? One equation, two unknowns Intuitivel, what does this constraint mean? I The component of the flow perpendicular p to the gradient (i.e., parallel to the edge) is unknown I ( u, v) + I = t t 0 If (u, v) satisfies the equation, so does (u+u, v+v ) if I ( ( u', v') = 0 gradient (u,v) (u,v ) (u+u,v+v ) edge

The aperture problem Perceived motion

The aperture problem Actual motion

The barber pole illusion http://en.wikipedia.org/wiki/barberpole_illusion

The barber pole illusion http://en.wikipedia.org/wiki/barberpole_illusion

The barber pole illusion http://en.wikipedia.org/wiki/barberpole_illusion

Solving the aperture problem How to get more equations for a pixel? Spatial coherence constraint: pretend the pixel s neighbors have the same (u,v) If we use a 5x5 window, that gives us 25 equations per pixel B. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the International Joint Conference on Artificial Intelligence, pp. 674 679, 1981.

Solving the aperture problem Least squares problem: When is this sstem solvable? What if the window contains just a single straight edge? B. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the International Joint Conference on Artificial Intelligence, pp. 674 679, 1981.

Conditions for solvabilit Bad case: single straight edge

Conditions for solvabilit Good case

Lucas-Kanade flow Linear least squares problem Solution given b The summations are over all pixels in the window B. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the International Joint Conference on Artificial Intelligence, pp. 674 679, 1981.

Lucas-Kanade flow Recall the Harris corner detector: M = A T A is the second moment matrix We can figure out whether the sstem is solvable b looking at the eigenvalues of the second moment matrix The eigenvectors and eigenvalues of M relate to edge direction and magnitude The eigenvector associated with the larger eigenvalue points in the direction of fastest intensit change, and the other eigenvector is orthogonal to it

Interpreting the eigenvalues Classification of image points using eigenvalues of the second moment matrix: λ 2 Edge λ 2 >> λ 1 Corner λ 1 and λ 2 are large, λ 1 ~ λ 2 λ 1 and λ 2 are small Flat region Edge λ 1 >> λ 2 λ 1

Uniform region gradients have small magnitude small λ 1, small λ 2 sstem is ill-conditioned

Edge gradients have one dominant direction large λ 1, small λ 2 sstem is ill-conditioned

High-texture or corner region gradients have different directions, large magnitudes large λ 1, large λ 2 sstem is well-conditioned

Errors in Lucas-Kanade The motion is large (larger than a pixel) Iterative refinement Coarse-to-fine estimation Exhaustive neighborhood search (feature matching) A point does not move like its neighbors Motion segmentation Brightness constanc does not hold Exhaustive neighborhood search with normalized correlation

Feature tracking So far, we have onl considered optical flow estimation in a pair of images If we have more than two images, we can compute the optical flow from each frame to the next Given a point in the first image, we can in principle reconstruct its path b simpl following the arrows

Tracking challenges Ambiguit of optical flow Need to find good features to track Large motions, changes in appearance, occlusions, disocclusions Need mechanism for deleting, adding new features Drift errors ma accumulate over time Need to know when to terminate a track

Tracking over man frames Select features in first frame For each frame: Update positions of tracked features Discrete search or Lucas-Kanade (or a combination of the two) Terminate inconsistent tracks Compute similarit with corresponding feature in the previous frame or in the first frame where it s visible Find more features to track

Shi-Tomasi feature tracker Find good features using eigenvalues of secondmoment matrix Ke idea: good features to track are the ones whose motion can be estimated reliabl From frame to frame, track with Lucas-Kanade This amounts to assuming a translation model for frame-to-frame feature movement Check consistenc of tracks b affine registration to the first observed instance of the feature Affine model is more accurate for larger displacements Comparing to the first frame helps to minimize drift J. Shi and C. Tomasi. Good Features to Track. CVPR 1994.

Tracking example J. Shi and C. Tomasi. Good Features to Track. CVPR 1994.