Image Transfer Methods. Satya Prakash Mallick Jan 28 th, 2003

Similar documents
But First: Multi-View Projective Geometry

Lecture 6 Stereo Systems Multi-view geometry

Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15

Multiple Views Geometry

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Epipolar Geometry and Stereo Vision

Lecture 9: Epipolar Geometry

Computer Vision Lecture 17

Computer Vision Lecture 17

Lecture 14: Basic Multi-View Geometry

Epipolar Geometry and Stereo Vision

Structure from Motion and Multi- view Geometry. Last lecture

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Stereo vision. Many slides adapted from Steve Seitz

calibrated coordinates Linear transformation pixel coordinates

1 Projective Geometry

Step-by-Step Model Buidling

Epipolar Geometry and Stereo Vision

Unit 3 Multiple View Geometry

BIL Computer Vision Apr 16, 2014

Stereo Vision. MAN-522 Computer Vision

Image Processing: Motivation Rendering from Images. Related Work. Overview. Image Morphing Examples. Overview. View and Image Morphing CS334

Stereo and Epipolar geometry

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Rectification and Distortion Correction

Lecture 5 Epipolar Geometry

Epipolar Geometry and Stereo Vision

Assignment 2: Stereo and 3D Reconstruction from Disparity

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Lecture 9 & 10: Stereo Vision

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Two-view geometry Computer Vision Spring 2018, Lecture 10

Lecture'9'&'10:'' Stereo'Vision'

Rectification and Disparity

View Synthesis by Trinocular Edge Matching and Transfer

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

Epipolar geometry. x x

Shape as a Perturbation to Projective Mapping

A virtual tour of free viewpoint rendering

Rectification. Dr. Gerhard Roth

EE795: Computer Vision and Intelligent Systems

Projective Rectification from the Fundamental Matrix

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

Stereo Vision Computer Vision (Kris Kitani) Carnegie Mellon University

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

Lecture 10: Multi-view geometry

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

Multiple View Geometry. Frank Dellaert

FLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE

Stereo CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Final project bits and pieces

Cameras and Stereo CSE 455. Linda Shapiro

Midterm Examination CS 534: Computational Photography

Structure from motion

Image-Based Rendering Methods

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Geometry of Multiple views

N-Views (1) Homographies and Projection

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Dense 3D Reconstruction. Christiano Gava

Announcements. Stereo

Image warping and stitching

Multiple View Geometry in Computer Vision

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017

Computer Vision I. Announcement. Stereo Vision Outline. Stereo II. CSE252A Lecture 15

Chaplin, Modern Times, 1936

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

Introduction à la vision artificielle X

The end of affine cameras

Model-Based Stereo. Chapter Motivation. The modeling system described in Chapter 5 allows the user to create a basic model of a

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

EECS 442 Computer vision. Stereo systems. Stereo vision Rectification Correspondence problem Active stereo vision systems

Image Based Reconstruction II

Structure from motion

Stereo matching. Francesco Isgrò. 3D Reconstruction and Stereo p.1/21

Multiple View Geometry

Image Morphing. Application: Movie Special Effects. Application: Registration /Alignment. Image Cross-Dissolve

Two-View Geometry (Course 23, Lecture D)

Stereo. Many slides adapted from Steve Seitz

Mosaics wrapup & Stereo

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Image-Based Rendering

Announcements. Stereo

Application questions. Theoretical questions

IMAGE-BASED RENDERING TECHNIQUES FOR APPLICATION IN VIRTUAL ENVIRONMENTS

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry

Visual Recognition: Image Formation

Binocular stereo. Given a calibrated binocular stereo pair, fuse it to produce a depth image. Where does the depth information come from?

Rectification for Any Epipolar Geometry

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Stereo pairs from linear morphing

Computer Vision I. Announcements. Random Dot Stereograms. Stereo III. CSE252A Lecture 16

Perception and Action using Multilinear Forms

Dense 3D Reconstruction. Christiano Gava

1-2 Feature-Based Image Mosaicing

Transcription:

Image Transfer Methods Satya Prakash Mallick Jan 28 th, 2003

Objective Given two or more images of the same scene, the objective is to synthesize a novel view of the scene from a view point where there is no camera. Left Image Synthesized Images Right Image

Approaches 1. Traditional Approach: Generate a 3D representation of the scene Render the scene, given a view point Problems: Generating 3D from 2D images is generally very noisy. Generating 3D from 2D can be computationally prohibitive for real-time applications.

Approaches Light Field Rendering/ Lumigraph It s an elegant method in which one doesn t need to compute 3D information Problems: Calculation of the light field requires additional machinery. Calculation of light field can be very memory intensive.

Approaches Image Transfer Methods: Only 2D image operations are used to generate the novel view. 3D is never calculated explicitly. Its extremely fast. ( The OpenCV implementation of Stiez s algorithm works in real time!). Problems: Usually these methods cannot generate views from any arbitrary view point. There is a limitation on the quality of rendering that can be achieved. Ex. Its difficult to render using a different lighting condition.

Papers ( In Chronological Order) E. Chen, L. Williams,View Interpolation for Image Synthesis, SIGGRAPH 1993 S. Seitz, C. Dyer, View Morphing SIGGRAPH, pp. 21--30, 1996. S Avidan, A. Shashua Novel View Synthesis by Cascading Trilinear Tensors IEEE Transactions on visualization and computer graphics, Vol 4, No. 4, October-November 1998. Y. Genc, J. Ponce. Image-Based Rendering Using Parameterized Image Varieties. International Journal of Computer Vision, Vol. 41, No. 3, pp. 143-170, 2001.

What do these paper represent? 1993: Image based rendering was still at it s infancy. The first paper DOESNOT represent an image transfer method 1996: One of the first image transfer methods which used two images. However, only in-between views could be generated. 1998: An image transfer method using 2 cameras was introduced which use trilinear tensor to overcome the limitation of view generation along a line only. 2001: A new image transfer method is introduced. It relies on minimum parameterization of image space.

Theoretical background Image Morphing: (Term coined by T. Beier, 1992) [0,1], ) ( )C (1 ), W ( ) ( C ) (1 ), W ( Correspondence Maps : C : C : Input images, 1 1 0 0 1 1 1 1 1 0 0 0 0 0 0 1 1 1 0 0 1 0 + = + = s I p I p sp p s s p p s p s s p I I I I I I The warped image is an average of the pixel intensities obtained using forward warping and backward warping.

Physical Validity of Morphing If two images of the same object are morphed, the resulting image may not represent the same image geometrically.

Z - Buffering 1. Clear the color buffer to the background color 2. Initialize all xy coordinates in the Z buffer to one 3. For each fragment of each surface, compare depth values to those already stored in the Z buffer - Calculate the distance from the projection plane for each xy position on the surface - If the distance is less than the value currently stored in the Z buffer: Set the corresponding position in the color buffer to the color of the fragment Set the value in the Z buffer to the distance to that object - Otherwise: Leave the color and Z buffers unchanged

View Interpolation Given: Many images of the same scene. Their range data ( 3D is already given!). Camera Transformation Objective: To have a smooth navigation system through the scene ( Virtual Walkthrough ). Note: This is NOT an image transfer method.

View Interpolation The algorithm: 1. Pixel Correspondence Since Pixel s screen coordinates are known (x,y,z) and the relative orientation of the cameras is also known, a 4x4 transformation relates the points in one image to other. The above transformation can be stored as a offset vector for each pixel. This is called morph map. 2. Interpolating Correspondences: Offset vectors are interpolated linearly and the pixels in the source image are moved by the interpolated vector to the destination image.

View Interpolation The algorithm ( Contd. ) 2. Interpolating Correspondences: The interpolation is an approximation of pixel coordinate transformation by a perspective viewing matrix. 3. Composting Images: Visibility Resolution: Z-buffering can be used. A view independent visible priority can be used. Holes: Can be filled using interpolation of color from neighboring pixels. 4. Block Compression: Neighboring pixel move together. Hence morphing map can be defined for blocks of images rather than pixels.

Results

Results

Drawbacks For real scenes having dense disparity or/and perfect correspondence between pixels is extremely difficult. The underlying assumptions make it suitable only for synthetic data. No wonder, only synthetic results were shown.

The Geometry of Multiple Views Epipolar Geometry The Essential Matrix The Fundamental Matrix The Trifocal Tensor

Epipolar Geometry Epipolar Plane Epipolar Lines Epipoles Baseline

Epipolar Constraint Potential matches for p have to lie on the corresponding epipolar line l. Potential matches for p have to lie on the epipolar line l.

Epipolar Constraint: Ideal Case Essential Matrix (Longuet-Higgins, 1981)

Properties of the Essential Matrix Ep is the epipolar line associated with p. T E p is the epipolar line associated with p. E e =0 and E e=0. E is singular. T E has two equal non-zero singular values (Huang and Faugeras, 1989).

Epipolar Constraint: Real Case Fundamental Matrix (Faugeras and Luong, 1992)

Image Rectification

An Example Due to Josh Wills

Image Rectification To make the epipolar lines parallel to each other and aligned to one of the scanlines, we should send the epipoles to infinity. Rectified image is obtained by a rotation of the original image keeping the optical center fixed. Hence, they are related by a homography.

Image Rectification Such a Homography is given by: H = GR, R = a rotation which takes epipole(e) to ( T f,0,1) G= 1 0 1/ f 0 1 0 0 0 1 As is easily seen, T T G* ( f,0,1) = ( f,0,0) = point at infinity If the images are related by a fundamental matrix, F = [ e] M, Then the homography which ensures that both images are aligned is given by: H ' = ( I + Hea T ) HM

View Morphing Image morphing is shape preserving only in the special case when the images are taken using cameras having parallel optical axes. An additional step called image rectification is introduced. Image interpolation on rectified images produces physically valid in-between images. Rectification is a process of aligning corresponding epipolar lines in the two images along their scanlines.

The algorithm Find a few point correspondences in both images. Estimate the fundamental matrix using 8- point algorithm along with RANSAC. Rectify both the images. Find dense correspondence and morph the rectified images Post warp the morphed image. Remove holes ( moirs ).

Step By Step

Step By Step

Surprise!

Some Results

Novel View Synthesis using Trilinear Tensors What is a Tensor? Any array of numbers: A scalar is a zero dimensional tensor, a vector is a one dimensional tensor, a matrix is a two dimensional tensor and so on. A Trilinear Tensor is a 3x3x3 array. Three views satisfy certain matching constraints represented by a tensor. Given two views in correspondence and a tensor, the corresponding third view can be generated by means of a warping function.

Constraint Equation p i s µ j r ρ k T i jk = 0

More Hand waiving! Let p, p are known. i jk Then p s µ j Ti is a point that coincides with all lines passing through p. Hence, given images i p, p of a point in two views and a tensor we can find the image of the point in the third view by the reprojection equation given by: i µ jk k p s j Ti p"

Basic Tensor Operator Basic Tensor Operator describes how to modify a tensor so as to represent a new set of cameras. Let, [I;0], [A,V],[B,V ] and [C,V ] be the camera matrices associated with the views in question. Let be jk T i the jk tensor between views 1,2,3 and G i be the tensor between views 1,2,4. If the motion parameters between views 3 and 4 are represented by D, then we have a relationship: jk k jk G i = di Ti + Hence, corresponding to different D s, we can have different viewpoints. t k a j i

The Algorithm Pre-processing steps: Compute dense correspondence (optical-flow) between the pair of reference images. Construct the trilinear tensor of the two images Extract the rotation matrix of the two images from the trilinear tensor. View Synthesis steps: Accept the rotation and translation of the novel image. Construct the new tensor Reproject the novel image.

Results

Rendering using PIV Punch Line: The set of all images of a rigid set of m points and n lines observed using a weak perspective camera can be represented by a 2( m n six dimensional variety embedded in R + and can be parameterized by the image coordinates of three reference points ) Note: Loosely speaking, a variety can be thought to be a subspace

Basic Ideas Given: a) Images of a rigid scene taken from different view points. These images are called training images. b) A set of point( m in no. ) and line ( n in no. ) correspondences in all training images. c) Coordinates of 3 reference points in all images.

The Algorithm The position of m points and the position and orientation of n lines can be used to calculate some parameters which define the scene. Training Image1 Training Image2 Training Image3 Parameters p1, p2,p3..

The view point of the image to be synthesized is expressed by the location of the three points in the synthesized image. Training Image1 Training Image2 Training Image3 Synthesized Image The positions of three points in the Synthesized image specifies the viewpoint ( Instead of the normal viewing matrix or R and T)

The parameters can then be used to calculate the position and orientation of all m points and n lines in the synthesized image. Parameters p1, p2, p3. position of points which specify the viewpoint + Synthesized Image: We get the position and orientation of lines and points in the synthesized image.

Training Image1 Training Image2 Training Image3 Synthesized Image Triangulation based Rendering Complete Synthesized Image

Results