Silhouette Coherence for Camera Calibration under Circular Motion

Similar documents
Silhouette Coherence for Camera Calibration under Circular Motion

Silhouette Coherence for Camera Calibration under Circular Motion

Using silhouette coherence for 3D image-based object modeling under circular motion

CALIBRATING A NETWORK OF CAMERAS FROM LIVE OR ARCHIVED VIDEO. Sudipta N. Sinha, Marc Pollefeys

Camera Calibration from Dynamic Silhouettes Using Motion Barcodes

Multi-view stereo. Many slides adapted from S. Seitz

Visual-Hull Reconstruction from Uncalibrated and Unsynchronized Video Streams

3D Modeling using multiple images Exam January 2008

Silhouette-based Multiple-View Camera Calibration

Efficiently Determining Silhouette Consistency

Motion Segmentation Based on Factorization Method and Discriminant Criterion

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

1D camera geometry and Its application to circular motion estimation. Creative Commons: Attribution 3.0 Hong Kong License

Spherical Embedding of Inlier Silhouette Dissimilarities

Shape-from-Silhouette with Two Mirrors and an Uncalibrated Camera

Spherical Embedding of Inlier Silhouette Dissimilarities

arxiv: v1 [cs.cv] 28 Sep 2018

Midterm Exam Solutions

Euclidean Reconstruction Independent on Camera Intrinsic Parameters

Precise Omnidirectional Camera Calibration

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Shape from Texture: Surface Recovery Through Texture-Element Extraction

Overview. Related Work Tensor Voting in 2-D Tensor Voting in 3-D Tensor Voting in N-D Application to Vision Problems Stereo Visual Motion

A Statistical Consistency Check for the Space Carving Algorithm.

A Desktop 3D Scanner Exploiting Rotation and Visual Rectification of Laser Profiles

Local quadrics surface approximation for real-time tracking of textureless 3D rigid curved objects

Computer Vision Lecture 17

Computer Vision Lecture 17

Chapter 3 Image Registration. Chapter 3 Image Registration

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

Filtering Images. Contents

Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Volumetric stereo with silhouette and feature constraints

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

CS 223B Computer Vision Problem Set 3

Camera Network Calibration from Dynamic Silhouettes

Chaplin, Modern Times, 1936

Motion Segmentation Based on Factorization Method and Discriminant Criterion

BIL Computer Vision Apr 16, 2014

3D Shape Recovery of Smooth Surfaces: Dropping the Fixed Viewpoint Assumption

Edge and local feature detection - 2. Importance of edge detection in computer vision

Step-by-Step Model Buidling

Gaze interaction (2): models and technologies

Constructing a 3D Object Model from Multiple Visual Features

Stereo: Disparity and Matching

Stereo Vision. MAN-522 Computer Vision

Reconstruction of sculpture from its profiles with unknown camera positions

3D and Appearance Modeling from Images

Factorization with Missing and Noisy Data

Motion Tracking and Event Understanding in Video Sequences

arxiv: v1 [cs.cv] 14 Apr 2017

CS 231A Computer Vision (Fall 2012) Problem Set 3

A Novel Stereo Camera System by a Biprism

GOPRO CAMERAS MATRIX AND DEPTH MAP IN COMPUTER VISION

STRUCTURE AND MOTION ESTIMATION FROM DYNAMIC SILHOUETTES UNDER PERSPECTIVE PROJECTION *

Making Machines See. Roberto Cipolla Department of Engineering. Research team

3D Accuracy Improvement from an Image Evaluation and Viewpoint Dependency. Keisuke Kinoshita ATR Human Information Science Laboratories

A Stratified Approach for Camera Calibration Using Spheres

Camera calibration with two arbitrary coaxial circles

Image Analysis. Edge Detection

Camera Calibration from Surfaces of Revolution

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

3D Tracking Using Two High-Speed Vision Systems

Camera Calibration with a Simulated Three Dimensional Calibration Object

PACS Spectrometer Simulation and the Extended to Point Correction

FAIR: Towards A New Feature for Affinely-Invariant Recognition

Lecture 10: Multi-view geometry

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Topological Mapping. Discrete Bayes Filter

3D Computer Vision. Dense 3D Reconstruction II. Prof. Didier Stricker. Christiano Gava

13. Learning Ballistic Movementsof a Robot Arm 212

INFERENCE OF SEGMENTED, VOLUMETRIC SHAPE FROM INTENSITY IMAGES

Fundamental Matrices from Moving Objects Using Line Motion Barcodes

Multi-View Reconstruction using Narrow-Band Graph-Cuts and Surface Normal Optimization

Structure from Motion and Multi- view Geometry. Last lecture

Comments on Consistent Depth Maps Recovery from a Video Sequence

3D Models from Contours: Further Identification of Unexposed Areas

Image Analysis. Edge Detection

Object and Motion Recognition using Plane Plus Parallax Displacement of Conics

Optic Flow and Basics Towards Horn-Schunck 1

Multiple View Geometry in computer vision

A Calibration Algorithm for POX-Slits Camera

Project 4 Results. Representation. Data. Learning. Zachary, Hung-I, Paul, Emanuel. SIFT and HoG are popular and successful.

Reconstruction of sculpture from its profiles with unknown camera positions

Peripheral drift illusion

6-DOF Model Based Tracking via Object Coordinate Regression Supplemental Note

Camera Geometry II. COS 429 Princeton University

Multi-stable Perception. Necker Cube

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

Tracking human joint motion for turntable-based static model reconstruction

Structure and Motion Estimation from Apparent Contours under Circular Motion

Camera Calibration and Light Source Estimation from Images with Shadows

Edge detection. Convert a 2D image into a set of curves. Extracts salient features of the scene More compact than pixels

Combinatorial optimization and its applications in image Processing. Filip Malmberg

A Factorization Method for Structure from Planar Motion

Depth from Refraction Using a Transparent Medium with Unknown Pose and Refractive Index

Transcription:

Silhouette Coherence for Camera Calibration under Circular Motion Carlos Hernández, Francis Schmitt and Roberto Cipolla Appendix I

2 I. ERROR ANALYSIS OF THE SILHOUETTE COHERENCE AS A FUNCTION OF SILHOUETTE NOISE In this experiment we have added noise to the exact silhouette contours in order to statistically measure the accuracy of the epipolar tangency criterion and the silhouette coherence criterion. The noise has been simply added along the normal of the contours. The amplitude of the noise is computed as a uniform noise smoothed by a Gaussian filter, which avoids unrealistic jaggedness along the silhouette contour. The noise variance is computed after the Gaussian filtering. For each noise standard deviation σ, we have computed 2 samples in order to obtain reliable results. We have estimated the rotation axis, the translation direction and the focal length. We have not recovered the camera angles in order to avoid too many outliers due to local minima. The implemented epipolar tangency criterion has the following form: n C et = n i= j N (i) K K ij C et (S i, S j ), () ij i= j N (i) if e ij S i or e ji S j K ij =, else where n is the number of available silhouettes, N (i) is the set of peer silhouettes for the silhouette i, C et (S i, S j ) is the epipolar tangency coherence between two silhouettes, and K ij takes into account if we can compute C et (S i, S j ) or not. In the particular case where the epipole is inside a silhouette, C et (S i, S j ) cannot be computed so C et is weighted accordingly. It remains to precise the meaning of N in Eqn. (). If we only want to compare each silhouette against the n others, then N all (i) = {i+,, n}. However, this is not the original definition given used in [3], where N 3 (i) = {i +,, min(i + 3, n)} is used. Although this choice is not justified, it is probably due to the resulting computation time saving (N 3 is faster than N all ) and to the convergence properties of C et : using a small number of silhouettes around the current silhouette should help to smooth the energy shape. But this leads to a worse accuracy than when using all the silhouettes. We show the rotation axis estimation, the translation angle recovery and the focal length estimation results infig. 2. We have tested the epipolar tangency criterion with 4 different sets of peer silhouettes N (see Eqn. ): silhouette N (i) = {min(i +, n)}, 3 silhouettes N 3 (i) = {i+,, min(i+3, n)}, 6 silhouettes N 6 (i) = {i+,, min(i+6, n)} and all the silhouettes

3 N all (i) = {i +,, n}. Note that, although the total number of silhouettes is 2, N 6 and N all are not the same because silhouette and silhouette 2 will never be compared using N 6 but they will using N all. Concerning the silhouette coherence criterion, we have tested two different values of the sampling offset δ: δ =.5 pixels and δ = pixels. The criterion used to measure the error between the recovered rotation axis a recovered and the real axis a in Fig. 2 is the angle between both axes: acos(a recovered a). The estimation results of the camera motion (rotation axis and translation direction) are shown in figures 2 left and middle. For large noise standard deviation (σ >.3 pixels), the silhouette coherence criteria (δ =.5 pixels and δ = pixels) perform better than any of the epipolar tangency criteria. This is justified by the fact that the silhouette coherence criterion uses the entire contours for the computation while the epipolar tangency criterion is actually using only the epipolar tangent points. However, the behavior of the silhouette coherence changes with low noise, where the epipolar tangency criterion performs better, and what is even more surprising, the silhouette coherence curves are not monotone with the noise standard deviation: e.g., for δ =, we obtain better accuracy with σ =.3 than with σ =.. This strange behavior is caused by the rapid saturation of the silhouette coherence criterion. For small noise, the silhouette coherence is very easily maximized if δ is not small enough. As a result, around the optimum there is a constant platform whose size depends on the values of δ and σ. If we compare the results for the translation estimation, i.e., the α t parameter, we observe a strong dependence of the epipolar tangency criterion on the number of silhouettes used (see Fig. 2 middle). The epipolar tangency criterion needs a large baseline to estimate α t accurately. This is justified by the fact that, when the baseline is small, the epipolar tangencies change very little with large variations of α t. Even when using all the possible silhouette pairs for the epipolar tangency criterion (N = N all ), the silhouette coherence criterion performs much better than the epipolar tangency one (see Fig. 2 middle). Finally, we show the focal length estimation results in Fig. 2 right. Both the epipolar tangency criterion (with N all ) and the silhouette coherence criterion perform very well, with less than 3% of error for a noise standard deviation of pixel. An interesting remark is the fact that, with δ = pixel, the focal length error is almost the same as with σ =. and σ =. pixels. As explained before, it shows the saturation effect of the silhouette coherence criterion when the

4 2 4 6 8 2 4 6 8 2 22 Fig.. Teapot sequence with 2 views, their corresponding exact silhouettes and their camera angles in..4.2..8.6.4 rotation axis mean error epipolar tangencies ± silhouette coherence δ=.5.3.25.2.5. translation direction mean error epipolar tangencies ± silhouette coherence δ=.5 6 5 4 3 2 focal length mean error epipolar tangencies epipolar tangencies 3 epipolar tangencies 6 silhouette coherence δ=.5.2.5..2.3.4.6.8.4.2..8.6.4.2 rotation axis standard deviation epipolar tangencies ± silhouette coherence δ=.5.25.5.5..2.3.4.6.8.3.2. translation direction standard deviation epipolar tangencies ± silhouette coherence δ=.5..2.3.4.6.8 6 5 4 3 2 focal length standard deviation epipolar tangencies epipolar tangencies 3 epipolar tangencies 6 silhouette coherence δ=.5..2.3.4.6.8..2.3.4.6.8..2.3.4.6.8 Fig. 2. Rotation axis, translation direction and focal length estimation error as a function of the noise standard deviation σ. Top: mean error. Bottom: standard deviation. values of δ are large compared to the noise of the silhouettes. II. ERROR ANALYSIS OF THE SILHOUETTE COHERENCE AS A FUNCTION OF OBJECT SHAPE In order to show why the silhouette coherence criterion exploits more information than just epipolar tangents alone, we have created a synthetic object with a parameterized shape. The object is composed of two items: a deformed tube and an ellipsoid. Since the tube length is larger than the image height, the tube silhouettes do not provide any epipolar tangency. The only possible epipolar tangencies may come from the ellipsoid. We note that, in order to exploit the epipolar tangencies of this configuration, we would need to handle the visibility of the epipolar

5 tangencies and match them between different views as proposed in [], [2]. We have parameterized the shape of the ellipsoid with a s, and we have computed 6 different synthetic objects for s = {,.2,.4,.6,.8,.} (see Fig. 3). For each synthetic object, we have generated a sequence of 8 silhouettes, using a constant camera angle step of 2. Then, for each silhouette sequence, we have studied the accuracy of the silhouette coherence to recover the rotation axis, the translation and the focal length with a noise standard deviation of.5 pixels. To obtain reliable results, we have done 2 trials per. We show the estimation results in Fig. 4. For the s s = and s =.2 there are no epipolar tangency points at all, i.e., algorithms based on the notion of epipolar tangency will not work. However, we show in Fig. 4 that the silhouette coherence is still able to recover the translation and the focal length with very good accuracy. The recovery of the rotation axis shows an interesting behavior: the accuracy of φ a is very good (see dashed curve in Fig. 4 top) while the accuracy of θ a strongly depends on the (see solid curve in Fig. 4 top). This behavior is expected since the silhouettes of the tube almost do not change for strong differences of θ a, which corresponds to the angle between the rotation axis a and the camera viewing axis z. As we increase s, we introduce cues in the silhouettes about the rotation axis, since the silhouettes of the ellipsoid do depend on θ a. This experiment shows how the silhouette coherence is able to automatically extract the maximum information of the silhouettes, without the need of matching or handling the visibility of any epipolar tangency point, and even when no epipolar tangency point is available at all. It also shows that epipolar tangency points do encode a substantial amount of the silhouette information, but not all the information a silhouette can provide. REFERENCES [] Y. Furukawa, A. Sethi, J. Ponce, and D. Kriegman, Structure and motion from images of smooth textureless objects, in ECCV 24, vol. 2, Prague, Czech Republic, May 24, pp. 287 298. [2] S. N. Sinha, M. Pollefeys, and L. McMillan, Camera network calibration from dynamic silhouettes, in CVPR, vol., 24, pp. 95 22. [3] K.-Y. K. Wong and R. Cipolla, Structure and motion from silhouettes, in 8th IEEE International Conference on Computer Vision, vol. II, Vancouver, Canada, July 2, pp. 27 222.

6 s = s =.2 s =.4 s =.6 s =.8 s =. Fig. 3. Different synthetic objects created with one deformed tube and one ellipsoide as a function of the ellipsoid s. Top: front view. Bottom: bottom view..4 rotation axis mean error θ a.5 translation direction mean error.4 focal length mean error φ a.3.2. (x 3 ).5.3.2..2.4.6.8.4 rotation axis standard deviation θ a φ a.2.4.6.8.5 translation direction standard deviation.2.4.6.8.4 focal length standard deviation.3.2. (x 3 ).5.3.2..2.4.6.8.2.4.6.8.2.4.6.8 Fig. 4. Rotation axis, translation direction and focal length estimation error as a function of the ellipsoid s. Top: mean error. Bottom: standard deviation.