Closed-form Linear Solution To Motion Estimation In Disparity Space

Size: px
Start display at page:

Download "Closed-form Linear Solution To Motion Estimation In Disparity Space"

Transcription

1 Closed-form Linear Solution To Motion Estimation In Disparity Space Konstantinos G. Derpanis Vision Technologies Sarnoff Corporation Princeton, NJ 8543 Peng Chang ABSTRACT Real-time stereovision systems play an important role in automotive related applications. This paper concerns the problem of rigid motion estimation with a stereovision sensor. Given a set of corresponding 3D points in Euclidean space reconstructed from stereovision, efficient linear algorithms exist to solve for the rigid motion. However it has been well known that the noise in the Euclidean reconstruction from stereovision is heteroscedastic and anisotropic, therefore the linear algorithms are only sub-optimal. Recently a d-motion based algorithm has been developed to solve for the rigid motion directly in the disparity space, in which the noise can be approximated to be homogenous and isotropic. However this algorithm is nonlinear and requires iterative least-squares solution. By reformulating the problem, a closed-form linear solution is presented in this paper to solve for the rigid motion in disparity space. Synthetic experimental results show that this new algorithm outperforms the d-motion based algorithm in terms of both accuracy and computational cost. We believe that the closed-form linear solution is potentially very useful for applications making use of stereovision to estimate rigid motion. I. INTRODUCTION Stereovision based approaches have played a very important role in automotive related applications, and recently gained increased popularity due to the fact that the real-time stereovision systems are becoming increasingly affordable [], [2], [3], [4], [5], [6], [7]. In this paper, we focus on the problem of rigid motion estimation with a stereovision sensor. To estimate the rigid motion from a stereovision sensor, a typical approach often involves extracting and matching image feature points in subsequent images from a designated camera of the stereo pair. Meanwhile, the disparities of those points are obtained from the stereo matching process. As a result, the correspondences in 3D across time can be established, which provide enough information to infer the 3D rigid motion. Many previous approaches perform the motion estimation with the 3D reconstruction in Euclidean space [8], [9], []. Given corresponding 3D point sets, the rigid motion can be estimated with closed-form least-squares solutions as reviewed in []. However, least-squares algorithms are optimal only when the data is corrupted with independent and identically distributed (i.i.d.), zero mean and isotropic noise, and give poor estimation with the stereo-reconstructed 3D points, which are known to be corrupted with heteroscedastic and anisotropic noise. In [2], a heteroscedastic error in variables (HEIV) based algorithm is presented, which yields optimal rigid motion estimates with 3D points reconstructed from stereo, however such method requires estimation of the underlying uncertainty for each reconstructed 3D point, which itself is nontrivial. Alternatively, methods have been proposed to estimate the rigid motion directly in the disparity space, in which the noise can be reasonably assumed to be isotropic and i.i.d. [3], [4]. In [3], a d-motion based algorithm has been presented to solve for the rigid motion in disparity space directly. Unfortunately, this algorithm is nonlinear and uses an iterative weighted least-squares method, which is not guaranteed to converge. In addition, it is assumed to only handle small rotations. In this paper, we present a closed-form linear algorithm to solve for general rigid motion in disparity space. The synthetic experimental results clearly show its advantage over the previous approach. In our approach, the rigid motion estimation is based on discrete feature points. Alternatively, dense disparity maps and grey scale images can be used to perform the rigid motion estimation [5], [6]. Our primary concern is computational cost, while detailed comparison is beyond the scope of this paper. In Section II the disparity space formulation and the previous algorithm as in [3] are first reviewed. Then the closedform linear solution is presented, which come in two possible forms. In Section III synthesized experiments are carried out, and comparisons are made among the new algorithm and previous algorithms. In Section IV discussions are made and future work outlined. II. ESTIMATION IN DISPARITY SPACE Rather than estimating the rigid motion in the Euclidean space the points are corrupted by heteroscedastic noise, Demirdjian and Darrell [7] demonstrate that the estimation can be done in disparity space, the noise is purported to be independent and isotropic for parallel stereo images. We first review the results in their work.

2 A. Disparity Space In homogeneous coordinates the rigid transformation (R rotation and t translation) that maps point P = (X, Y, Z) to P = (X, Y, Z ) is given as follows, ( ) ( ) ( ) P R t P = () Let ω = (x, y, d) be a point in disparity space, (x, y) represents the image coordinate of a point on the left image with respect to the optical center, and d the disparity. After a little matrix manipulation, we arrive at, ( ) ( ) ω P Γ (2) Γ = B (3) and indicates the transformation is up to a scale factor. And and represent the camera focal length expressed in the effective horizontal and vertical pixel size, respectively, and B the camera baseline. Both the baseline and focal length are assumed known. Similarly for the point P, ( ) ( ) ω P Γ (4) Substituting the inverse of (2) and (4) into (), yields ( ) ( ) ω ω H(R, t) ( R t H(R, t) = Γ (5) ) Γ (6) H(R, t) is a homography dependent on the 3D rotation R and translation t. B. Previous Nonlinear Algorithm As pointed out in the literature, the noise in the Euclidean space reconstructed from stereo is heteroscedastic and anisotropic, which in turn calls for HEIV (Heteroscedastic Error In Variables) based method for motion estimation. Alternatively, the noise in the disparity space is much betterbehaved, and can be reasonably assumed to be homogenous and isotropic, which allows for more efficient algorithms. We first review the algorithm proposed in [7], then present a closed-form linear solution to the motion estimation problem in disparity space, which turns out to be much more efficient and accurate in comparison. In [7] the covariance of the noise corrupting the disparity space coordinate is given by Λ = σ 2 I, I is the identity matrix. Given a list of N correspondences in the disparity space {(ω i, ω i ), i =...N}, the estimation of rigid motion in [7] is formulated as a non-linear least-squares minimization problem, and the error metric is based on the summarized squared error of the so called d-motion. In addition, the authors focus on the case of small motion and approximate the rotation R as follows, R = I + ω c ω b ω c ω a (7) ω b ω a Note that this matrix is not orthogonal and thus not truly a rotation matrix. A procedure to orthogonalize the estimated rotation matrix is summarized below. The total error metric E 2 is formulated as,, E 2 = i w 2 i P i u + v i (8) u = (t x, t y, t z, ω a, ω b, ω c ), (9) and w i is a scalar, v i a 3-vector and P i a 3 6 matrix which all depend on ω i and ω i. Given this form of the total error E 2, an iterative reweighted least-squares estimation algorithm is given: Algorithm Disparity space re-weighted least-squares estimation[7] : Initialize w i () =. Estimate u. 2: Evaluate w i (k + ) from the current solution for u. 3: Minimize E 2 (k + ) using a standard weighted leastsquares method while keeping w i fixed. 4: Termination estimation if E 2 (k + ) E 2 (k) ɛ then stop, else return to Step 2; ɛ is a user defined threshold. The author also pointed out that this approach is not guaranteed to converge. A problem with the approach as detailed above is that the rotation matrix (7) is not orthonormal. A procedure for converting the estimated rotation matrix (7) to the closest orthonormal matrix (in the Frobenius-norm sense ) is found by computing the SVD of the estimated rotation matrix and setting the matrix of singular values W to the identity matrix, yielding, ˆR = UWV () UIV () UV (2) In the case of larger motions, a global non-linear minimization method is required such as Levenberg-Marquardt [8], [9]. The closest matrix in the Frobenius-norm sense entails minimizing the least-squares distance between the individual components of two matrices.

3 C. Linear solution in disparity space In this section a novel linear solution is proposed for estimating the Euclidean rigid motion using the disparity space representation. Let (x i, y i, d i ) be the points in disparity space selected from the video frame at time t = j and (x i, y i, d i ) be their correspondents taken from the frame at t = j+, i =,..., n. Let m = ( x y d ) and ( x2 y 2 d 2 ) be the points in homogeneous coordinates (for clarity the index of the measurements has been dropped). The mapping from the points at t = j to t = j + is a 4 4 homography matrix [7] given by (5). Assuming that there is no noise present in the observations, the following constraint holds, x 2 = h m h 4 m y 2 = h 2 m h 4 m d 2 = h 3 m h 4 m (3) h k denotes the k th row of the homography H (index of measurements dropped for clarity). Rewriting (3) yields the following system of linear equations, Z i = and Z i θ = (4) x i y i d i... x i y i d i... x i... x 2i x i y 2i x i x 2i y i y 2i y i x 2i d i y 2i d i x 2i y 2i y i d i d 2i x i d 2i y i d 2i d i d 2i (5) θ = ( h h 2 h 3 h 4 h 5 h 6 h 7 h 8 h 9... h h h 2 h 3 h 4 h 5 h 6 ) (6) which is a vector consisting of the components of H. Given ( all the point correspondences we define matrix Z = ) Z. Z2 Zn Each match contributes three equations to the homography constraint, therefore the minimum number of measurement correspondences required for the estimation of the homography is five; due to the fact that the homography is defined up to a multiplicative scalar. A closed-form 2 estimate of Ĥ in the presence of noise is obtained by computing the smallest singular vector of the matrix Z. Since linear solutions for homography estimates are 2 In fact, this algorithm is stictly non-iterative if one discounts the solution of the eigensystem or the determination of the singular value decomposition which is used in computing the singular values. notoriously unstable for significant amounts of noise, the measurements in both video frames must be centered and scaled [2]. Note that the linearization introduces heteroscedastic noise [2], [22]; the normalization represents only a partial solution. From a statistical point of view better solutions exist (e.g., [2]) in the literature but they require the non-trivial estimation of the heteroscedastic noise characteristics prior to parameter estimation and are iterative computations which may preclude them from real-time applications. Let and represent the camera focal length and B the camera baseline, both are assumed to be known. The homography H in (6) can be expressed as follows, H(R, t) = r r 2 r 2 r 22 t y t xb r 3 B r 23 r 3 r 32 t z B r 33 (7) r ij represent the elements of the rotation matrix R and t = (t x, t y, y z ). Let the estimated homography be represented as follows, Ĥ = ĥ ĥ 2 ĥ 3 ĥ 4 ĥ 5 ĥ 6 ĥ 7 ĥ 8 ĥ 9 ĥ ĥ ĥ 2 ĥ 3 ĥ 4 ĥ 5 ĥ 6 (8) Given the estimated homography (8) and the homography parameterized in terms of the rigid motion parameters and the known camera parameters (7), the estimate of the rigid motion parameters are recovered as follows, ˆR = ĥ ĥ 2 ĥ 4 ĥ fy ĥ 5 ĥ 8 6 ĥ 3 ĥ 4 ĥ 6 ˆt = Bĥ3 B ĥ 7 Bĥ5 (9) (2) We term this approach the Normalized Disparity Space Direct Linear Transform (N). An alternative solution, is to fix h 9 = h = h 2 = and h = by leveraging the structure of the disparity space homography (7), this yields the following three constraints (two homogeneous plus a homogeneous constraint), Z i = Z i θ = c (2) x i y i d i... x i y i d i x i x 2i y i x 2i d i x 2i x 2i x i y 2i y i y 2i d i y 2i y 2i x i d 2i y i d 2i d i d 2i d 2i (22)

4 c i = d i (23) θ = ( h h 2 h 3 h 4 h 5 h 6 h 7 h 8 h 3 h 4 h 5 h 6 ) (24) which is a vector consisting of the components of H. Here only four points correspondences (as opposed to five as in N) are required for the solution, given as, θ = Z c (25) Z = ( ) Z Z2 Zn ( ) and c = c,... c n. If more than four points matches are used, yielding an overdetermined system of equations, the inverse in (25) is replaced with the pseudoinverse. Given the solution to θ the estimate of the rigid motion parameters are recovered as follows, ˆR = ĥ ĥ 2 ĥ 4 ĥ fy ĥ 5 ĥ 8 6 ĥ 3 ĥ 4 ĥ 6 ˆt = Bĥ3 B ĥ 7 Bĥ5 (26) (27) We term this approach the Disparity Space Direct Linear Transform (). Both N and are followed by the step to orthogonalize the estimated rotation matrix as described in Section II-B. A disadvantage of the is that it does not allow for the inclusion of the Hartley normalization [2] as in the N solution. However, often in real applications, a necessary first step is to reject the outliers in the 3D correspondence set, which is usually achieved by a RANSAC process. When used as a front end to generate the initial motion hypothesis in the RANSAC process, the fact that requires one less correspondence point than N is appealing, since the smaller the number of minimum points required, the smaller amount of samples are required by the RANSAC process, which in turn translates into more efficient implementation. III. EXPERIMENTAL RESULTS Experiments with simulated data are carried out to compare the performance of proposed linear algorithms with the previous algorithms, in particular, the nonlinear algorithm in disparity space as in [23] and the well-known SVD based algorithm in Euclidean space as in [24]. Apparently the algorithm performance depends on many variables, such as noise level, underlying motion and even the scene structure. Here we simulate the scenario when a stereo sensor is mounted on a host car and detects imminent collision with other cars. The relative motion is critical in deciding whether the two cars will collide or not, and it can be estimated from the 3D point correspondence set from the stereo sensor. With this application in mind, a synthesized scene of points is used, which are randomly distributed inside a cuboid with roughly the size of a car. A random rigid motion is then generated, simulating a car moving on the ground plane. The 3D ground truth points are projected onto the stereo image planes, generating the ground truth points in disparity space. Then Gaussian noise (at the image level) with varying standard deviation is added to the points in disparity space, to obtain the noise corrupted correspondence set in disparity space, which is used as the input to the motion estimation algorithms in disparity space. By triangulation with the noise corrupted point in disparity space, we can obtain the noise corrupted 3D reconstruction in Euclidean space, which is used as the input to the SVD based algorithm for comparison. To obtain statistically meaningful comparison, at each noise level, we repeat the process for 5 times, and compare the absolute mean and standard deviation of the estimation error. The translation error is computed as the angle subtended by the ground truth translation vector and the estimated translation vector. The rotation error is computed as the difference in the ground truth rotation angle and the estimated rotation angle along the ground truth rotation axis. Figure shows the comparison of the N,, and with increasing noise levels, refers to the iterative least-squares algorithm, and is followed by a Levenberg-Marquardt nonlinear optimization. The underlying rigid motion is t = (.5.5) and with a rotation angle of 5 about the axis (), X, Z axis are parallel to the ground plane and Y axis vertical to the ground. Figure shows the N and algorithms have very similar performance under different noise levels, with N generating slightly better result in the absolute mean error in the estimate rotation angle. Both N and outperform the iterative least-squares method by a large margin. While the Levenberg-Marquardt optimization can improve the iterative least-squares result, its rotation angle estimation still has much larger standard deviation than N and algorithms. Figure 2 shows the comparison result under the same setting, except that the rotation angle is now increased to. As it can be seen from the figure, under large rotation, the Levenberg-Marquardt optimization cannot improve the iterative least-squares result, possibly due to the wrong starting point. Nonetheless N and still generate quite consistent motion estimates. IV. CONCLUSION AND DISCUSSION In this paper, we present a novel closed-form linear algorithm to solve for rigid motion in the disparity space. To our knowledge this is the first linear algorithm for rigid motion estimation in disparity space, which can be potentially very useful for real-time automotive applications, due to its high efficiency.

5 2 Mean Translation Error.4 STD of Translation Error N N Mean Rotation Error.25 STD of Rotation Error N N Fig.. Comparison of N and with D-motion based methods under different noise levels To apply any rigid motion estimation algorithm to real applications, at least one hurdle has to be overcome, the outliers in the correspondence set. RANSAC has been a popular method in rejecting the outliers. The efficient closedform linear algorithm then can be used in the initial hypothesis generation step to reject the outliers, which otherwise can be extremely expensive if nonlinear method is used. After the inliers are located, more expensive and optimal nonlinear algorithm can be applied to achieve better result. REFERENCES [] Z. Sun, G. Bebis, and R. Miller, On-road vehicle detection using optical sensors: a review, 24. [2] M. Bertozzi, A. Broggi, A. Fascioli, and S. Nichele, Stereo vision-based vehicle detection, 2. [3] M. Suwa, Y. Wu, and M. Kobayashi, A stereo-based vehicle detection method under windy conditions, 2. [4] N. Srinivasa, A vision-based vehicle detection and tracking method for forward collision warning, 22. [5] D. Gavrila, Pedestrian detection from a moving vehicle, in European Conference on Computer Vision, 2, pp [6] M. Bertozzi, A. Broggi, R. Chapuis, F. Chausse, A. Fascioli, and A. Tibaldi, Shape based pedestrian detection and localization, 23. [7] S. Nedevschi, R. Danescu, T. Marita, F. Oniga, C. Pocol, S. Sobol, T. Graf, and R. Schmidt, Driving environment perception using stereovision, in IEEE Intelligent Vehicles Symposium, 25. [8] L. Matthies and S. Shafer, Error modeling in stereo navigation, IEEE Trans. on Rob. and Auto., vol. 3, pp , 987. [9] C. Olson, L. Matthies, M. Schoppers, and M. Maimone, Robust stereo ego-motion for long distance navigation, in CVPR, 999. [] P. Chang, T. Camus, and R. Mandelbaum, Stereo based vision system for automotive imminent collision detection, in IEEE Intelligent Vehicles Symposium, 24. [] D. Eggert, A. Lorusso, and R. Fisher, Estimating 3d rigid body transformations: a comparison of four major algorithms, Mach. Vis. and Appl., vol. 9, pp , 997. [2] B. Matei and P. Meer, Optimal rigid motion estimation and performance evaluation with bootstrap, in CVPR, 999. [3] D. Demirdjian and T. Darrell, Using multiple-hypothesis disparity maps and image velocity for 3-d motion estimation. IJCV, vol. 47, no. -3, pp , 22. [4] M. Agrawal, K. Konolige, and L. Iocchi, Real-time detection of independent motion using stereo, in Motion, 25. [5] M. Harville, A. Rahimi, T.Darrell, G. Gordon, and J. Woodfill, 3d pose tracking with linear depth and brightness constraints, in ICCV, 999. [6] R. Mandelbaum, G. Salgian, and H. Sawhney, Correlation-based estimation of ego-motion and structure from motion and stereo, in ICCV, 999. [7] D. Demirdjian and T. Darrell, Motion estimation from disparity images, in IEEE International Conference on Computer Vision, 2, pp. I: [8] K. Levenberg, A method for the solution of certain problems in least squares, in Quarterly of Applied Mathematics, vol. 2, 944. [9] D. Marquardt, An algorithm for least-squares estimation of nonlinear parameters, SIAM Journal on Applied Mathematics, vol., pp , 963. [2] R. Hartley, In defense of the eight-point algorithm, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 9, no. 6, pp , June 997.

6 3 Mean Translation Error 35 STD of Translation Error N N Mean Rotation Error 8 STD of Rotation Error N N Fig. 2. Comparison of N and with D-motion based methods under different noise levels undergoing large rotation [2] B. Matei and P. Meer, A general method for errors-in-variables problems in computer vision, in CVPR, 2, pp. II: [22] M. Mühlich and R. Mester, The role of total least squares in motion analysis, in European Conference on Computer Vision, 998. [23] D. Demirdjian and T. Darrell, Using multiple-hypothesis disparity maps and image velocity for 3-d motion estimation, Internation Journal of Computer Vision, vol. 47, no. -3, pp , April 22. [24] K. Arun, T. Huang, and S. Blostein, Least-squares fitting of two 3- D point sets, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 9, no. 5, pp , September 987.

Real-time detection of independent motion using stereo

Real-time detection of independent motion using stereo Real-time detection of independent motion using stereo Motilal Agrawal, Kurt Konolige and Luca Iocchi SRI International 333 Ravenswood Ave. Menlo Park, CA 94025 agrawal, konolige, iocchi@ai.sri.com Abstract

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Motion Estimation from Disparity Images

Motion Estimation from Disparity Images Motion Estimation from Disparity Images D. Demirdjian T. Darrell Massachusetts Institute of Technology, I Laboratory ambridge, M 239 demirdjiai.mit.edu trevorai.mit.edu bstract new method for 3D rigid

More information

A Factorization Method for Structure from Planar Motion

A Factorization Method for Structure from Planar Motion A Factorization Method for Structure from Planar Motion Jian Li and Rama Chellappa Center for Automation Research (CfAR) and Department of Electrical and Computer Engineering University of Maryland, College

More information

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne Hartley - Zisserman reading club Part I: Hartley and Zisserman Appendix 6: Iterative estimation methods Part II: Zhengyou Zhang: A Flexible New Technique for Camera Calibration Presented by Daniel Fontijne

More information

Parameter estimation. Christiano Gava Gabriele Bleser

Parameter estimation. Christiano Gava Gabriele Bleser Parameter estimation Christiano Gava Christiano.Gava@dfki.de Gabriele Bleser gabriele.bleser@dfki.de Introduction Previous lectures: P-matrix 2D projective transformations Estimation (direct linear transform)

More information

Multiple-View Structure and Motion From Line Correspondences

Multiple-View Structure and Motion From Line Correspondences ICCV 03 IN PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, NICE, FRANCE, OCTOBER 003. Multiple-View Structure and Motion From Line Correspondences Adrien Bartoli Peter Sturm INRIA

More information

Egomotion Estimation by Point-Cloud Back-Mapping

Egomotion Estimation by Point-Cloud Back-Mapping Egomotion Estimation by Point-Cloud Back-Mapping Haokun Geng, Radu Nicolescu, and Reinhard Klette Department of Computer Science, University of Auckland, New Zealand hgen001@aucklanduni.ac.nz Abstract.

More information

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches Reminder: Lecture 20: The Eight-Point Algorithm F = -0.00310695-0.0025646 2.96584-0.028094-0.00771621 56.3813 13.1905-29.2007-9999.79 Readings T&V 7.3 and 7.4 Essential/Fundamental Matrix E/F Matrix Summary

More information

Factorization with Missing and Noisy Data

Factorization with Missing and Noisy Data Factorization with Missing and Noisy Data Carme Julià, Angel Sappa, Felipe Lumbreras, Joan Serrat, and Antonio López Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona,

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253 Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine

More information

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10 Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263 Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine

More information

Real Time Stereo Image Registration for Planar Structure and 3D Sensor Pose Estimation

Real Time Stereo Image Registration for Planar Structure and 3D Sensor Pose Estimation 18 Real Time Stereo Image Registration for Planar Structure and 3D Sensor Pose Estimation Fadi Dornaika 1 and Angel D. Sappa 2 1 Institut Géographique National, 2 avenue Pasteur, 94165 Saint-Mandé, 2 Computer

More information

A Summary of Projective Geometry

A Summary of Projective Geometry A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]

More information

CS 664 Structure and Motion. Daniel Huttenlocher

CS 664 Structure and Motion. Daniel Huttenlocher CS 664 Structure and Motion Daniel Huttenlocher Determining 3D Structure Consider set of 3D points X j seen by set of cameras with projection matrices P i Given only image coordinates x ij of each point

More information

An idea which can be used once is a trick. If it can be used more than once it becomes a method

An idea which can be used once is a trick. If it can be used more than once it becomes a method An idea which can be used once is a trick. If it can be used more than once it becomes a method - George Polya and Gabor Szego University of Texas at Arlington Rigid Body Transformations & Generalized

More information

Midterm Exam Solutions

Midterm Exam Solutions Midterm Exam Solutions Computer Vision (J. Košecká) October 27, 2009 HONOR SYSTEM: This examination is strictly individual. You are not allowed to talk, discuss, exchange solutions, etc., with other fellow

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Euclidean Reconstruction Independent on Camera Intrinsic Parameters

Euclidean Reconstruction Independent on Camera Intrinsic Parameters Euclidean Reconstruction Independent on Camera Intrinsic Parameters Ezio MALIS I.N.R.I.A. Sophia-Antipolis, FRANCE Adrien BARTOLI INRIA Rhone-Alpes, FRANCE Abstract bundle adjustment techniques for Euclidean

More information

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Epipolar Geometry and Stereo Vision

Epipolar Geometry and Stereo Vision Epipolar Geometry and Stereo Vision Computer Vision Jia-Bin Huang, Virginia Tech Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X x

More information

Combining Two-view Constraints For Motion Estimation

Combining Two-view Constraints For Motion Estimation ombining Two-view onstraints For Motion Estimation Venu Madhav Govindu Somewhere in India venu@narmada.org Abstract In this paper we describe two methods for estimating the motion parameters of an image

More information

Multiple View Geometry in Computer Vision

Multiple View Geometry in Computer Vision Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Structure Computation Lecture 18 March 22, 2005 2 3D Reconstruction The goal of 3D reconstruction

More information

Vision par ordinateur

Vision par ordinateur Epipolar geometry π Vision par ordinateur Underlying structure in set of matches for rigid scenes l T 1 l 2 C1 m1 l1 e1 M L2 L1 e2 Géométrie épipolaire Fundamental matrix (x rank 2 matrix) m2 C2 l2 Frédéric

More information

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry 55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz

Stereo II CSE 576. Ali Farhadi. Several slides from Larry Zitnick and Steve Seitz Stereo II CSE 576 Ali Farhadi Several slides from Larry Zitnick and Steve Seitz Camera parameters A camera is described by several parameters Translation T of the optical center from the origin of world

More information

Two-view geometry Computer Vision Spring 2018, Lecture 10

Two-view geometry Computer Vision Spring 2018, Lecture 10 Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of

More information

On Board 6D Visual Sensors for Intersection Driving Assistance Systems

On Board 6D Visual Sensors for Intersection Driving Assistance Systems On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro

More information

Iterative Closest Point Algorithm in the Presence of Anisotropic Noise

Iterative Closest Point Algorithm in the Presence of Anisotropic Noise Iterative Closest Point Algorithm in the Presence of Anisotropic Noise L. Maier-Hein, T. R. dos Santos, A. M. Franz, H.-P. Meinzer German Cancer Research Center, Div. of Medical and Biological Informatics

More information

CS 395T Lecture 12: Feature Matching and Bundle Adjustment. Qixing Huang October 10 st 2018

CS 395T Lecture 12: Feature Matching and Bundle Adjustment. Qixing Huang October 10 st 2018 CS 395T Lecture 12: Feature Matching and Bundle Adjustment Qixing Huang October 10 st 2018 Lecture Overview Dense Feature Correspondences Bundle Adjustment in Structure-from-Motion Image Matching Algorithm

More information

Automatic 3D Model Construction for Turn-Table Sequences - a simplification

Automatic 3D Model Construction for Turn-Table Sequences - a simplification Automatic 3D Model Construction for Turn-Table Sequences - a simplification Fredrik Larsson larsson@isy.liu.se April, Background This report introduces some simplifications to the method by Fitzgibbon

More information

Lecture 3: Camera Calibration, DLT, SVD

Lecture 3: Camera Calibration, DLT, SVD Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P

More information

3D Motion from Image Derivatives Using the Least Trimmed Square Regression

3D Motion from Image Derivatives Using the Least Trimmed Square Regression 3D Motion from Image Derivatives Using the Least Trimmed Square Regression Fadi Dornaika and Angel D. Sappa Computer Vision Center Edifici O, Campus UAB 08193 Bellaterra, Barcelona, Spain {dornaika, sappa}@cvc.uab.es

More information

Two-View Geometry (Course 23, Lecture D)

Two-View Geometry (Course 23, Lecture D) Two-View Geometry (Course 23, Lecture D) Jana Kosecka Department of Computer Science George Mason University http://www.cs.gmu.edu/~kosecka General Formulation Given two views of the scene recover the

More information

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS Cha Zhang and Zhengyou Zhang Communication and Collaboration Systems Group, Microsoft Research {chazhang, zhang}@microsoft.com ABSTRACT

More information

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE

MOTION. Feature Matching/Tracking. Control Signal Generation REFERENCE IMAGE Head-Eye Coordination: A Closed-Form Solution M. Xie School of Mechanical & Production Engineering Nanyang Technological University, Singapore 639798 Email: mmxie@ntuix.ntu.ac.sg ABSTRACT In this paper,

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Extrinsic camera calibration method and its performance evaluation Jacek Komorowski 1 and Przemyslaw Rokita 2 arxiv:1809.11073v1 [cs.cv] 28 Sep 2018 1 Maria Curie Sklodowska University Lublin, Poland jacek.komorowski@gmail.com

More information

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology

More information

Last lecture. Passive Stereo Spacetime Stereo

Last lecture. Passive Stereo Spacetime Stereo Last lecture Passive Stereo Spacetime Stereo Today Structure from Motion: Given pixel correspondences, how to compute 3D structure and camera motion? Slides stolen from Prof Yungyu Chuang Epipolar geometry

More information

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW Thorsten Thormählen, Hellward Broszio, Ingolf Wassermann thormae@tnt.uni-hannover.de University of Hannover, Information Technology Laboratory,

More information

EigenTracking: Robust Matching and Tracking of Articulated Objects Using a View-Based Representation

EigenTracking: Robust Matching and Tracking of Articulated Objects Using a View-Based Representation EigenTracking: Robust Matching and Tracking of Articulated Objects Using a View-Based Representation Michael J. Black and Allan D. Jepson Xerox Palo Alto Research Center, 3333 Coyote Hill Road, Palo Alto,

More information

Image correspondences and structure from motion

Image correspondences and structure from motion Image correspondences and structure from motion http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2017, Lecture 20 Course announcements Homework 5 posted.

More information

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah

Srikumar Ramalingam. Review. 3D Reconstruction. Pose Estimation Revisited. School of Computing University of Utah School of Computing University of Utah Presentation Outline 1 2 3 Forward Projection (Reminder) u v 1 KR ( I t ) X m Y m Z m 1 Backward Projection (Reminder) Q K 1 q Q K 1 u v 1 What is pose estimation?

More information

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Heiko Hirschmüller, Peter R. Innocent and Jon M. Garibaldi Centre for Computational Intelligence, De Montfort

More information

CS231A Course Notes 4: Stereo Systems and Structure from Motion

CS231A Course Notes 4: Stereo Systems and Structure from Motion CS231A Course Notes 4: Stereo Systems and Structure from Motion Kenji Hata and Silvio Savarese 1 Introduction In the previous notes, we covered how adding additional viewpoints of a scene can greatly enhance

More information

CS201 Computer Vision Camera Geometry

CS201 Computer Vision Camera Geometry CS201 Computer Vision Camera Geometry John Magee 25 November, 2014 Slides Courtesy of: Diane H. Theriault (deht@bu.edu) Question of the Day: How can we represent the relationships between cameras and the

More information

Nonlinear Mean Shift for Robust Pose Estimation

Nonlinear Mean Shift for Robust Pose Estimation Nonlinear Mean Shift for Robust Pose Estimation Raghav Subbarao Yakup Genc Peter Meer ECE Department Real-time Vision and Modeling Department Rutgers University Siemens Corporate Research Piscataway, NJ

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

Lecture'9'&'10:'' Stereo'Vision'

Lecture'9'&'10:'' Stereo'Vision' Lecture'9'&'10:'' Stereo'Vision' Dr.'Juan'Carlos'Niebles' Stanford'AI'Lab' ' Professor'FeiAFei'Li' Stanford'Vision'Lab' 1' Dimensionality'ReducIon'Machine'(3D'to'2D)' 3D world 2D image Point of observation

More information

Plane-based Calibration Algorithm for Multi-camera Systems via Factorization of Homography Matrices

Plane-based Calibration Algorithm for Multi-camera Systems via Factorization of Homography Matrices Plane-based Calibration Algorithm for Multi-camera Systems via Factorization of Homography Matrices Toshio Ueshiba Fumiaki Tomita National Institute of Advanced Industrial Science and Technology (AIST)

More information

Robust Camera Calibration from Images and Rotation Data

Robust Camera Calibration from Images and Rotation Data Robust Camera Calibration from Images and Rotation Data Jan-Michael Frahm and Reinhard Koch Institute of Computer Science and Applied Mathematics Christian Albrechts University Kiel Herman-Rodewald-Str.

More information

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo -- Computational Optical Imaging - Optique Numerique -- Multiple View Geometry and Stereo -- Winter 2013 Ivo Ihrke with slides by Thorsten Thormaehlen Feature Detection and Matching Wide-Baseline-Matching

More information

Identifying Car Model from Photographs

Identifying Car Model from Photographs Identifying Car Model from Photographs Fine grained Classification using 3D Reconstruction and 3D Shape Registration Xinheng Li davidxli@stanford.edu Abstract Fine grained classification from photographs

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching --

Computational Optical Imaging - Optique Numerique. -- Single and Multiple View Geometry, Stereo matching -- Computational Optical Imaging - Optique Numerique -- Single and Multiple View Geometry, Stereo matching -- Autumn 2015 Ivo Ihrke with slides by Thorsten Thormaehlen Reminder: Feature Detection and Matching

More information

Epipolar Geometry and Stereo Vision

Epipolar Geometry and Stereo Vision Epipolar Geometry and Stereo Vision Computer Vision Shiv Ram Dubey, IIIT Sri City Many slides from S. Seitz and D. Hoiem Last class: Image Stitching Two images with rotation/zoom but no translation. X

More information

Unsupervised learning in Vision

Unsupervised learning in Vision Chapter 7 Unsupervised learning in Vision The fields of Computer Vision and Machine Learning complement each other in a very natural way: the aim of the former is to extract useful information from visual

More information

arxiv: v1 [cs.cv] 18 Sep 2017

arxiv: v1 [cs.cv] 18 Sep 2017 Direct Pose Estimation with a Monocular Camera Darius Burschka and Elmar Mair arxiv:1709.05815v1 [cs.cv] 18 Sep 2017 Department of Informatics Technische Universität München, Germany {burschka elmar.mair}@mytum.de

More information

Spatio-Temporal Stereo Disparity Integration

Spatio-Temporal Stereo Disparity Integration Spatio-Temporal Stereo Disparity Integration Sandino Morales and Reinhard Klette The.enpeda.. Project, The University of Auckland Tamaki Innovation Campus, Auckland, New Zealand pmor085@aucklanduni.ac.nz

More information

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006 Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics

More information

Miniature faking. In close-up photo, the depth of field is limited.

Miniature faking. In close-up photo, the depth of field is limited. Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Jae-Hak Kim 1, Richard Hartley 1, Jan-Michael Frahm 2 and Marc Pollefeys 2 1 Research School of Information Sciences and Engineering

More information

Absolute Scale Structure from Motion Using a Refractive Plate

Absolute Scale Structure from Motion Using a Refractive Plate Absolute Scale Structure from Motion Using a Refractive Plate Akira Shibata, Hiromitsu Fujii, Atsushi Yamashita and Hajime Asama Abstract Three-dimensional (3D) measurement methods are becoming more and

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

1 Projective Geometry

1 Projective Geometry CIS8, Machine Perception Review Problem - SPRING 26 Instructions. All coordinate systems are right handed. Projective Geometry Figure : Facade rectification. I took an image of a rectangular object, and

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

arxiv: v1 [cs.cv] 2 May 2016

arxiv: v1 [cs.cv] 2 May 2016 16-811 Math Fundamentals for Robotics Comparison of Optimization Methods in Optical Flow Estimation Final Report, Fall 2015 arxiv:1605.00572v1 [cs.cv] 2 May 2016 Contents Noranart Vesdapunt Master of Computer

More information

Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy

Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy Depth Measurement and 3-D Reconstruction of Multilayered Surfaces by Binocular Stereo Vision with Parallel Axis Symmetry Using Fuzzy Sharjeel Anwar, Dr. Shoaib, Taosif Iqbal, Mohammad Saqib Mansoor, Zubair

More information

EECS 442: Final Project

EECS 442: Final Project EECS 442: Final Project Structure From Motion Kevin Choi Robotics Ismail El Houcheimi Robotics Yih-Jye Jeffrey Hsu Robotics Abstract In this paper, we summarize the method, and results of our projective

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

On Road Vehicle Detection using Shadows

On Road Vehicle Detection using Shadows On Road Vehicle Detection using Shadows Gilad Buchman Grasp Lab, Department of Computer and Information Science School of Engineering University of Pennsylvania, Philadelphia, PA buchmag@seas.upenn.edu

More information

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert Week 2: Two-View Geometry Padua Summer 08 Frank Dellaert Mosaicking Outline 2D Transformation Hierarchy RANSAC Triangulation of 3D Points Cameras Triangulation via SVD Automatic Correspondence Essential

More information

Computer Vision CSCI-GA Assignment 1.

Computer Vision CSCI-GA Assignment 1. Computer Vision CSCI-GA.2272-001 Assignment 1. September 22, 2017 Introduction This assignment explores various methods for aligning images and feature extraction. There are four parts to the assignment:

More information

Epipolar Geometry and the Essential Matrix

Epipolar Geometry and the Essential Matrix Epipolar Geometry and the Essential Matrix Carlo Tomasi The epipolar geometry of a pair of cameras expresses the fundamental relationship between any two corresponding points in the two image planes, and

More information

Camera Geometry II. COS 429 Princeton University

Camera Geometry II. COS 429 Princeton University Camera Geometry II COS 429 Princeton University Outline Projective geometry Vanishing points Application: camera calibration Application: single-view metrology Epipolar geometry Application: stereo correspondence

More information

Calibration of a Multi-Camera Rig From Non-Overlapping Views

Calibration of a Multi-Camera Rig From Non-Overlapping Views Calibration of a Multi-Camera Rig From Non-Overlapping Views Sandro Esquivel, Felix Woelk, and Reinhard Koch Christian-Albrechts-University, 48 Kiel, Germany Abstract. A simple, stable and generic approach

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy 1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:

More information

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1

More information

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

Multiple View Geometry of Projector-Camera Systems from Virtual Mutual Projection

Multiple View Geometry of Projector-Camera Systems from Virtual Mutual Projection Multiple View Geometry of rojector-camera Systems from Virtual Mutual rojection Shuhei Kobayashi, Fumihiko Sakaue, and Jun Sato Department of Computer Science and Engineering Nagoya Institute of Technology

More information

MAPI Computer Vision. Multiple View Geometry

MAPI Computer Vision. Multiple View Geometry MAPI Computer Vision Multiple View Geometry Geometry o Multiple Views 2- and 3- view geometry p p Kpˆ [ K R t]p Geometry o Multiple Views 2- and 3- view geometry Epipolar Geometry The epipolar geometry

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Epipolar Geometry Based On Line Similarity

Epipolar Geometry Based On Line Similarity 2016 23rd International Conference on Pattern Recognition (ICPR) Cancún Center, Cancún, México, December 4-8, 2016 Epipolar Geometry Based On Line Similarity Gil Ben-Artzi Tavi Halperin Michael Werman

More information

Multiple Motion Scene Reconstruction from Uncalibrated Views

Multiple Motion Scene Reconstruction from Uncalibrated Views Multiple Motion Scene Reconstruction from Uncalibrated Views Mei Han C & C Research Laboratories NEC USA, Inc. meihan@ccrl.sj.nec.com Takeo Kanade Robotics Institute Carnegie Mellon University tk@cs.cmu.edu

More information

Quasi-Euclidean Uncalibrated Epipolar Rectification

Quasi-Euclidean Uncalibrated Epipolar Rectification Dipartimento di Informatica Università degli Studi di Verona Rapporto di ricerca Research report September 2006 RR 43/2006 Quasi-Euclidean Uncalibrated Epipolar Rectification L. Irsara A. Fusiello Questo

More information

3D Visualization through Planar Pattern Based Augmented Reality

3D Visualization through Planar Pattern Based Augmented Reality NATIONAL TECHNICAL UNIVERSITY OF ATHENS SCHOOL OF RURAL AND SURVEYING ENGINEERS DEPARTMENT OF TOPOGRAPHY LABORATORY OF PHOTOGRAMMETRY 3D Visualization through Planar Pattern Based Augmented Reality Dr.

More information