Projective reconstruction from line-correspondences in multiple uncalibrated images

Size: px
Start display at page:

Download "Projective reconstruction from line-correspondences in multiple uncalibrated images"

Transcription

1 Pattern Recognition 39 (2006) Projective reconstruction from line-correspondences in multiple uncalibrated images A.W.K. Tang, T.P. Ng, Y.S. Hung, C.H. Leung Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam Road, Hong Kong Received 30 June 2005 Abstract A new approach is proposed for reconstructing 3D lines and cameras from 2D corresponding lines across multiple uncalibrated views. There is no requirement that the 2D corresponding lines on different images represent the same segment of a 3D line, which may not appear on all images. A 3D line is reconstructed by minimizing a geometric cost function that measures the distance of the reprojected end points of the 3D segment from the measured 2D lines on different images. An algorithmic procedure is provided with guaranteed convergence to a solution where the geometric cost function achieves a (local) minimum Pattern Recognition Society. Published by Elsevier Ltd. All rights reserved. Keywords: Multiple views; Projective reconstruction; Line reconstruction; Line correspondence 1. Introduction The recovery of a 3D scene from uncalibrated images has long been a major area in computer vision. Existing reconstruction methods mainly make use of feature point correspondences [1 3] across multiple views. Recently, 3D reconstruction from line correspondences [4 12] has attracted some attention since there are many merits of reconstruction from lines over points. In the matching process, for example, reconstruction from lines allows correspondence of different sections of a line in different views. Dornaika and Garcia [13] proposed two 3D reconstruction methods from point and line correspondences for weak perspective and paraperspective camera models. Ansar and Daniilidis [14] proposed a linear method for 3D reconstruction from point and line correspondences under the assumption that the camera intrinsic parameters are known. Both of these methods (i.e. Refs. [13] and [14]), however, make a priori assumptions about the cameras. Corresponding author. Tel.: ; fax: addresses: wktang@eee.hku.hk (A.W.K. Tang), NgTuenPui@graduate.hku.hk (T.P. Ng), yshung@eee.hku.hk (Y.S. Hung), chleung@eee.hku.hk (C.H. Leung). For uncalibrated cameras, Hartley [4] introduced the geometry between line correspondences across three images which was later developed into the trifocal tensor [6] governing point and line correspondences in three images analogous to that of the fundamental matrix for two images. Explicit linear algorithms for computing the trifocal tensor from line correspondences are available and camera projection matrices can be readily obtained from the trifocal tensor. Methods based on these multilinear constraints [4 6], however, are noise sensitive and applicable only for three or four views, and they minimize an algebraic rather than geometric error. A different approach for line reconstruction is to parameterize a 3D line as a 6-vector in Plücker coordinates. Faugeras and Mourrain [15] propose to transform a general projection matrix for points to a line projection matrix in Plücker coordinates, so that the linear relationship of projecting a 3D line onto an image plane as a 2D line can be represented by the line projection matrix similar to the projection equation for points. There are several methods [1,8 10,15] based on the linear projection equation for lines in Plücker coordinates. Maolin et al. [8] propose an iterative projective reconstruction for estimating scale factors, and Martinec and Pajdla [9] propose a non-iterative /$ Pattern Recognition Society. Published by Elsevier Ltd. All rights reserved. doi: /j.patcog

2 890 A.W.K. Tang et al. / Pattern Recognition 39 (2006) factorization method. The main problem of the 3D line reconstruction in the Plücker coordinates is that both the line and line projection matrix should satisfy some conditions which cannot be enforced directly during the reconstruction process. Instead, they are enforced at the end of the reconstruction process with the introduction of unpredictable errors due to an algebraic condition with no physical meaning. Moreover, the reconstructed line projection matrices need to be transformed back to general projection matrices for upgrading the projective to Euclidean space. Bartoli and Sturm [16] show empirically that the existing linear methods perform poorly in most cases, and they propose a bundle adjustment method based on orthonormal representation of 3D lines in the Plücker coordinates for minimizing a geometric quantity, namely the orthogonal distances between the reprojected lines and the two 2D end points of the measured line segments. Soe and Hong [10] propose a sequential line reconstruction method using iterative extended Kalman filter (IEKF) by adding short baseline frames one by one to update the scene reconstructed from three wide-baseline frames. As this method relies on the initial estimates of the projection matrices computed from the first three views and they are updated iteratively, the method is biased towards the images input earlier. In Ref. [12], Triggs uses the factorization-based method to perform 3D reconstruction from both point and line correspondences. However, the method needs a set of point correspondences to compute fundamental matrices which are then used to convert the line correspondences into additional point correspondences. Furthermore, the algebraic rather than geometric error is minimized, and no missing data is allowed. In this paper, we will propose a line reconstruction method adopted from Ref. [1], which performs projective reconstruction from point correspondences in multiple views by minimizing 2D reprojection errors using a quasi-linear bundle adjustment approach. The proposed method reconstructs a line in 3D space as a segment with two suitably chosen end points whose projections on images are as close as possible in a geometric sense to measured lines that correspond with each other. Despite the apparent simplicity of the idea, there are many intricacies in the design of a complete solution using this approach, and we are not aware of any similar method for line reconstruction having the following characteristics: the cost function represents a geometric measure of the goodness of reconstruction; the reconstruction is truly based on line correspondences with no need for point correspondences; the method is able to handle missing lines; the optimization algorithm incorporates bundle adjustment and is guaranteed to converge to a local minimum of the geometric cost function. The paper is organized as follows. In Section 2, we will give a brief summary of the point-based quasi-linear bundle adjustment method of Ref. [1] that will be adopted for the proposed line reconstruction method. In Section 3, the line reconstruction problem is formulated as a minimization of a geometric cost and an algorithmic solution is developed. Experimental results are given in Section 4. Section 5 contains some concluding remarks. Notation: The Hadamard product of two matrices A=[a ij ] and B =[b ij ] of the same size is denoted A B =[a ij b ij ]. 2. Point-based method for projective reconstruction Suppose a set of n image point correspondences x ij are established across m views. The jth 3D point X j R 4 1 projected on the ith view by a projection matrix P i R 3 4 as a 2D point x ij =[u ij v ij 1] T can be expressed as λ ij x ij = P i X j, (1) where λ ij is the depth of X j to the ith camera. The projection of all the 3D points onto all the cameras can be represented in matrix form as λ 11 x 11 λ 1n x 1n M = = PX R 3m n, (2) λ m1 x m1 λ mn x mn where M is the scaled measurement matrix, P =[P T 1 P T 2 P T m ]T R 3m 4 is the joint projection matrix and X = [X 1 X 2 X n ] R 4 n is the shape matrix. If the projective depths [λ ij ] are known, the scaled measurement matrix can be factorized into P and X by means of singular value decomposition. Various methods are proposed to determine the projective depths [1,2]. In particular, the depths are estimated iteratively in Ref. [1] as a minimization problem with the cost function i=m,j=n min γ ij (x ij β ij P i X j ) 2, (3) P,X,β i=1,j=1 where β ij (=1/λ ij ) is the inverse depth, γ ij =[11γγ ij ] T, γ is a control factor to force the cost (3) to approach the 2D reprojection error and γ ij are weighting factors for balancing the magnitude of pixel coordinates. This method has the advantages that it minimizes the 2D reprojection error and is capable of dealing with missing data by considering only the visible 2D points in (3). An iterative algorithm can be used to estimate the unknown parameters β ij, P i and X j alternately as linear least-squares problems and therefore the solution process is a quasi-linear bundle adjustment. In the next section, we will extend this formulation to the case of line reconstruction.

3 A.W.K. Tang et al. / Pattern Recognition 39 (2006) Line reconstruction Estimated 3D Points 3.1. Representing lines In Fig. 1, a 2D line is represented as a 3-vector l. The minimum distance (i.e. orthogonal distance) between a 2D point x and the 2D line l is given by d (x, l) = l T x, (4) (l) (l)2 2 where (l) k is the kth element of the vector l. Ifx lies on l, d (x, l) = 0. A line can also be represented by any two distinct points (e.g. x 1 and x 2 for 2D cases or X 1 and X 2 for 3D cases) which lie on it. The points lying on that line between these two points can be expressed as αx 1 + (1 α)x 2 for a 2D line or αx 1 + (1 α)x 2 for a 3D line, 0 α 1. The proposed method relies on these two kinds of line representations The error measure for line reconstruction A critical issue of line reconstruction methods is how to measure the difference between a measured line and its reprojected line from the reconstructed 3D line. Let the jth measured 2D line segment on the ith view be denoted l ij and its two end points be denoted x ij and x i,n+j so that l T ij x ij =l T ij x i,n+j =0, where n is the total number of 3D lines. Let the measured 2D line segments be indexed by ordered pairs of the set A ={(i, j) l ij with end points x ij and x i,n+j is observed as the jth line on the ith view}. The idea of the proposed method is to reconstruct a 3D line segment with end points X j and X n+j whose reprojected end points on the ith view will be close to l ij. Denote the reprojected end points of X j and X n+j on the ith view by ˆx ij and ˆx i,n+j, respectively. Making use of (4), the measure of error in the line reconstruction can be defined as ε 2 = 1 2 n(a) (i,j) A (lij T ˆx i,j ) 2 + (lij T ˆx i,n+j ) 2 (l ij ) (l ij ) 2, (5) 2 where n(a) is the number of elements of the set A. ε can be regarded as the root mean square error (RMSE) of the O l (l) (l) 1 + (l) 2 d Fig. 1. Line on a 2D plane. x Estimated Camera Centre Measured line with two end points Reprojected Points Fig. 2. The projection of a line. reprojected lines to the measured line segments. It is also possible to simplify (5) by normalizing l ij so that (l ij ) (l ij ) 2 2 = 1, i, j. In the sequel, we will assume that all lines are already normalized Analyses on using lines for projective reconstruction In Fig. 2, the reprojection of the estimated end points of a 3D line segment on an image is illustrated. Formulating a minimization problem directly from (5) to estimate the 3D line segments without extra constraint(s) will not provide a reasonable solution. This is because it is always easier to fit a single point than two end points of a line segment to the projection model, and therefore there is a tendency during optimization for the two end points of a line segment to converge to one single point that is closest to a measured line. In order to prevent the two end points from converging into one point, we require the two reprojected end points of the 3D line segment to approach the end points of a particular measured line segment on one of the image planes, and this line will be referred to as the reference line segment. Thus, the reference line segment acts as an anchor for fixing the section of the 3D line to be reconstructed. For each 3D line, only one line segment is chosen as the reference line segment. For all other lines l ij, the distance of the reprojected end points of the 3D line segment X j X n+j from l ij is taken to be the orthogonal distance from the (extended) line l ij without any regard to the actual positions of the end points of the measured line. Fig. 3 shows the idea of reconstructing the 3D line segment by requiring its reprojected end points to approach the measured end points of a reference line segment (leftmost image), whereas the reprojected end points on the other two views are only required to be as close as possible to the measured lines with no other restrictions on the locations of the reprojected points Choice of reference line segments The strategy of reference line segment selection directly affects the accuracy of the final reconstruction. The

4 892 A.W.K. Tang et al. / Pattern Recognition 39 (2006) ˆx i,n+j xˆ ij l rj x i,n+j θ n+j θ j x ij Fig. 4. The relationship between 2D reprojection error and orthogonal distance. is related to the orthogonal distance to the line as: Fig. 3. The reconstruction process. recommended strategy is to choose the line segment having the maximum length in pixel coordinate among the group of corresponding lines across multiple views. If the same amount of disturbance is applied to the end points of line segments having different lengths, the disturbed line will be closer to the original line when the ratio of disturbance to the length of line segment is smaller. Let the reference line segments be indexed by ordered pairs of the set R ={(r, j) l rj is chosen as a reference line segment} Cost function for line reconstruction To formulate the line reconstruction method stated above, the cost function of the proposed method is defined as F L (P, X, β) = 1 1 γ 2n(A) r,kn+j (x r,kn+j k=0 (r,j) R β r,kn+j P r X kn+j ) 2 + (β i,kn+j lij T P ix kn+j ) 2 (i,j) A\R + γ 2 γ 2 i,kn+j (1 β i,kn+j Pi 3 X kn+j ) 2, (6) where P s i is the sth row of P i and γ i,kn+j =[11γγ i,kn+j ] T. The first summation in Eq. (6) is the mean squares of the reprojection errors for all the end points of the reference line segments. This quantity becomes the 2D reprojection error when γ tends to infinity. The second summation becomes the mean squares of all the orthogonal distances of the reprojected points to the measured lines when γ tends to infinity. The cost (6) is also the upper bound error of Eq. (5). In Fig. 4, the 2D reprojection error of the reprojected points ˆx i,j and ˆx i,n+j from the end points of the reference line l ij sum of squared orthogonal distances = ˆx r,j x r,j 2 sin 2 θ j + ˆx r,n+j x r,n+j 2 sin 2 θ n+j, θ j, θ n+j R ˆx r,j x r,j 2 + ˆx r,n+j x r,n+j 2 sum of squared 2D reprojection errors. The cost function (6) can therefore be treated as an approximation to the error measure (5) with a constraint on the end points of the reconstructed line Choice of weighting factors The purpose of the weighting factor γ ij is to balance the magnitude of the pixel coordinates with the last element (i.e. 1) of the homogeneous coordinates. During the iterations of the proposed method, γ ij is fixed. There are different considerations for the two terms in Eq. (6). For the first term of the minimization of 2D reprojection error, we choose γ rj = max( u rj, v rj ), (r, j) R. (7) This weighting factor helps to balance the distribution among the three components of the 2D homogeneous coordinates. For the second term of the minimization of orthogonal distances, γ ij is chosen as (l ij ) 3 (l γ ij = rj ) 3 max( u rj, v rj, u r,n+j, v r,n+j ), (l rj ) 3 = 0 (l ij ) 3, (l rj ) 3 = 0 (i, j) A\R,(r,j) R and i = r, (8) where the lines l ij and l rj are normalized and the 2D line l rj on the rth view is chosen as the reference line segment for the jth 3D line. Note that (l ij ) 3 is the orthogonal distance between the 2D line from the image origin. Similar to (7), the weighting factor (8) is meant to scale the third component of the homogeneous coordinates to magnitudes comparable with the pixel coordinates of the reprojected points on the ith view. We note however that the final results are not unduly sensitive to the choice of γ ij, and as γ is increased

5 progressively in the algorithm (see step 7 of algorithm below), eventually γγ ij irrespective of the choice of γ ij A quasi-linear bundle adjustment algorithm for line reconstruction A.W.K. Tang et al. / Pattern Recognition 39 (2006) Put k = 0, γ = 1, ε b = 0; Set β 0 ij = 1, i, j and compute P 0 and X 0 from a rank-4 SVD approximation of [(1/β 0 ij )x ij ]. Select reference line segments and compute γ ij. 2. Put k = k Fix X k 1 and β k 1 and determine P k by solving min P kf L (P k,x k 1, β k 1 ). 4. Fix P k and β k 1 and determine X k by solving min X kf L (P k,x k, β k 1 ). 5. Fix P k and X k and determine β k by solving ε k = min β kf L (P k,x k, β k ). 6. Repeat steps 2 5 until ε k converges. 7. If ε k ε b is bigger than a threshold value, then put ε b =ε k, γ = 1.1 γ,k= 0 and return to step 2, else stop. At the end of the algorithm, the projection matrices are taken as Pi k and the 3D lines are defined by the points Xj k and Xn+j k. When some 3D lines are not visible to all the views, missing entries in the measurement matrix should be filled in the initialization stage. The end points of a missing line can be filled by the centroid of the end points of the measured line segments on the view. This method for initializing missing data is used in all examples in this paper. We found that this initialization works well in all cases when the percentage of missing line segments is less than 60% Convergence The proposed method minimizes the cost function, F L (P, X, β), by estimating P, X and β iteratively as three weighted least squares problems. It can be shown that as γ, F L (P, X, β) is guaranteed to converge to a (local) minimum of the cost function with a geometric meaning given in Section Experimental results The proposed method is evaluated by both synthetic and real data. The method is implemented using Matlab 6.5 running on a Pentium-42.8 GHz PC Synthetic data Fig. 5 shows a view of the synthetic scene consisting of a box and five cameras. The box of size 1m 1m 1m is made up of 29 3D lines with patterns of triangles and rectangles on its sides. Five cameras are randomly generated RMS 2D reprojection error / pixel Fig. 5. A view of the synthetic scene noise level / pixel Fig. 6. Performance of the algorithm with data of difference noise levels. such that all of them point at the centroid of the box and the image of the box almost occupies the whole image. The intrinsic parameters are fixed over all views and the image size is pixels. The two vertices of each 2D line segment are chosen as its two measured points and they are perturbed by Gaussian noise independently. Line segments in the first view are chosen as the reference line segments. The reconstruction results for noisy data having Gaussian noise level ranging from 0 to 4 pixels with increments of 0.5 pixels are shown in Fig. 6. The noises are added to the x- and y-coordinates of the corner points independently. For each noise level, the estimated reprojection error is taken as the mean of the root mean square values calculated in 30 trials with different randomly generated noise. Note from Fig. 6 that the RMS error is proportional to the added noise. We have also implemented the line factorization method [9] in Matlab. Although the method works fine for noisefree data, the results are however very sensitive to added noise and it is not always possible to obtain a reasonable solution even at low noise levels. Hence we cannot perform a sensible comparison with our method Real data We consider two real image sequences, the clock tower sequence and the toy house sequence. Feature lines of the objects are manually extracted and matched.

6 894 A.W.K. Tang et al. / Pattern Recognition 39 (2006) Fig. 7. Image sequence of the clock tower. Fig. 9. The reconstructed tower. Fig. 8. The reconstructed lines of the reconstructed tower Clock tower Four images were taken around the clock tower shown in Fig. 7 by a Canon D60 digital camera with a 70 mm lens. The image size is pixels and the intrinsic parameters are varied due to auto-focusing. Line segments on the images corresponding to 22 3D line segments of the tower are visible in all views but different portions of the same segment are seen in different views. The measured line segments are shown as blue solid lines superimposed on the four images in Fig. 7. Note that some parts of the feature lines are blocked by the palm trees in front of the tower in the second image from the left. The proposed method took s and 679 iterations to converge. The RMS 2D reprojection error of the reference line segments is pixels and the RMS orthogonal distance is pixels. Using the estimated projection matrices, the reconstructed 3D line segments can be reprojected onto the images and these are plotted as dotted green line segments on two of the images. Fig. 8 shows the reprojected line segments (dotted green segments) superimposed on the solid blue measured lines for comparison. Fig. 9 shows different views of the reconstructed tower in the Euclidean frame. Fig. 9(a) shows the estimated positions of the cameras taking the images. Fig. 9(b) shows a close-up view of the tower. Fig. 9(c) is the top view of the tower, which shows that the angle between two sides of the tower is almost Toy house Nine images were taken around a toy house by a Canon A80 digital camera, as shown in Fig. 10. The image size is Fig. 10. Image sequence of toy house. Fig. 11. Distribution of measured line segments for image sequence of toy house pixels and the intrinsic parameters are varied due to auto-focusing. As not all the parts of the house are visible in all images, this image sequence demonstrates the performance of the proposed method for dealing with missing line segments in some of the images. In these images, each line segment is visible in at least three images. Line segments on the images corresponding to 57 3D line segments are extracted manually and there are altogether 271 missing lines on the images (about 57.2% of all 2D lines). The manually extracted lines are shown as blue lines in Fig. 10. The distribution of measured line segments are shown in Fig. 11 where a blue cross in the (i, j) position indicates

7 A.W.K. Tang et al. / Pattern Recognition 39 (2006) Fig. 12. The reconstructed house in Euclidean frame. that the jth 3D line can be observed by the ith camera as a 2D line l ij. The proposed method took 466 s with 1187 iterations to converge. The RMS 2D reprojection error of the reference line segments is pixels and the RMS orthogonal distance is pixels. The results of the reconstruction are shown in Fig. 12, where different views of the reconstructed house after being upgraded to the Euclidean frame are presented. 5. Conclusion In this paper, a projective reconstruction method for line correspondences has been developed and evaluated by synthetic data and real image sequences. We reconstruct a 3D line segment by minimizing the sum of orthogonal distances of the reprojected end points of the 3D segment from the measured 2D lines in all the images, except for a reference line segment where the distance is measured from the actual end points of the measured segment. The problem of line reconstruction is thus formulated purely on the basis of line correspondences in multiple views. There is no requirement that the 2D corresponding lines visible on different images represent the same segment of a 3D line. The method does not require point correspondence and its ability to handle missing lines is demonstrated by real image sequences. Experimental results based on synthetic data show that the proposed method is robust to added noise and the RMS orthogonal distance is almost linearly proportional to the added noise. 6. Summary In this paper, the problem of using line correspondences for projective reconstruction from multiple uncalibrated images is addressed. We propose a new approach to reconstruct 3D lines from 2D corresponding lines across multiple views. The reconstruction problem is formulated purely on the basis of line correspondences in multiple views. There is no requirement that the 2D corresponding lines visible on different images represent the same segment of a 3D line, and the lines are not required to be visible in all the views. Most 3D reconstruction methods make use of feature point correspondences across multiple views. However, reconstruction from line correspondences has many merits over point-based methods. Existing methods for projective reconstruction from line correspondences are mainly based on multilinear geometric constraints or the linear projection equation for lines in Plücker coordinates. These methods have the drawback of being excessively noise sensitive and they minimize an algebraic rather than geometric error. A further problem of 3D line reconstruction in Plücker coordinates is that both the line and line projection matrix should satisfy some conditions which cannot be enforced directly during the reconstruction process. Instead, they are enforced at the end of the reconstruction process with the introduction of unpredictable errors due to the imposition of an algebraic condition with no physical meaning. In this paper, we propose a new approach for reconstructing 3D lines from 2D line correspondences. A 3D line segment is reconstructed by minimizing a geometric cost function that measures the distance of the reprojected end points of the 3D segment from the measured 2D lines on different images. The distance is defined to be the orthogonal distance of the reprojected end points of the 3D segment from the measured 2D lines in all the images, except for a reference line segment where the distance is measured from the actual end points of the measured segment. As the choice of reference line segments directly affects the accuracy of the final reconstruction, we propose a strategy to choose reference line segments among the groups of corresponding lines across multiple views. A quasi-linear bundle adjustment algorithm is provided which is guaranteed to converge to a solution where the geometric cost function achieves a (local) minimum. Experimental results based on synthetic data show that the proposed method is robust to added noise and the RMS error measure is almost linearly proportional to the added noise. Real image sequences demonstrate the performance of the proposed method for dealing with cases when different sections of the same segment are seen in different views and some lines are missing in some of the views. Acknowledgements The work described in this paper was substantially supported by a grant from the Research Grants Council of Hong Kong Special Administrative Region, China (Project No. HKU7058/02E).

8 896 A.W.K. Tang et al. / Pattern Recognition 39 (2006) References [1] W.K. Tang, Y.S. Hung, A factorization-based method for projective reconstruction with minimization of 2D reprojection errors, in: Proceedings of the DAGM 2002, September, 2002, pp [2] P. Sturm, B. Triggs, A factorization based algorithm for multipleimage projective structure and motion, in: Proceedings of the European Conference on Computer Vision, Cambridge, England, 1996, pp [3] C. Tomasi, T. Kanade, Shape and motion from image streams under orthography: a factorization method, Int. J. Comput. Vision 9 (2) (1992) [4] R. I. Hartley, Projective reconstruction from line correspondences, in: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1994, pp [5] R.I. Hartley, Multilinear relationships between coordinates of corresponding image points and lines, in: Proceedings of the Sophus Lie Symposium, Nordfjordeid, Norway, [6] R.I. Hartley, Lines and points in three views and the trifocal tensor, Int. J. Comput. Vision 22 (2) (1997) [7] Y. Liu, S. Huang, A linear algorithm for motion estimation using straight line correspondences, in: Proceedings of the IEEE Ninth International Conference on Pattern Recognition, vol. 1, 1988, pp [8] H. Maolin, Z. Quanbing, W. Sui, Projective reconstruction from lines based on SVD, in: Proceedings of the SPIE The International Society for Optical Engineering, vol. 4875, 2002, pp [9] D. Martinec, T. Pajdla, Line reconstruction from many perspective images by factorization, in: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, 2003, pp [10] Y. Seo, K. S. Hong, Sequential reconstruction of lines in projective space, in: Proceedings of the 13th International Conference in Pattern Recognition, vol. 1, 1996, pp [11] J. Weng, S. Huang, N. Ahuja, Motion and structure from line correspondences: closed-form solution, uniqueness and optimization, IEEE Trans. Pattern Anal. Mach. Intell. 14 (3) (1992) [12] B. Triggs, Factorization methods for projective structure and motion, in: Proceedings of the Conference on Computer Vision and Pattern Recognition, San Franciso, June, 1996, pp [13] F. Donaika, C. Garcia, Pose estimation using point and line correspondences, Real-Time Imaging 5 (1999) [14] A. Ansar, K. Daniilidis, Linear pose estimation from points or lines, IEEE Trans. Pattern Anal. Mach. Intell. 25 (5) (2003) [15] O. Faugeras, B. Mourrain, On the geometry and algebra of the point and line correspondences between N images, in: Proceedings of the Fifth International Conference on Computer Vision, 1995, pp [16] A. Bartoli, P. Sturm, Multiple-view structure and motion from line correspondences, in: Proceedings of the International Conference on Computer Vision, 2003.

Perception and Action using Multilinear Forms

Perception and Action using Multilinear Forms Perception and Action using Multilinear Forms Anders Heyden, Gunnar Sparr, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: {heyden,gunnar,kalle}@maths.lth.se Abstract

More information

Multiple-View Structure and Motion From Line Correspondences

Multiple-View Structure and Motion From Line Correspondences ICCV 03 IN PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, NICE, FRANCE, OCTOBER 003. Multiple-View Structure and Motion From Line Correspondences Adrien Bartoli Peter Sturm INRIA

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t R 2 3,t 3 Camera 1 Camera

More information

Structure from Motion CSC 767

Structure from Motion CSC 767 Structure from Motion CSC 767 Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R,t R 2,t 2 R 3,t 3 Camera??

More information

Structure from Motion

Structure from Motion 11/18/11 Structure from Motion Computer Vision CS 143, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

A Factorization Method for Structure from Planar Motion

A Factorization Method for Structure from Planar Motion A Factorization Method for Structure from Planar Motion Jian Li and Rama Chellappa Center for Automation Research (CfAR) and Department of Electrical and Computer Engineering University of Maryland, College

More information

Factorization with Missing and Noisy Data

Factorization with Missing and Noisy Data Factorization with Missing and Noisy Data Carme Julià, Angel Sappa, Felipe Lumbreras, Joan Serrat, and Antonio López Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona,

More information

A Summary of Projective Geometry

A Summary of Projective Geometry A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]

More information

Multiple View Geometry in Computer Vision

Multiple View Geometry in Computer Vision Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Structure Computation Lecture 18 March 22, 2005 2 3D Reconstruction The goal of 3D reconstruction

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Haowei Liu LECTURE 16 Structure from Motion from Tracked Points 16.1. Introduction In the last lecture we learned how to track point features

More information

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View

More information

Camera Calibration Using Line Correspondences

Camera Calibration Using Line Correspondences Camera Calibration Using Line Correspondences Richard I. Hartley G.E. CRD, Schenectady, NY, 12301. Ph: (518)-387-7333 Fax: (518)-387-6845 Email : hartley@crd.ge.com Abstract In this paper, a method of

More information

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy 1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:

More information

The end of affine cameras

The end of affine cameras The end of affine cameras Affine SFM revisited Epipolar geometry Two-view structure from motion Multi-view structure from motion Planches : http://www.di.ens.fr/~ponce/geomvis/lect3.pptx http://www.di.ens.fr/~ponce/geomvis/lect3.pdf

More information

Structure from Motion

Structure from Motion Structure from Motion Outline Bundle Adjustment Ambguities in Reconstruction Affine Factorization Extensions Structure from motion Recover both 3D scene geoemetry and camera positions SLAM: Simultaneous

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Multiple Motion Scene Reconstruction from Uncalibrated Views

Multiple Motion Scene Reconstruction from Uncalibrated Views Multiple Motion Scene Reconstruction from Uncalibrated Views Mei Han C & C Research Laboratories NEC USA, Inc. meihan@ccrl.sj.nec.com Takeo Kanade Robotics Institute Carnegie Mellon University tk@cs.cmu.edu

More information

CS 664 Structure and Motion. Daniel Huttenlocher

CS 664 Structure and Motion. Daniel Huttenlocher CS 664 Structure and Motion Daniel Huttenlocher Determining 3D Structure Consider set of 3D points X j seen by set of cameras with projection matrices P i Given only image coordinates x ij of each point

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Unit 3 Multiple View Geometry

Unit 3 Multiple View Geometry Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover

More information

Linear Multi View Reconstruction and Camera Recovery Using a Reference Plane

Linear Multi View Reconstruction and Camera Recovery Using a Reference Plane International Journal of Computer Vision 49(2/3), 117 141, 2002 c 2002 Kluwer Academic Publishers. Manufactured in The Netherlands. Linear Multi View Reconstruction and Camera Recovery Using a Reference

More information

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.

CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt. CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt. Session 4 Affine Structure from Motion Mani Golparvar-Fard Department of Civil and Environmental Engineering 329D, Newmark Civil Engineering

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263 Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine

More information

Two-view geometry Computer Vision Spring 2018, Lecture 10

Two-view geometry Computer Vision Spring 2018, Lecture 10 Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of

More information

CS231A Course Notes 4: Stereo Systems and Structure from Motion

CS231A Course Notes 4: Stereo Systems and Structure from Motion CS231A Course Notes 4: Stereo Systems and Structure from Motion Kenji Hata and Silvio Savarese 1 Introduction In the previous notes, we covered how adding additional viewpoints of a scene can greatly enhance

More information

A Calibration Algorithm for POX-Slits Camera

A Calibration Algorithm for POX-Slits Camera A Calibration Algorithm for POX-Slits Camera N. Martins 1 and H. Araújo 2 1 DEIS, ISEC, Polytechnic Institute of Coimbra, Portugal 2 ISR/DEEC, University of Coimbra, Portugal Abstract Recent developments

More information

EXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES

EXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES EXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES Camillo RESSL Institute of Photogrammetry and Remote Sensing University of Technology, Vienna,

More information

EECS 442: Final Project

EECS 442: Final Project EECS 442: Final Project Structure From Motion Kevin Choi Robotics Ismail El Houcheimi Robotics Yih-Jye Jeffrey Hsu Robotics Abstract In this paper, we summarize the method, and results of our projective

More information

Agenda. Rotations. Camera calibration. Homography. Ransac

Agenda. Rotations. Camera calibration. Homography. Ransac Agenda Rotations Camera calibration Homography Ransac Geometric Transformations y x Transformation Matrix # DoF Preserves Icon translation rigid (Euclidean) similarity affine projective h I t h R t h sr

More information

Structure from Motion

Structure from Motion /8/ Structure from Motion Computer Vision CS 43, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from motion

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253 Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine

More information

Euclidean Reconstruction and Auto-Calibration from Continuous Motion

Euclidean Reconstruction and Auto-Calibration from Continuous Motion Euclidean Reconstruction and Auto-Calibration from Continuous Motion Fredrik Kahl and Anders Heyden Λ Centre for Mathematical Sciences Lund University Box 8, SE- Lund, Sweden {fredrik, andersp}@maths.lth.se

More information

Lecture 9: Epipolar Geometry

Lecture 9: Epipolar Geometry Lecture 9: Epipolar Geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Why is stereo useful? Epipolar constraints Essential and fundamental matrix Estimating F (Problem Set 2

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Epipolar Geometry in Stereo, Motion and Object Recognition

Epipolar Geometry in Stereo, Motion and Object Recognition Epipolar Geometry in Stereo, Motion and Object Recognition A Unified Approach by GangXu Department of Computer Science, Ritsumeikan University, Kusatsu, Japan and Zhengyou Zhang INRIA Sophia-Antipolis,

More information

CS231M Mobile Computer Vision Structure from motion

CS231M Mobile Computer Vision Structure from motion CS231M Mobile Computer Vision Structure from motion - Cameras - Epipolar geometry - Structure from motion Pinhole camera Pinhole perspective projection f o f = focal length o = center of the camera z y

More information

Motion Segmentation Based on Factorization Method and Discriminant Criterion

Motion Segmentation Based on Factorization Method and Discriminant Criterion Motion Segmentation Based on Factorization Method and Discriminant Criterion Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-, Umezono, Tsukuba Ibaraki, 35-8568 Japan ichimura@etl.go.jp Abstract A motion

More information

Epipolar geometry. x x

Epipolar geometry. x x Two-view geometry Epipolar geometry X x x Baseline line connecting the two camera centers Epipolar Plane plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections

More information

N-View Methods. Diana Mateus, Nassir Navab. Computer Aided Medical Procedures Technische Universität München. 3D Computer Vision II

N-View Methods. Diana Mateus, Nassir Navab. Computer Aided Medical Procedures Technische Universität München. 3D Computer Vision II 1/66 N-View Methods Diana Mateus, Nassir Navab Computer Aided Medical Procedures Technische Universität München 3D Computer Vision II Inspired by Slides from Adrien Bartoli 2/66 Outline 1 Structure from

More information

Agenda. Rotations. Camera models. Camera calibration. Homographies

Agenda. Rotations. Camera models. Camera calibration. Homographies Agenda Rotations Camera models Camera calibration Homographies D Rotations R Y = Z r r r r r r r r r Y Z Think of as change of basis where ri = r(i,:) are orthonormal basis vectors r rotated coordinate

More information

Structure and motion in 3D and 2D from hybrid matching constraints

Structure and motion in 3D and 2D from hybrid matching constraints Structure and motion in 3D and 2D from hybrid matching constraints Anders Heyden, Fredrik Nyberg and Ola Dahl Applied Mathematics Group Malmo University, Sweden {heyden,fredrik.nyberg,ola.dahl}@ts.mah.se

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera

More information

Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15

Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15 Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15 Lecture 6 Stereo Systems Multi- view geometry Stereo systems

More information

Rectification and Distortion Correction

Rectification and Distortion Correction Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification

More information

Vision par ordinateur

Vision par ordinateur Epipolar geometry π Vision par ordinateur Underlying structure in set of matches for rigid scenes l T 1 l 2 C1 m1 l1 e1 M L2 L1 e2 Géométrie épipolaire Fundamental matrix (x rank 2 matrix) m2 C2 l2 Frédéric

More information

1 Projective Geometry

1 Projective Geometry CIS8, Machine Perception Review Problem - SPRING 26 Instructions. All coordinate systems are right handed. Projective Geometry Figure : Facade rectification. I took an image of a rectangular object, and

More information

3D reconstruction class 11

3D reconstruction class 11 3D reconstruction class 11 Multiple View Geometry Comp 290-089 Marc Pollefeys Multiple View Geometry course schedule (subject to change) Jan. 7, 9 Intro & motivation Projective 2D Geometry Jan. 14, 16

More information

Triangulation from Two Views Revisited: Hartley-Sturm vs. Optimal Correction

Triangulation from Two Views Revisited: Hartley-Sturm vs. Optimal Correction Triangulation from Two Views Revisited: Hartley-Sturm vs. Optimal Correction Kenichi Kanatani 1, Yasuyuki Sugaya 2, and Hirotaka Niitsuma 1 1 Department of Computer Science, Okayama University, Okayama

More information

Mathematics of a Multiple Omni-Directional System

Mathematics of a Multiple Omni-Directional System Mathematics of a Multiple Omni-Directional System A. Torii A. Sugimoto A. Imiya, School of Science and National Institute of Institute of Media and Technology, Informatics, Information Technology, Chiba

More information

Metric Structure from Motion

Metric Structure from Motion CS443 Final Project Metric Structure from Motion Peng Cheng 1 Objective of the Project Given: 1. A static object with n feature points and unknown shape. 2. A camera with unknown intrinsic parameters takes

More information

Euclidean Reconstruction Independent on Camera Intrinsic Parameters

Euclidean Reconstruction Independent on Camera Intrinsic Parameters Euclidean Reconstruction Independent on Camera Intrinsic Parameters Ezio MALIS I.N.R.I.A. Sophia-Antipolis, FRANCE Adrien BARTOLI INRIA Rhone-Alpes, FRANCE Abstract bundle adjustment techniques for Euclidean

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Planar pattern for automatic camera calibration

Planar pattern for automatic camera calibration Planar pattern for automatic camera calibration Beiwei Zhang Y. F. Li City University of Hong Kong Department of Manufacturing Engineering and Engineering Management Kowloon, Hong Kong Fu-Chao Wu Institute

More information

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Jae-Hak Kim 1, Richard Hartley 1, Jan-Michael Frahm 2 and Marc Pollefeys 2 1 Research School of Information Sciences and Engineering

More information

Chapter 7: Computation of the Camera Matrix P

Chapter 7: Computation of the Camera Matrix P Chapter 7: Computation of the Camera Matrix P Arco Nederveen Eagle Vision March 18, 2008 Arco Nederveen (Eagle Vision) The Camera Matrix P March 18, 2008 1 / 25 1 Chapter 7: Computation of the camera Matrix

More information

Minimal Projective Reconstruction for Combinations of Points and Lines in Three Views

Minimal Projective Reconstruction for Combinations of Points and Lines in Three Views Minimal Projective Reconstruction for Combinations of Points and Lines in Three Views Magnus Oskarsson, Andrew Zisserman and Kalle Åström Centre for Mathematical Sciences Lund University,SE 221 00 Lund,

More information

Lecture 3: Camera Calibration, DLT, SVD

Lecture 3: Camera Calibration, DLT, SVD Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P

More information

Epipolar Geometry and the Essential Matrix

Epipolar Geometry and the Essential Matrix Epipolar Geometry and the Essential Matrix Carlo Tomasi The epipolar geometry of a pair of cameras expresses the fundamental relationship between any two corresponding points in the two image planes, and

More information

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS Cha Zhang and Zhengyou Zhang Communication and Collaboration Systems Group, Microsoft Research {chazhang, zhang}@microsoft.com ABSTRACT

More information

The Geometry of Dynamic Scenes On Coplanar and Convergent Linear Motions Embedded in 3D Static Scenes

The Geometry of Dynamic Scenes On Coplanar and Convergent Linear Motions Embedded in 3D Static Scenes EXTENDED VERSION SHORT VERSION APPEARED IN THE 13TH BMVC, CARDIFF, SEPTEMBER 2002. The Geometry of Dynamic Scenes On Coplanar and Convergent Linear Motions Embedded in 3D Static Scenes Adrien Bartoli,

More information

Visual Recognition: Image Formation

Visual Recognition: Image Formation Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know

More information

A Canonical Framework for Sequences of Images

A Canonical Framework for Sequences of Images A Canonical Framework for Sequences of Images Anders Heyden, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: andersp@maths.lth.se kalle@maths.lth.se Abstract This

More information

Mei Han Takeo Kanade. January Carnegie Mellon University. Pittsburgh, PA Abstract

Mei Han Takeo Kanade. January Carnegie Mellon University. Pittsburgh, PA Abstract Scene Reconstruction from Multiple Uncalibrated Views Mei Han Takeo Kanade January 000 CMU-RI-TR-00-09 The Robotics Institute Carnegie Mellon University Pittsburgh, PA 1513 Abstract We describe a factorization-based

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

Linear Multi-View Reconstruction of Points, Lines, Planes and Cameras using a Reference Plane

Linear Multi-View Reconstruction of Points, Lines, Planes and Cameras using a Reference Plane Linear Multi-View Reconstruction of Points, Lines, Planes and Cameras using a Reference Plane Carsten Rother Computational Vision and Active Perception Laboratory (CVAP) Royal Institute of Technology (KTH),

More information

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert Week 2: Two-View Geometry Padua Summer 08 Frank Dellaert Mosaicking Outline 2D Transformation Hierarchy RANSAC Triangulation of 3D Points Cameras Triangulation via SVD Automatic Correspondence Essential

More information

THE TRIFOCAL TENSOR AND ITS APPLICATIONS IN AUGMENTED REALITY

THE TRIFOCAL TENSOR AND ITS APPLICATIONS IN AUGMENTED REALITY THE TRIFOCAL TENSOR AND ITS APPLICATIONS IN AUGMENTED REALITY Jia Li A Thesis submitted to the Faculty of Graduate and Postdoctoral Studies in partial fulfillment of the requirements for the degree of

More information

Factorization Method Using Interpolated Feature Tracking via Projective Geometry

Factorization Method Using Interpolated Feature Tracking via Projective Geometry Factorization Method Using Interpolated Feature Tracking via Projective Geometry Hideo Saito, Shigeharu Kamijima Department of Information and Computer Science, Keio University Yokohama-City, 223-8522,

More information

COMP 558 lecture 19 Nov. 17, 2010

COMP 558 lecture 19 Nov. 17, 2010 COMP 558 lecture 9 Nov. 7, 2 Camera calibration To estimate the geometry of 3D scenes, it helps to know the camera parameters, both external and internal. The problem of finding all these parameters is

More information

Combining Two-view Constraints For Motion Estimation

Combining Two-view Constraints For Motion Estimation ombining Two-view onstraints For Motion Estimation Venu Madhav Govindu Somewhere in India venu@narmada.org Abstract In this paper we describe two methods for estimating the motion parameters of an image

More information

CS231A Midterm Review. Friday 5/6/2016

CS231A Midterm Review. Friday 5/6/2016 CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric

More information

MAPI Computer Vision. Multiple View Geometry

MAPI Computer Vision. Multiple View Geometry MAPI Computer Vision Multiple View Geometry Geometry o Multiple Views 2- and 3- view geometry p p Kpˆ [ K R t]p Geometry o Multiple Views 2- and 3- view geometry Epipolar Geometry The epipolar geometry

More information

C280, Computer Vision

C280, Computer Vision C280, Computer Vision Prof. Trevor Darrell trevor@eecs.berkeley.edu Lecture 11: Structure from Motion Roadmap Previous: Image formation, filtering, local features, (Texture) Tues: Feature-based Alignment

More information

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry 55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence

More information

arxiv: v1 [cs.cv] 2 May 2016

arxiv: v1 [cs.cv] 2 May 2016 16-811 Math Fundamentals for Robotics Comparison of Optimization Methods in Optical Flow Estimation Final Report, Fall 2015 arxiv:1605.00572v1 [cs.cv] 2 May 2016 Contents Noranart Vesdapunt Master of Computer

More information

HW 1: Project Report (Camera Calibration)

HW 1: Project Report (Camera Calibration) HW 1: Project Report (Camera Calibration) ABHISHEK KUMAR (abhik@sci.utah.edu) 1 Problem The problem is to calibrate a camera for a fixed focal length using two orthogonal checkerboard planes, and to find

More information

Towards the completion of assignment 1

Towards the completion of assignment 1 Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?

More information

Affine-invariant shape matching and recognition under partial occlusion

Affine-invariant shape matching and recognition under partial occlusion Title Affine-invariant shape matching and recognition under partial occlusion Author(s) Mai, F; Chang, CQ; Hung, YS Citation The 17th IEEE International Conference on Image Processing (ICIP 2010), Hong

More information

Identifying Car Model from Photographs

Identifying Car Model from Photographs Identifying Car Model from Photographs Fine grained Classification using 3D Reconstruction and 3D Shape Registration Xinheng Li davidxli@stanford.edu Abstract Fine grained classification from photographs

More information

A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space

A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-4, Umezono, Tsukuba Ibaraki, 35-8568 Japan ichimura@etl.go.jp

More information

Comments on Consistent Depth Maps Recovery from a Video Sequence

Comments on Consistent Depth Maps Recovery from a Video Sequence Comments on Consistent Depth Maps Recovery from a Video Sequence N.P. van der Aa D.S. Grootendorst B.F. Böggemann R.T. Tan Technical Report UU-CS-2011-014 May 2011 Department of Information and Computing

More information

Camera Model and Calibration

Camera Model and Calibration Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the

More information

Lecture 6 Stereo Systems Multi-view geometry

Lecture 6 Stereo Systems Multi-view geometry Lecture 6 Stereo Systems Multi-view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-5-Feb-4 Lecture 6 Stereo Systems Multi-view geometry Stereo systems

More information

Motion Segmentation Based on Factorization Method and Discriminant Criterion

Motion Segmentation Based on Factorization Method and Discriminant Criterion Motion Segmentation Based on Factorization Method and Discriminant Criterion Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-, Umezono, Tsukuba Ibaraki, -6 Japan ichimura@etl.go.jp Abstract A motion segmentation

More information

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Computer Vision Projective Geometry and Calibration. Pinhole cameras Computer Vision Projective Geometry and Calibration Professor Hager http://www.cs.jhu.edu/~hager Jason Corso http://www.cs.jhu.edu/~jcorso. Pinhole cameras Abstract camera model - box with a small hole

More information

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10 Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley

More information

Multi-View AAM Fitting and Camera Calibration

Multi-View AAM Fitting and Camera Calibration To appear in the IEEE International Conference on Computer Vision Multi-View AAM Fitting and Camera Calibration Seth Koterba, Simon Baker, Iain Matthews, Changbo Hu, Jing Xiao, Jeffrey Cohn, and Takeo

More information

Factorization Methods for Projective Structure and Motion

Factorization Methods for Projective Structure and Motion Factorization Methods for Projective Structure and Motion Bill Triggs LIFIA, INRIA Rhône-Alpes, 6, avenue Félix Viallet, 383 Grenoble, France. Bill.Triggs@imag.fr Abstract This paper describes a family

More information

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with

More information

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp

More information

Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds

Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds Marta Wilczkowiak Edmond Boyer Peter Sturm Movi Gravir Inria Rhône-Alpes, 655 Avenue de l Europe, 3833 Montbonnot, France

More information

An Overview of Matchmoving using Structure from Motion Methods

An Overview of Matchmoving using Structure from Motion Methods An Overview of Matchmoving using Structure from Motion Methods Kamyar Haji Allahverdi Pour Department of Computer Engineering Sharif University of Technology Tehran, Iran Email: allahverdi@ce.sharif.edu

More information

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry

More information

Computer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman Assignment #1. (Due date: 10/23/2012) x P. = z

Computer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman   Assignment #1. (Due date: 10/23/2012) x P. = z Computer Vision I Name : CSE 252A, Fall 202 Student ID : David Kriegman E-Mail : Assignment (Due date: 0/23/202). Perspective Projection [2pts] Consider a perspective projection where a point = z y x P

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES

URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment

More information