Projective reconstruction from line-correspondences in multiple uncalibrated images
|
|
- Posy Jackson
- 5 years ago
- Views:
Transcription
1 Pattern Recognition 39 (2006) Projective reconstruction from line-correspondences in multiple uncalibrated images A.W.K. Tang, T.P. Ng, Y.S. Hung, C.H. Leung Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam Road, Hong Kong Received 30 June 2005 Abstract A new approach is proposed for reconstructing 3D lines and cameras from 2D corresponding lines across multiple uncalibrated views. There is no requirement that the 2D corresponding lines on different images represent the same segment of a 3D line, which may not appear on all images. A 3D line is reconstructed by minimizing a geometric cost function that measures the distance of the reprojected end points of the 3D segment from the measured 2D lines on different images. An algorithmic procedure is provided with guaranteed convergence to a solution where the geometric cost function achieves a (local) minimum Pattern Recognition Society. Published by Elsevier Ltd. All rights reserved. Keywords: Multiple views; Projective reconstruction; Line reconstruction; Line correspondence 1. Introduction The recovery of a 3D scene from uncalibrated images has long been a major area in computer vision. Existing reconstruction methods mainly make use of feature point correspondences [1 3] across multiple views. Recently, 3D reconstruction from line correspondences [4 12] has attracted some attention since there are many merits of reconstruction from lines over points. In the matching process, for example, reconstruction from lines allows correspondence of different sections of a line in different views. Dornaika and Garcia [13] proposed two 3D reconstruction methods from point and line correspondences for weak perspective and paraperspective camera models. Ansar and Daniilidis [14] proposed a linear method for 3D reconstruction from point and line correspondences under the assumption that the camera intrinsic parameters are known. Both of these methods (i.e. Refs. [13] and [14]), however, make a priori assumptions about the cameras. Corresponding author. Tel.: ; fax: addresses: wktang@eee.hku.hk (A.W.K. Tang), NgTuenPui@graduate.hku.hk (T.P. Ng), yshung@eee.hku.hk (Y.S. Hung), chleung@eee.hku.hk (C.H. Leung). For uncalibrated cameras, Hartley [4] introduced the geometry between line correspondences across three images which was later developed into the trifocal tensor [6] governing point and line correspondences in three images analogous to that of the fundamental matrix for two images. Explicit linear algorithms for computing the trifocal tensor from line correspondences are available and camera projection matrices can be readily obtained from the trifocal tensor. Methods based on these multilinear constraints [4 6], however, are noise sensitive and applicable only for three or four views, and they minimize an algebraic rather than geometric error. A different approach for line reconstruction is to parameterize a 3D line as a 6-vector in Plücker coordinates. Faugeras and Mourrain [15] propose to transform a general projection matrix for points to a line projection matrix in Plücker coordinates, so that the linear relationship of projecting a 3D line onto an image plane as a 2D line can be represented by the line projection matrix similar to the projection equation for points. There are several methods [1,8 10,15] based on the linear projection equation for lines in Plücker coordinates. Maolin et al. [8] propose an iterative projective reconstruction for estimating scale factors, and Martinec and Pajdla [9] propose a non-iterative /$ Pattern Recognition Society. Published by Elsevier Ltd. All rights reserved. doi: /j.patcog
2 890 A.W.K. Tang et al. / Pattern Recognition 39 (2006) factorization method. The main problem of the 3D line reconstruction in the Plücker coordinates is that both the line and line projection matrix should satisfy some conditions which cannot be enforced directly during the reconstruction process. Instead, they are enforced at the end of the reconstruction process with the introduction of unpredictable errors due to an algebraic condition with no physical meaning. Moreover, the reconstructed line projection matrices need to be transformed back to general projection matrices for upgrading the projective to Euclidean space. Bartoli and Sturm [16] show empirically that the existing linear methods perform poorly in most cases, and they propose a bundle adjustment method based on orthonormal representation of 3D lines in the Plücker coordinates for minimizing a geometric quantity, namely the orthogonal distances between the reprojected lines and the two 2D end points of the measured line segments. Soe and Hong [10] propose a sequential line reconstruction method using iterative extended Kalman filter (IEKF) by adding short baseline frames one by one to update the scene reconstructed from three wide-baseline frames. As this method relies on the initial estimates of the projection matrices computed from the first three views and they are updated iteratively, the method is biased towards the images input earlier. In Ref. [12], Triggs uses the factorization-based method to perform 3D reconstruction from both point and line correspondences. However, the method needs a set of point correspondences to compute fundamental matrices which are then used to convert the line correspondences into additional point correspondences. Furthermore, the algebraic rather than geometric error is minimized, and no missing data is allowed. In this paper, we will propose a line reconstruction method adopted from Ref. [1], which performs projective reconstruction from point correspondences in multiple views by minimizing 2D reprojection errors using a quasi-linear bundle adjustment approach. The proposed method reconstructs a line in 3D space as a segment with two suitably chosen end points whose projections on images are as close as possible in a geometric sense to measured lines that correspond with each other. Despite the apparent simplicity of the idea, there are many intricacies in the design of a complete solution using this approach, and we are not aware of any similar method for line reconstruction having the following characteristics: the cost function represents a geometric measure of the goodness of reconstruction; the reconstruction is truly based on line correspondences with no need for point correspondences; the method is able to handle missing lines; the optimization algorithm incorporates bundle adjustment and is guaranteed to converge to a local minimum of the geometric cost function. The paper is organized as follows. In Section 2, we will give a brief summary of the point-based quasi-linear bundle adjustment method of Ref. [1] that will be adopted for the proposed line reconstruction method. In Section 3, the line reconstruction problem is formulated as a minimization of a geometric cost and an algorithmic solution is developed. Experimental results are given in Section 4. Section 5 contains some concluding remarks. Notation: The Hadamard product of two matrices A=[a ij ] and B =[b ij ] of the same size is denoted A B =[a ij b ij ]. 2. Point-based method for projective reconstruction Suppose a set of n image point correspondences x ij are established across m views. The jth 3D point X j R 4 1 projected on the ith view by a projection matrix P i R 3 4 as a 2D point x ij =[u ij v ij 1] T can be expressed as λ ij x ij = P i X j, (1) where λ ij is the depth of X j to the ith camera. The projection of all the 3D points onto all the cameras can be represented in matrix form as λ 11 x 11 λ 1n x 1n M = = PX R 3m n, (2) λ m1 x m1 λ mn x mn where M is the scaled measurement matrix, P =[P T 1 P T 2 P T m ]T R 3m 4 is the joint projection matrix and X = [X 1 X 2 X n ] R 4 n is the shape matrix. If the projective depths [λ ij ] are known, the scaled measurement matrix can be factorized into P and X by means of singular value decomposition. Various methods are proposed to determine the projective depths [1,2]. In particular, the depths are estimated iteratively in Ref. [1] as a minimization problem with the cost function i=m,j=n min γ ij (x ij β ij P i X j ) 2, (3) P,X,β i=1,j=1 where β ij (=1/λ ij ) is the inverse depth, γ ij =[11γγ ij ] T, γ is a control factor to force the cost (3) to approach the 2D reprojection error and γ ij are weighting factors for balancing the magnitude of pixel coordinates. This method has the advantages that it minimizes the 2D reprojection error and is capable of dealing with missing data by considering only the visible 2D points in (3). An iterative algorithm can be used to estimate the unknown parameters β ij, P i and X j alternately as linear least-squares problems and therefore the solution process is a quasi-linear bundle adjustment. In the next section, we will extend this formulation to the case of line reconstruction.
3 A.W.K. Tang et al. / Pattern Recognition 39 (2006) Line reconstruction Estimated 3D Points 3.1. Representing lines In Fig. 1, a 2D line is represented as a 3-vector l. The minimum distance (i.e. orthogonal distance) between a 2D point x and the 2D line l is given by d (x, l) = l T x, (4) (l) (l)2 2 where (l) k is the kth element of the vector l. Ifx lies on l, d (x, l) = 0. A line can also be represented by any two distinct points (e.g. x 1 and x 2 for 2D cases or X 1 and X 2 for 3D cases) which lie on it. The points lying on that line between these two points can be expressed as αx 1 + (1 α)x 2 for a 2D line or αx 1 + (1 α)x 2 for a 3D line, 0 α 1. The proposed method relies on these two kinds of line representations The error measure for line reconstruction A critical issue of line reconstruction methods is how to measure the difference between a measured line and its reprojected line from the reconstructed 3D line. Let the jth measured 2D line segment on the ith view be denoted l ij and its two end points be denoted x ij and x i,n+j so that l T ij x ij =l T ij x i,n+j =0, where n is the total number of 3D lines. Let the measured 2D line segments be indexed by ordered pairs of the set A ={(i, j) l ij with end points x ij and x i,n+j is observed as the jth line on the ith view}. The idea of the proposed method is to reconstruct a 3D line segment with end points X j and X n+j whose reprojected end points on the ith view will be close to l ij. Denote the reprojected end points of X j and X n+j on the ith view by ˆx ij and ˆx i,n+j, respectively. Making use of (4), the measure of error in the line reconstruction can be defined as ε 2 = 1 2 n(a) (i,j) A (lij T ˆx i,j ) 2 + (lij T ˆx i,n+j ) 2 (l ij ) (l ij ) 2, (5) 2 where n(a) is the number of elements of the set A. ε can be regarded as the root mean square error (RMSE) of the O l (l) (l) 1 + (l) 2 d Fig. 1. Line on a 2D plane. x Estimated Camera Centre Measured line with two end points Reprojected Points Fig. 2. The projection of a line. reprojected lines to the measured line segments. It is also possible to simplify (5) by normalizing l ij so that (l ij ) (l ij ) 2 2 = 1, i, j. In the sequel, we will assume that all lines are already normalized Analyses on using lines for projective reconstruction In Fig. 2, the reprojection of the estimated end points of a 3D line segment on an image is illustrated. Formulating a minimization problem directly from (5) to estimate the 3D line segments without extra constraint(s) will not provide a reasonable solution. This is because it is always easier to fit a single point than two end points of a line segment to the projection model, and therefore there is a tendency during optimization for the two end points of a line segment to converge to one single point that is closest to a measured line. In order to prevent the two end points from converging into one point, we require the two reprojected end points of the 3D line segment to approach the end points of a particular measured line segment on one of the image planes, and this line will be referred to as the reference line segment. Thus, the reference line segment acts as an anchor for fixing the section of the 3D line to be reconstructed. For each 3D line, only one line segment is chosen as the reference line segment. For all other lines l ij, the distance of the reprojected end points of the 3D line segment X j X n+j from l ij is taken to be the orthogonal distance from the (extended) line l ij without any regard to the actual positions of the end points of the measured line. Fig. 3 shows the idea of reconstructing the 3D line segment by requiring its reprojected end points to approach the measured end points of a reference line segment (leftmost image), whereas the reprojected end points on the other two views are only required to be as close as possible to the measured lines with no other restrictions on the locations of the reprojected points Choice of reference line segments The strategy of reference line segment selection directly affects the accuracy of the final reconstruction. The
4 892 A.W.K. Tang et al. / Pattern Recognition 39 (2006) ˆx i,n+j xˆ ij l rj x i,n+j θ n+j θ j x ij Fig. 4. The relationship between 2D reprojection error and orthogonal distance. is related to the orthogonal distance to the line as: Fig. 3. The reconstruction process. recommended strategy is to choose the line segment having the maximum length in pixel coordinate among the group of corresponding lines across multiple views. If the same amount of disturbance is applied to the end points of line segments having different lengths, the disturbed line will be closer to the original line when the ratio of disturbance to the length of line segment is smaller. Let the reference line segments be indexed by ordered pairs of the set R ={(r, j) l rj is chosen as a reference line segment} Cost function for line reconstruction To formulate the line reconstruction method stated above, the cost function of the proposed method is defined as F L (P, X, β) = 1 1 γ 2n(A) r,kn+j (x r,kn+j k=0 (r,j) R β r,kn+j P r X kn+j ) 2 + (β i,kn+j lij T P ix kn+j ) 2 (i,j) A\R + γ 2 γ 2 i,kn+j (1 β i,kn+j Pi 3 X kn+j ) 2, (6) where P s i is the sth row of P i and γ i,kn+j =[11γγ i,kn+j ] T. The first summation in Eq. (6) is the mean squares of the reprojection errors for all the end points of the reference line segments. This quantity becomes the 2D reprojection error when γ tends to infinity. The second summation becomes the mean squares of all the orthogonal distances of the reprojected points to the measured lines when γ tends to infinity. The cost (6) is also the upper bound error of Eq. (5). In Fig. 4, the 2D reprojection error of the reprojected points ˆx i,j and ˆx i,n+j from the end points of the reference line l ij sum of squared orthogonal distances = ˆx r,j x r,j 2 sin 2 θ j + ˆx r,n+j x r,n+j 2 sin 2 θ n+j, θ j, θ n+j R ˆx r,j x r,j 2 + ˆx r,n+j x r,n+j 2 sum of squared 2D reprojection errors. The cost function (6) can therefore be treated as an approximation to the error measure (5) with a constraint on the end points of the reconstructed line Choice of weighting factors The purpose of the weighting factor γ ij is to balance the magnitude of the pixel coordinates with the last element (i.e. 1) of the homogeneous coordinates. During the iterations of the proposed method, γ ij is fixed. There are different considerations for the two terms in Eq. (6). For the first term of the minimization of 2D reprojection error, we choose γ rj = max( u rj, v rj ), (r, j) R. (7) This weighting factor helps to balance the distribution among the three components of the 2D homogeneous coordinates. For the second term of the minimization of orthogonal distances, γ ij is chosen as (l ij ) 3 (l γ ij = rj ) 3 max( u rj, v rj, u r,n+j, v r,n+j ), (l rj ) 3 = 0 (l ij ) 3, (l rj ) 3 = 0 (i, j) A\R,(r,j) R and i = r, (8) where the lines l ij and l rj are normalized and the 2D line l rj on the rth view is chosen as the reference line segment for the jth 3D line. Note that (l ij ) 3 is the orthogonal distance between the 2D line from the image origin. Similar to (7), the weighting factor (8) is meant to scale the third component of the homogeneous coordinates to magnitudes comparable with the pixel coordinates of the reprojected points on the ith view. We note however that the final results are not unduly sensitive to the choice of γ ij, and as γ is increased
5 progressively in the algorithm (see step 7 of algorithm below), eventually γγ ij irrespective of the choice of γ ij A quasi-linear bundle adjustment algorithm for line reconstruction A.W.K. Tang et al. / Pattern Recognition 39 (2006) Put k = 0, γ = 1, ε b = 0; Set β 0 ij = 1, i, j and compute P 0 and X 0 from a rank-4 SVD approximation of [(1/β 0 ij )x ij ]. Select reference line segments and compute γ ij. 2. Put k = k Fix X k 1 and β k 1 and determine P k by solving min P kf L (P k,x k 1, β k 1 ). 4. Fix P k and β k 1 and determine X k by solving min X kf L (P k,x k, β k 1 ). 5. Fix P k and X k and determine β k by solving ε k = min β kf L (P k,x k, β k ). 6. Repeat steps 2 5 until ε k converges. 7. If ε k ε b is bigger than a threshold value, then put ε b =ε k, γ = 1.1 γ,k= 0 and return to step 2, else stop. At the end of the algorithm, the projection matrices are taken as Pi k and the 3D lines are defined by the points Xj k and Xn+j k. When some 3D lines are not visible to all the views, missing entries in the measurement matrix should be filled in the initialization stage. The end points of a missing line can be filled by the centroid of the end points of the measured line segments on the view. This method for initializing missing data is used in all examples in this paper. We found that this initialization works well in all cases when the percentage of missing line segments is less than 60% Convergence The proposed method minimizes the cost function, F L (P, X, β), by estimating P, X and β iteratively as three weighted least squares problems. It can be shown that as γ, F L (P, X, β) is guaranteed to converge to a (local) minimum of the cost function with a geometric meaning given in Section Experimental results The proposed method is evaluated by both synthetic and real data. The method is implemented using Matlab 6.5 running on a Pentium-42.8 GHz PC Synthetic data Fig. 5 shows a view of the synthetic scene consisting of a box and five cameras. The box of size 1m 1m 1m is made up of 29 3D lines with patterns of triangles and rectangles on its sides. Five cameras are randomly generated RMS 2D reprojection error / pixel Fig. 5. A view of the synthetic scene noise level / pixel Fig. 6. Performance of the algorithm with data of difference noise levels. such that all of them point at the centroid of the box and the image of the box almost occupies the whole image. The intrinsic parameters are fixed over all views and the image size is pixels. The two vertices of each 2D line segment are chosen as its two measured points and they are perturbed by Gaussian noise independently. Line segments in the first view are chosen as the reference line segments. The reconstruction results for noisy data having Gaussian noise level ranging from 0 to 4 pixels with increments of 0.5 pixels are shown in Fig. 6. The noises are added to the x- and y-coordinates of the corner points independently. For each noise level, the estimated reprojection error is taken as the mean of the root mean square values calculated in 30 trials with different randomly generated noise. Note from Fig. 6 that the RMS error is proportional to the added noise. We have also implemented the line factorization method [9] in Matlab. Although the method works fine for noisefree data, the results are however very sensitive to added noise and it is not always possible to obtain a reasonable solution even at low noise levels. Hence we cannot perform a sensible comparison with our method Real data We consider two real image sequences, the clock tower sequence and the toy house sequence. Feature lines of the objects are manually extracted and matched.
6 894 A.W.K. Tang et al. / Pattern Recognition 39 (2006) Fig. 7. Image sequence of the clock tower. Fig. 9. The reconstructed tower. Fig. 8. The reconstructed lines of the reconstructed tower Clock tower Four images were taken around the clock tower shown in Fig. 7 by a Canon D60 digital camera with a 70 mm lens. The image size is pixels and the intrinsic parameters are varied due to auto-focusing. Line segments on the images corresponding to 22 3D line segments of the tower are visible in all views but different portions of the same segment are seen in different views. The measured line segments are shown as blue solid lines superimposed on the four images in Fig. 7. Note that some parts of the feature lines are blocked by the palm trees in front of the tower in the second image from the left. The proposed method took s and 679 iterations to converge. The RMS 2D reprojection error of the reference line segments is pixels and the RMS orthogonal distance is pixels. Using the estimated projection matrices, the reconstructed 3D line segments can be reprojected onto the images and these are plotted as dotted green line segments on two of the images. Fig. 8 shows the reprojected line segments (dotted green segments) superimposed on the solid blue measured lines for comparison. Fig. 9 shows different views of the reconstructed tower in the Euclidean frame. Fig. 9(a) shows the estimated positions of the cameras taking the images. Fig. 9(b) shows a close-up view of the tower. Fig. 9(c) is the top view of the tower, which shows that the angle between two sides of the tower is almost Toy house Nine images were taken around a toy house by a Canon A80 digital camera, as shown in Fig. 10. The image size is Fig. 10. Image sequence of toy house. Fig. 11. Distribution of measured line segments for image sequence of toy house pixels and the intrinsic parameters are varied due to auto-focusing. As not all the parts of the house are visible in all images, this image sequence demonstrates the performance of the proposed method for dealing with missing line segments in some of the images. In these images, each line segment is visible in at least three images. Line segments on the images corresponding to 57 3D line segments are extracted manually and there are altogether 271 missing lines on the images (about 57.2% of all 2D lines). The manually extracted lines are shown as blue lines in Fig. 10. The distribution of measured line segments are shown in Fig. 11 where a blue cross in the (i, j) position indicates
7 A.W.K. Tang et al. / Pattern Recognition 39 (2006) Fig. 12. The reconstructed house in Euclidean frame. that the jth 3D line can be observed by the ith camera as a 2D line l ij. The proposed method took 466 s with 1187 iterations to converge. The RMS 2D reprojection error of the reference line segments is pixels and the RMS orthogonal distance is pixels. The results of the reconstruction are shown in Fig. 12, where different views of the reconstructed house after being upgraded to the Euclidean frame are presented. 5. Conclusion In this paper, a projective reconstruction method for line correspondences has been developed and evaluated by synthetic data and real image sequences. We reconstruct a 3D line segment by minimizing the sum of orthogonal distances of the reprojected end points of the 3D segment from the measured 2D lines in all the images, except for a reference line segment where the distance is measured from the actual end points of the measured segment. The problem of line reconstruction is thus formulated purely on the basis of line correspondences in multiple views. There is no requirement that the 2D corresponding lines visible on different images represent the same segment of a 3D line. The method does not require point correspondence and its ability to handle missing lines is demonstrated by real image sequences. Experimental results based on synthetic data show that the proposed method is robust to added noise and the RMS orthogonal distance is almost linearly proportional to the added noise. 6. Summary In this paper, the problem of using line correspondences for projective reconstruction from multiple uncalibrated images is addressed. We propose a new approach to reconstruct 3D lines from 2D corresponding lines across multiple views. The reconstruction problem is formulated purely on the basis of line correspondences in multiple views. There is no requirement that the 2D corresponding lines visible on different images represent the same segment of a 3D line, and the lines are not required to be visible in all the views. Most 3D reconstruction methods make use of feature point correspondences across multiple views. However, reconstruction from line correspondences has many merits over point-based methods. Existing methods for projective reconstruction from line correspondences are mainly based on multilinear geometric constraints or the linear projection equation for lines in Plücker coordinates. These methods have the drawback of being excessively noise sensitive and they minimize an algebraic rather than geometric error. A further problem of 3D line reconstruction in Plücker coordinates is that both the line and line projection matrix should satisfy some conditions which cannot be enforced directly during the reconstruction process. Instead, they are enforced at the end of the reconstruction process with the introduction of unpredictable errors due to the imposition of an algebraic condition with no physical meaning. In this paper, we propose a new approach for reconstructing 3D lines from 2D line correspondences. A 3D line segment is reconstructed by minimizing a geometric cost function that measures the distance of the reprojected end points of the 3D segment from the measured 2D lines on different images. The distance is defined to be the orthogonal distance of the reprojected end points of the 3D segment from the measured 2D lines in all the images, except for a reference line segment where the distance is measured from the actual end points of the measured segment. As the choice of reference line segments directly affects the accuracy of the final reconstruction, we propose a strategy to choose reference line segments among the groups of corresponding lines across multiple views. A quasi-linear bundle adjustment algorithm is provided which is guaranteed to converge to a solution where the geometric cost function achieves a (local) minimum. Experimental results based on synthetic data show that the proposed method is robust to added noise and the RMS error measure is almost linearly proportional to the added noise. Real image sequences demonstrate the performance of the proposed method for dealing with cases when different sections of the same segment are seen in different views and some lines are missing in some of the views. Acknowledgements The work described in this paper was substantially supported by a grant from the Research Grants Council of Hong Kong Special Administrative Region, China (Project No. HKU7058/02E).
8 896 A.W.K. Tang et al. / Pattern Recognition 39 (2006) References [1] W.K. Tang, Y.S. Hung, A factorization-based method for projective reconstruction with minimization of 2D reprojection errors, in: Proceedings of the DAGM 2002, September, 2002, pp [2] P. Sturm, B. Triggs, A factorization based algorithm for multipleimage projective structure and motion, in: Proceedings of the European Conference on Computer Vision, Cambridge, England, 1996, pp [3] C. Tomasi, T. Kanade, Shape and motion from image streams under orthography: a factorization method, Int. J. Comput. Vision 9 (2) (1992) [4] R. I. Hartley, Projective reconstruction from line correspondences, in: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1994, pp [5] R.I. Hartley, Multilinear relationships between coordinates of corresponding image points and lines, in: Proceedings of the Sophus Lie Symposium, Nordfjordeid, Norway, [6] R.I. Hartley, Lines and points in three views and the trifocal tensor, Int. J. Comput. Vision 22 (2) (1997) [7] Y. Liu, S. Huang, A linear algorithm for motion estimation using straight line correspondences, in: Proceedings of the IEEE Ninth International Conference on Pattern Recognition, vol. 1, 1988, pp [8] H. Maolin, Z. Quanbing, W. Sui, Projective reconstruction from lines based on SVD, in: Proceedings of the SPIE The International Society for Optical Engineering, vol. 4875, 2002, pp [9] D. Martinec, T. Pajdla, Line reconstruction from many perspective images by factorization, in: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1, 2003, pp [10] Y. Seo, K. S. Hong, Sequential reconstruction of lines in projective space, in: Proceedings of the 13th International Conference in Pattern Recognition, vol. 1, 1996, pp [11] J. Weng, S. Huang, N. Ahuja, Motion and structure from line correspondences: closed-form solution, uniqueness and optimization, IEEE Trans. Pattern Anal. Mach. Intell. 14 (3) (1992) [12] B. Triggs, Factorization methods for projective structure and motion, in: Proceedings of the Conference on Computer Vision and Pattern Recognition, San Franciso, June, 1996, pp [13] F. Donaika, C. Garcia, Pose estimation using point and line correspondences, Real-Time Imaging 5 (1999) [14] A. Ansar, K. Daniilidis, Linear pose estimation from points or lines, IEEE Trans. Pattern Anal. Mach. Intell. 25 (5) (2003) [15] O. Faugeras, B. Mourrain, On the geometry and algebra of the point and line correspondences between N images, in: Proceedings of the Fifth International Conference on Computer Vision, 1995, pp [16] A. Bartoli, P. Sturm, Multiple-view structure and motion from line correspondences, in: Proceedings of the International Conference on Computer Vision, 2003.
Perception and Action using Multilinear Forms
Perception and Action using Multilinear Forms Anders Heyden, Gunnar Sparr, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: {heyden,gunnar,kalle}@maths.lth.se Abstract
More informationMultiple-View Structure and Motion From Line Correspondences
ICCV 03 IN PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, NICE, FRANCE, OCTOBER 003. Multiple-View Structure and Motion From Line Correspondences Adrien Bartoli Peter Sturm INRIA
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t R 2 3,t 3 Camera 1 Camera
More informationStructure from Motion CSC 767
Structure from Motion CSC 767 Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R,t R 2,t 2 R 3,t 3 Camera??
More informationStructure from Motion
11/18/11 Structure from Motion Computer Vision CS 143, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationA Factorization Method for Structure from Planar Motion
A Factorization Method for Structure from Planar Motion Jian Li and Rama Chellappa Center for Automation Research (CfAR) and Department of Electrical and Computer Engineering University of Maryland, College
More informationFactorization with Missing and Noisy Data
Factorization with Missing and Noisy Data Carme Julià, Angel Sappa, Felipe Lumbreras, Joan Serrat, and Antonio López Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona,
More informationA Summary of Projective Geometry
A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]
More informationMultiple View Geometry in Computer Vision
Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Structure Computation Lecture 18 March 22, 2005 2 3D Reconstruction The goal of 3D reconstruction
More informationBIL Computer Vision Apr 16, 2014
BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm
More informationcalibrated coordinates Linear transformation pixel coordinates
1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial
More informationCSE 252B: Computer Vision II
CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Haowei Liu LECTURE 16 Structure from Motion from Tracked Points 16.1. Introduction In the last lecture we learned how to track point features
More informationComputer Vision I - Algorithms and Applications: Multi-View 3D reconstruction
Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View
More informationCamera Calibration Using Line Correspondences
Camera Calibration Using Line Correspondences Richard I. Hartley G.E. CRD, Schenectady, NY, 12301. Ph: (518)-387-7333 Fax: (518)-387-6845 Email : hartley@crd.ge.com Abstract In this paper, a method of
More informationMachine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy
1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:
More informationThe end of affine cameras
The end of affine cameras Affine SFM revisited Epipolar geometry Two-view structure from motion Multi-view structure from motion Planches : http://www.di.ens.fr/~ponce/geomvis/lect3.pptx http://www.di.ens.fr/~ponce/geomvis/lect3.pdf
More informationStructure from Motion
Structure from Motion Outline Bundle Adjustment Ambguities in Reconstruction Affine Factorization Extensions Structure from motion Recover both 3D scene geoemetry and camera positions SLAM: Simultaneous
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationMultiple Motion Scene Reconstruction from Uncalibrated Views
Multiple Motion Scene Reconstruction from Uncalibrated Views Mei Han C & C Research Laboratories NEC USA, Inc. meihan@ccrl.sj.nec.com Takeo Kanade Robotics Institute Carnegie Mellon University tk@cs.cmu.edu
More informationCS 664 Structure and Motion. Daniel Huttenlocher
CS 664 Structure and Motion Daniel Huttenlocher Determining 3D Structure Consider set of 3D points X j seen by set of cameras with projection matrices P i Given only image coordinates x ij of each point
More informationFlexible Calibration of a Portable Structured Light System through Surface Plane
Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured
More informationUnit 3 Multiple View Geometry
Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover
More informationLinear Multi View Reconstruction and Camera Recovery Using a Reference Plane
International Journal of Computer Vision 49(2/3), 117 141, 2002 c 2002 Kluwer Academic Publishers. Manufactured in The Netherlands. Linear Multi View Reconstruction and Camera Recovery Using a Reference
More informationCEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt.
CEE598 - Visual Sensing for Civil Infrastructure Eng. & Mgmt. Session 4 Affine Structure from Motion Mani Golparvar-Fard Department of Civil and Environmental Engineering 329D, Newmark Civil Engineering
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263
Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine
More informationTwo-view geometry Computer Vision Spring 2018, Lecture 10
Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of
More informationCS231A Course Notes 4: Stereo Systems and Structure from Motion
CS231A Course Notes 4: Stereo Systems and Structure from Motion Kenji Hata and Silvio Savarese 1 Introduction In the previous notes, we covered how adding additional viewpoints of a scene can greatly enhance
More informationA Calibration Algorithm for POX-Slits Camera
A Calibration Algorithm for POX-Slits Camera N. Martins 1 and H. Araújo 2 1 DEIS, ISEC, Polytechnic Institute of Coimbra, Portugal 2 ISR/DEEC, University of Coimbra, Portugal Abstract Recent developments
More informationEXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES
EXPERIMENTAL RESULTS ON THE DETERMINATION OF THE TRIFOCAL TENSOR USING NEARLY COPLANAR POINT CORRESPONDENCES Camillo RESSL Institute of Photogrammetry and Remote Sensing University of Technology, Vienna,
More informationEECS 442: Final Project
EECS 442: Final Project Structure From Motion Kevin Choi Robotics Ismail El Houcheimi Robotics Yih-Jye Jeffrey Hsu Robotics Abstract In this paper, we summarize the method, and results of our projective
More informationAgenda. Rotations. Camera calibration. Homography. Ransac
Agenda Rotations Camera calibration Homography Ransac Geometric Transformations y x Transformation Matrix # DoF Preserves Icon translation rigid (Euclidean) similarity affine projective h I t h R t h sr
More informationStructure from Motion
/8/ Structure from Motion Computer Vision CS 43, Brown James Hays Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz, and Martial Hebert This class: structure from motion
More informationIndex. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253
Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine
More informationEuclidean Reconstruction and Auto-Calibration from Continuous Motion
Euclidean Reconstruction and Auto-Calibration from Continuous Motion Fredrik Kahl and Anders Heyden Λ Centre for Mathematical Sciences Lund University Box 8, SE- Lund, Sweden {fredrik, andersp}@maths.lth.se
More informationLecture 9: Epipolar Geometry
Lecture 9: Epipolar Geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Why is stereo useful? Epipolar constraints Essential and fundamental matrix Estimating F (Problem Set 2
More informationCOMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION
COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA
More informationEpipolar Geometry in Stereo, Motion and Object Recognition
Epipolar Geometry in Stereo, Motion and Object Recognition A Unified Approach by GangXu Department of Computer Science, Ritsumeikan University, Kusatsu, Japan and Zhengyou Zhang INRIA Sophia-Antipolis,
More informationCS231M Mobile Computer Vision Structure from motion
CS231M Mobile Computer Vision Structure from motion - Cameras - Epipolar geometry - Structure from motion Pinhole camera Pinhole perspective projection f o f = focal length o = center of the camera z y
More informationMotion Segmentation Based on Factorization Method and Discriminant Criterion
Motion Segmentation Based on Factorization Method and Discriminant Criterion Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-, Umezono, Tsukuba Ibaraki, 35-8568 Japan ichimura@etl.go.jp Abstract A motion
More informationEpipolar geometry. x x
Two-view geometry Epipolar geometry X x x Baseline line connecting the two camera centers Epipolar Plane plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections
More informationN-View Methods. Diana Mateus, Nassir Navab. Computer Aided Medical Procedures Technische Universität München. 3D Computer Vision II
1/66 N-View Methods Diana Mateus, Nassir Navab Computer Aided Medical Procedures Technische Universität München 3D Computer Vision II Inspired by Slides from Adrien Bartoli 2/66 Outline 1 Structure from
More informationAgenda. Rotations. Camera models. Camera calibration. Homographies
Agenda Rotations Camera models Camera calibration Homographies D Rotations R Y = Z r r r r r r r r r Y Z Think of as change of basis where ri = r(i,:) are orthonormal basis vectors r rotated coordinate
More informationStructure and motion in 3D and 2D from hybrid matching constraints
Structure and motion in 3D and 2D from hybrid matching constraints Anders Heyden, Fredrik Nyberg and Ola Dahl Applied Mathematics Group Malmo University, Sweden {heyden,fredrik.nyberg,ola.dahl}@ts.mah.se
More informationStructure from motion
Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera
More informationLecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15
Lecture 6 Stereo Systems Multi- view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-24-Jan-15 Lecture 6 Stereo Systems Multi- view geometry Stereo systems
More informationRectification and Distortion Correction
Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification
More informationVision par ordinateur
Epipolar geometry π Vision par ordinateur Underlying structure in set of matches for rigid scenes l T 1 l 2 C1 m1 l1 e1 M L2 L1 e2 Géométrie épipolaire Fundamental matrix (x rank 2 matrix) m2 C2 l2 Frédéric
More information1 Projective Geometry
CIS8, Machine Perception Review Problem - SPRING 26 Instructions. All coordinate systems are right handed. Projective Geometry Figure : Facade rectification. I took an image of a rectangular object, and
More information3D reconstruction class 11
3D reconstruction class 11 Multiple View Geometry Comp 290-089 Marc Pollefeys Multiple View Geometry course schedule (subject to change) Jan. 7, 9 Intro & motivation Projective 2D Geometry Jan. 14, 16
More informationTriangulation from Two Views Revisited: Hartley-Sturm vs. Optimal Correction
Triangulation from Two Views Revisited: Hartley-Sturm vs. Optimal Correction Kenichi Kanatani 1, Yasuyuki Sugaya 2, and Hirotaka Niitsuma 1 1 Department of Computer Science, Okayama University, Okayama
More informationMathematics of a Multiple Omni-Directional System
Mathematics of a Multiple Omni-Directional System A. Torii A. Sugimoto A. Imiya, School of Science and National Institute of Institute of Media and Technology, Informatics, Information Technology, Chiba
More informationMetric Structure from Motion
CS443 Final Project Metric Structure from Motion Peng Cheng 1 Objective of the Project Given: 1. A static object with n feature points and unknown shape. 2. A camera with unknown intrinsic parameters takes
More informationEuclidean Reconstruction Independent on Camera Intrinsic Parameters
Euclidean Reconstruction Independent on Camera Intrinsic Parameters Ezio MALIS I.N.R.I.A. Sophia-Antipolis, FRANCE Adrien BARTOLI INRIA Rhone-Alpes, FRANCE Abstract bundle adjustment techniques for Euclidean
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationPlanar pattern for automatic camera calibration
Planar pattern for automatic camera calibration Beiwei Zhang Y. F. Li City University of Hong Kong Department of Manufacturing Engineering and Engineering Management Kowloon, Hong Kong Fu-Chao Wu Institute
More informationVisual Odometry for Non-Overlapping Views Using Second-Order Cone Programming
Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Jae-Hak Kim 1, Richard Hartley 1, Jan-Michael Frahm 2 and Marc Pollefeys 2 1 Research School of Information Sciences and Engineering
More informationChapter 7: Computation of the Camera Matrix P
Chapter 7: Computation of the Camera Matrix P Arco Nederveen Eagle Vision March 18, 2008 Arco Nederveen (Eagle Vision) The Camera Matrix P March 18, 2008 1 / 25 1 Chapter 7: Computation of the camera Matrix
More informationMinimal Projective Reconstruction for Combinations of Points and Lines in Three Views
Minimal Projective Reconstruction for Combinations of Points and Lines in Three Views Magnus Oskarsson, Andrew Zisserman and Kalle Åström Centre for Mathematical Sciences Lund University,SE 221 00 Lund,
More informationLecture 3: Camera Calibration, DLT, SVD
Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P
More informationEpipolar Geometry and the Essential Matrix
Epipolar Geometry and the Essential Matrix Carlo Tomasi The epipolar geometry of a pair of cameras expresses the fundamental relationship between any two corresponding points in the two image planes, and
More informationCALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang
CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS Cha Zhang and Zhengyou Zhang Communication and Collaboration Systems Group, Microsoft Research {chazhang, zhang}@microsoft.com ABSTRACT
More informationThe Geometry of Dynamic Scenes On Coplanar and Convergent Linear Motions Embedded in 3D Static Scenes
EXTENDED VERSION SHORT VERSION APPEARED IN THE 13TH BMVC, CARDIFF, SEPTEMBER 2002. The Geometry of Dynamic Scenes On Coplanar and Convergent Linear Motions Embedded in 3D Static Scenes Adrien Bartoli,
More informationVisual Recognition: Image Formation
Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know
More informationA Canonical Framework for Sequences of Images
A Canonical Framework for Sequences of Images Anders Heyden, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: andersp@maths.lth.se kalle@maths.lth.se Abstract This
More informationMei Han Takeo Kanade. January Carnegie Mellon University. Pittsburgh, PA Abstract
Scene Reconstruction from Multiple Uncalibrated Views Mei Han Takeo Kanade January 000 CMU-RI-TR-00-09 The Robotics Institute Carnegie Mellon University Pittsburgh, PA 1513 Abstract We describe a factorization-based
More informationMETRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS
METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires
More informationLinear Multi-View Reconstruction of Points, Lines, Planes and Cameras using a Reference Plane
Linear Multi-View Reconstruction of Points, Lines, Planes and Cameras using a Reference Plane Carsten Rother Computational Vision and Active Perception Laboratory (CVAP) Royal Institute of Technology (KTH),
More informationWeek 2: Two-View Geometry. Padua Summer 08 Frank Dellaert
Week 2: Two-View Geometry Padua Summer 08 Frank Dellaert Mosaicking Outline 2D Transformation Hierarchy RANSAC Triangulation of 3D Points Cameras Triangulation via SVD Automatic Correspondence Essential
More informationTHE TRIFOCAL TENSOR AND ITS APPLICATIONS IN AUGMENTED REALITY
THE TRIFOCAL TENSOR AND ITS APPLICATIONS IN AUGMENTED REALITY Jia Li A Thesis submitted to the Faculty of Graduate and Postdoctoral Studies in partial fulfillment of the requirements for the degree of
More informationFactorization Method Using Interpolated Feature Tracking via Projective Geometry
Factorization Method Using Interpolated Feature Tracking via Projective Geometry Hideo Saito, Shigeharu Kamijima Department of Information and Computer Science, Keio University Yokohama-City, 223-8522,
More informationCOMP 558 lecture 19 Nov. 17, 2010
COMP 558 lecture 9 Nov. 7, 2 Camera calibration To estimate the geometry of 3D scenes, it helps to know the camera parameters, both external and internal. The problem of finding all these parameters is
More informationCombining Two-view Constraints For Motion Estimation
ombining Two-view onstraints For Motion Estimation Venu Madhav Govindu Somewhere in India venu@narmada.org Abstract In this paper we describe two methods for estimating the motion parameters of an image
More informationCS231A Midterm Review. Friday 5/6/2016
CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric
More informationMAPI Computer Vision. Multiple View Geometry
MAPI Computer Vision Multiple View Geometry Geometry o Multiple Views 2- and 3- view geometry p p Kpˆ [ K R t]p Geometry o Multiple Views 2- and 3- view geometry Epipolar Geometry The epipolar geometry
More informationC280, Computer Vision
C280, Computer Vision Prof. Trevor Darrell trevor@eecs.berkeley.edu Lecture 11: Structure from Motion Roadmap Previous: Image formation, filtering, local features, (Texture) Tues: Feature-based Alignment
More information55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence
More informationarxiv: v1 [cs.cv] 2 May 2016
16-811 Math Fundamentals for Robotics Comparison of Optimization Methods in Optical Flow Estimation Final Report, Fall 2015 arxiv:1605.00572v1 [cs.cv] 2 May 2016 Contents Noranart Vesdapunt Master of Computer
More informationHW 1: Project Report (Camera Calibration)
HW 1: Project Report (Camera Calibration) ABHISHEK KUMAR (abhik@sci.utah.edu) 1 Problem The problem is to calibrate a camera for a fixed focal length using two orthogonal checkerboard planes, and to find
More informationTowards the completion of assignment 1
Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?
More informationAffine-invariant shape matching and recognition under partial occlusion
Title Affine-invariant shape matching and recognition under partial occlusion Author(s) Mai, F; Chang, CQ; Hung, YS Citation The 17th IEEE International Conference on Image Processing (ICIP 2010), Hong
More informationIdentifying Car Model from Photographs
Identifying Car Model from Photographs Fine grained Classification using 3D Reconstruction and 3D Shape Registration Xinheng Li davidxli@stanford.edu Abstract Fine grained classification from photographs
More informationA Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space
A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-4, Umezono, Tsukuba Ibaraki, 35-8568 Japan ichimura@etl.go.jp
More informationComments on Consistent Depth Maps Recovery from a Video Sequence
Comments on Consistent Depth Maps Recovery from a Video Sequence N.P. van der Aa D.S. Grootendorst B.F. Böggemann R.T. Tan Technical Report UU-CS-2011-014 May 2011 Department of Information and Computing
More informationCamera Model and Calibration
Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the
More informationLecture 6 Stereo Systems Multi-view geometry
Lecture 6 Stereo Systems Multi-view geometry Professor Silvio Savarese Computational Vision and Geometry Lab Silvio Savarese Lecture 6-5-Feb-4 Lecture 6 Stereo Systems Multi-view geometry Stereo systems
More informationMotion Segmentation Based on Factorization Method and Discriminant Criterion
Motion Segmentation Based on Factorization Method and Discriminant Criterion Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-, Umezono, Tsukuba Ibaraki, -6 Japan ichimura@etl.go.jp Abstract A motion segmentation
More informationComputer Vision Projective Geometry and Calibration. Pinhole cameras
Computer Vision Projective Geometry and Calibration Professor Hager http://www.cs.jhu.edu/~hager Jason Corso http://www.cs.jhu.edu/~jcorso. Pinhole cameras Abstract camera model - box with a small hole
More informationStructure from Motion. Introduction to Computer Vision CSE 152 Lecture 10
Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley
More informationMulti-View AAM Fitting and Camera Calibration
To appear in the IEEE International Conference on Computer Vision Multi-View AAM Fitting and Camera Calibration Seth Koterba, Simon Baker, Iain Matthews, Changbo Hu, Jing Xiao, Jeffrey Cohn, and Takeo
More informationFactorization Methods for Projective Structure and Motion
Factorization Methods for Projective Structure and Motion Bill Triggs LIFIA, INRIA Rhône-Alpes, 6, avenue Félix Viallet, 383 Grenoble, France. Bill.Triggs@imag.fr Abstract This paper describes a family
More informationECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt
ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with
More informationAccurate and Dense Wide-Baseline Stereo Matching Using SW-POC
Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp
More informationCamera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds
Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds Marta Wilczkowiak Edmond Boyer Peter Sturm Movi Gravir Inria Rhône-Alpes, 655 Avenue de l Europe, 3833 Montbonnot, France
More informationAn Overview of Matchmoving using Structure from Motion Methods
An Overview of Matchmoving using Structure from Motion Methods Kamyar Haji Allahverdi Pour Department of Computer Engineering Sharif University of Technology Tehran, Iran Email: allahverdi@ce.sharif.edu
More informationToday. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography
Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry
More informationComputer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman Assignment #1. (Due date: 10/23/2012) x P. = z
Computer Vision I Name : CSE 252A, Fall 202 Student ID : David Kriegman E-Mail : Assignment (Due date: 0/23/202). Perspective Projection [2pts] Consider a perspective projection where a point = z y x P
More informationStructure from Motion. Prof. Marco Marcon
Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)
More informationURBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES
URBAN STRUCTURE ESTIMATION USING PARALLEL AND ORTHOGONAL LINES An Undergraduate Research Scholars Thesis by RUI LIU Submitted to Honors and Undergraduate Research Texas A&M University in partial fulfillment
More information