Mosaics Construction from a Sparse Set of Views

Size: px
Start display at page:

Download "Mosaics Construction from a Sparse Set of Views"

Transcription

1 Mosaics Construction from a Sparse Set of Views W. Zhang, J. kosecka and F. Li Abstract In this paper we describe a flexible approach for constructing mosaics of architectural environments from a sparse set of uncalibrated views. The main contribution this paper is the use of environment constraints in order increase the efficiency and level of automation of the mosaic construction process. The observation that in architectural environments, the majority of lines is aligned with the principal orthogonal directions of the world coordinate frame, will be exploited in different stages of the mosaic construction pipeline. The automated detection of vanishing directions will enable us to partially calibrate the camera an estimate the relative orientation of the camera with respect to the scene from a single view. These initial estimates will facilitate efficient feature matching, computation of displacements between the views as well as alignment of multiple views. While the approach described here will be presented in the context of rotational mosaics, the alignment and matching techniques are applicable for general displacements, where the constraints of man-made environments are present and the displacement between the views is large. Key words: panoramic mosaic construction, vanishing point estimation, relative orientation, partial calibration using vanishing points. This work is supported by the George Mason University Research Initiative fund, {kosecka,wzhang2,fli}@cs.gmu.edu 1 Introduction The problem of construction of visually realistic models of the surrounding environments is actively being pursued by various researchers in computer vision, computer graphics and photogrammetry. One of the widely used approaches for acquisition of such models are so called image based rendering techniques. The philosophy of image based rendering techniques is to create realistic models of the environments from a set of photographs or a video stream, since these media capture both the geometry and the appearance of the environment. The popularity of image based rendering techniques partly lies in a broad dissemination of digital (still and video) cameras, ease of data acquisition as well as advances in the understanding of the geometric relationship between multiple views in an uncalibrated setting and associated algorithms. Given multiple views of the scene the approaches typically vary in the type of models they attempt to capture, the amount of 3D information represented in the model, the level of automation in the model acquisition process, model complexity and capability to create new views. The techniques which strive for the full 3D model acquisition enable creation of arbitrary novel views [12]. While it is possible to recover the relative camera displacements and camera calibration information from multiple views, in order to obtain photorealistic models some level of assistance it is often required. This is either at the level of choice of the model (planar surfaces and associated textures maps) or during the image matching and model instantiation process [11, 3]. Alternative representations, which sample the space of all possible appearances of the scene often require storage of large amounts of data [13, 8] and are more applicable for object level models. Alternative image-based representations of environments are so called image mosaics. In case of mosaics the capabilities of generating novel views of the scene are limited, the mosaics are however superior when it comes to ease of their acquisition. Mosaics have been used in a variety of applications including collection of photographs from aerial and satellite imagery, for video indexing, virtual reality and were also proposed as means of representing visual scenes. Comprehensive review as well as motivation and 1

2 associated algorithms can be found in [15]. Most commonly used are so called panoramic or spherical mosaics, which capture the complete 360 o view or a full spherical view of the surrounding environment. In such case the images are obtained from cameras mounted on tripods, where individual views are related by pure rotation. Some recent recent efforts focus on enriching the type mosaic representation to the case of general displacements [16] as well as improving the level of automation and efficiency of the existing techniques. The efficiency and level of automation of the mosaic construction process is mostly determined by the algorithms used for matching and alignment of neighboring views as well as availability of the knowledge of the camera intrinsic parameters. It is often assumed that the sequence of closely separated images (e.g. as acquired by video camera) is available [15] and that intrinsic parameters of the camera are partially known. In such case the individual views have relatively large overlap and are typically aligned using differential motion models (e.g. pure translation, affine models, pure rotation), which are estimated by iterative techniques [16]. In case of large displacements multi-scale representations are adopted and the iterative alignment and warping are interleaved across different levels of the pyramid. Once the views are aligned, and the camera internal parameters are known, the rotations between individual views can be calculated and are followed by the the mapping of each pixel into spherical or cylindrical coordinates. As the displacement between individual views becomes larger, the alignment techniques based of differential motion models often exhibit difficulties of converging to local minima or require large number of iterations. The flexibility of automatic mosaic construction can be greatly enhanced by developing techniques which enable mosaicing from a sparse set of uncalibrated views. In the sequel we will describe such approach by combining environment constraints and single view analysis in order to yield efficient matching and multi-view alignment of a sparse set of views. Alternative techniques for matching across widely separated views using affine invariants associated with textured regions was proposed in [14]. Paper Outline The goal of this paper is to demonstrate how the qualitative knowledge of the environment can increase the flexibility and efficiency of the image alignment process and mosaic acquisition. We will demonstrate how can the constraints of man-made environments be utilized for partial camera calibration and image matching across widely separated views. The main premise of the approach is the presence of sets of parallel and orthogonal lines and planes aligned with principal orthogonal directions of the world coordinate frame. We briefly review an efficient technique for simultaneous estimation and grouping of the detected lines into dominant vanishing directions and demonstrate how to use this information for partial calibration and estimation of camera orientation with respect to the environment from a single view. The partial calibration and orientation information is then used for guided image matching between widely separated views and estimation and refinement of planar homographies between adjacent views. In this paper we will demonstrate applicability of these techniques for construction of panoramic mosaics. The proposed approach can be applied for the alignment of views related by general displacements. There is a large body of work related to individual steps of our approach. We will point out the differences and commonalities with our approach along the way. 2 Single View Analysis Recent efforts in building large city models as well as basic surveillance and monitoring applications often encounter the alignment problem of registering current view to the model or registering widely separate views. Single view analysis can be very instrumental in providing some information about position of the camera and camera intrinsic parameters. The structural regularities of man-made environments, such as presence of sets of parallel and orthogonal lines and planes can be exploited towards determining the relative orientation of the camera with respect to the scene using the information about vanishing points and vanishing lines. The problem of vanishing point detection and estimation have been addressed numerous times in the past and comprehensive review can be found in more recent publications on the topic [1]. The geometric constraints imposed by vanishing directions on the camera intrinsic parameters and camera rotation as well as associated estimation techniques are well understood and have been used previously in the context of structure and motion recovery problems in the uncalibrated case [4]. Once the detected line segments are grouped into common vanishing directions, the MAP estimates of vanishing points can be obtained by minimizing the distance of the line end points from the estimated line segments leading to a nonlinear optimization problem [4]. An alternative to the nonlinear minimization is a covariance weighted linear least squares formulation suggested first in [6], which tries to minimize the algebraic errors. Vanishing points detection The line segments parallel in the 3D world intersect in the image in the vanishing point. Depending on the line orientation the vanishing point can be finite or infinite. The grouping of the line segments into vanishing directions has been often considered separately from the geometric estimation 2

3 problems, or it has been studied in the case of calibrated camera [1]. Consider the perspective camera projection model, where 3D coordinates of points X =[X, Y, Z, 1] T are related to their image projections x =[x, y, 1] T in a following way: λx = PgX where P =[I 3 3, 0] R 3 4 is the projection matrix, g = (R, T ) SE(3) is a rigid body transformation represented by 4 4 matrix using homogeneous coordinates and λ is the unknown scale corresponding to the depth Z of the point X. Given two image points x 1 and x 2, the line segment passing through the two endpoints is represented by a plane normal of a plane passing through the center of projection and intersecting the retinal plane in a line l, such that l = x 1 x 2 = x 1 x 1 2. The unit vectors corresponding to the plane normals l i can viewed as points on a unit sphere. The vectors l i corresponding to the parallel lines in 3D world all lie in some plane, whose intersection with the Gaussian sphere forms a great circle. The vanishing direction v then corresponds to the plane normal where all these lines lie and in the noise free case l T i v =0. Given two lines the common normal is determined by v = l 1 l 2 = l 1 l 2. Hence given a set of line segments belonging to the lines parallel in 3D, the common vanishing direction v can be obtained by solving the following linear least squares estimation problem: min v n (l T i v) 2 i=1 This corresponds to min v Av 2, where the rows of matrix A R n 3 are the lines segments l i. This particular least squares estimation problem has be studied in [2] assuming the unit vectors on the sphere are distributed according to Binghman distribution. The optimal solution to this type of orthogonal least squares problems is also described in [6]. Uncalibrated camera In order to be able to determine and adjust along the way the number of groups present in the image some notion of a distance between the line and vanishing direction or two vanishing directions is necessary. In the calibrated setting the angle between two directions is represented by the inner product between two vectors u T v with u, v R 3. In the case of uncalibrated camera the image coordinates undergo an additional transformation K which depends on the 2 5 is a skew symmetric matrix associ- 0 x 3 x 2 1 bx = 4 x 3 0 x 1 x 2 x 1 0 ated with x =[x 1,x 2,x 3 ] T. 3 internal camera parameters: x = Kx with K = α x α θ o x 0 α y o y = f α θ o x 0 kf o y where f is the focal length of the camera in pixel units, k is the aspect ratio and [o x,o y, 1] T is the principal point of the camera. In the uncalibrated setting the vectors undergo additional transformation K and we have u = Ku and v = Kv and the inner product u T v becomes: u T v = u T K T K 1 v where the unknown matrix S = K T K 1 can be interpreted as a metric of the uncalibrated space. In the following we will demonstrate that by transforming the measurements by an arbitrary nonsingular transformation A has no effect on the computation of the vanishing points. In order to proceed note that the following fact holds: If v R 3 and A SL(3),thenA T va = Â 1 v. Suppose that the endpoints of two lines are x 1, x 2 and x 3, x 4, such that l 1 = x 1 x 2 and l 2 = x 3 x 4,where the measurements x i = Ax i are related to the calibrated image coordinates by some unknown nonsingular transformation A. We can to show that the vanishing point v corresponding to the plane normal spanned by vectors l 1 and l 2 ; v = l 1 l 2 is related to the actual vanishing direction in the original space by the unknown transformation A, namely v = Av. Hence we have: v = l 1 l 2 =(Âx 1Ax 2 ) (Âx 3Ax 4 ) = (A T x 1 x 2 ) (A T x 3 x 4 )= (1) = (A T l 1 ) (A T l 2 )=A l 1 l 2 = Av (2) The above fact demonstrates that in the context of vanishing point estimation, transforming the image measurements by an arbitrary nonsingular transformation A andthentransforming the result back, does not affect the final estimate. We will use this fact in the normalization step of the least squares estimation in the context of EM algorithm. In the case of man-made environments we will exploit the fact that the dominant vanishing directions are aligned with the principal orthogonal axes e i,e j,e k of the world reference frame. Similarly as in [1] we address the grouping stage and vanishing point estimation stage simultaneously as a problem of probabilistic inference with an unknown model. We assume however that the camera is not calibrated. In such instances the algorithm of choice is the Expectation Maximization algorithm (EM), which estimates the coordinates of vanishing points as well as the probabilities of individual line segments belonging to a particular vanishing directions. We will demonstrate that with proper 3

4 normalization, the simultaneous grouping and estimation of the vanishing points using EM can be accomplished in the case of an uncalibrated camera. The posterior distribution of the vanishing points given line segments can be expressed using Bayes rule in terms of the conditional distribution and prior probability of the vanishing points: p(v k l i )= p(l i v k )p(v k ) (3) p(l i ) where p(l i v k ) is the likelihood of the line segment belonging to a particular vanishing direction v k. Hence for a particular line segment, p(l i ) can be expressed using the conditional mixture model representation: p(l i )= m p(v k )p(l i v k ) (4) k=1 The number of possible vanishing directions m, will vary depending on the image. We assume that there are at most four significant models, three corresponding to the dominant vanishing directions and an additional one modeling the outlier process. The choice of the likelihood term p(l i v k ) depends on the form of the objective being minimized as well as the error model. In the noise free case we have l T i v k =0. In the case of noisy measurements we assume that the error ξ i in l T i v k = ξ i is a normally distributed random variable with N(0,σ1 2 ). Then the likelihood term is given as: ( (l T p(l i v k ) exp i v k ) 2 ) (5) 2σ 2 1 Given initial estimates of the vanishing points v k, the membership probabilities of a line segment l i belonging to the k-th vanishing direction are computed in the following way: p(v k l i )= p(l i v k )p(v k ) m k=1 p(l i v k )p(v k ) The posterior probability terms p(v k l i ) represent so called membership probabilities, denoted by w ik and capture the probability of a line segment l i belonging to k-th vanishing direction v k. Initially we assume that the prior probabilities of all vanishing directions are equally likely and hence do not affect the posterior conditional probability. The prior probabilities of the vanishing directions can be estimated from the likelihoods and can affect favorably the convergence process as demonstrated in [1]. In the following paragraph we describe the main ingredients of the EM algorithm for simultaneous grouping and estimation of the vanishing directions. Prior to the estimation of vanishing points and grouping of the line segments into common vanishing directions, we first transform all the endpoint measurements by K 1 ; x = K 1 x 1. The transformation K (6) will make the line segments and the vanishing directions well separated on the unit sphere (as in the calibrated setting). Given an image of size [dx, dy] T the transformation K is simply related to the size of the image where f = dx, õ x = dx/2 and õ y = dy/2 and has the following form: K = f 0 õ x 0 f õy The E-step of the EM algorithm amounts to computation of posterior probabilities p(v k l i ) given the currently available vanishing points estimates. The M-step of the algorithm involves maximization of the expected complete log likelihood with respect to the unknown parameters v k [10]. This step yields a maximization of the following objective function: max v k n p(l i )= i=1 n log p(l i ) (7) i=1 where p(l i v k ) is the likelihood term defined in equation (5). The above objective function in the case of linear log likelihood model yields a solution to a weighted least squares problem; one for each model. Each line has an associated weight w ik determined by posterior probabilities computed in the E step. In such case the vanishing points are estimated by solving the following linear least-squares problem: J(v k ) = min v k w ik (l T i v k ) 2 i =min v k WAv k ) 2 (8) Where W R n n is a diagonal matrix associated with the weights and rows of A R 3 n are the detected line segments. Closed form solution corresponds to the eigenvector associated with the smallest eigenvalue of A T W T WAand yields the new estimate of v k. EM algorithm is an iterative technique guaranteed to increase the likelihood of the available measurements. The iterations of the EM algorithm are depicted in Figure 1. The initially large number of vanishing point estimates, is reduced through the merging process to three dominant directions and the process usually converges in 3-5 iterations. The EM process is efficiently initialized from orientation histograms of the detected line segments, with no need for costly computation of the vanishing points using Hough Transforms. For more technical details of the algorithm see [7]. 2.1 Partial Calibration from Single View The constraints among detected vanishing directions can be used for partial self-calibration. In the uncalibrated case 4

5 When three finite vanishing points are detected, they provide three independent constraints on matrix S: vi T Sv j = 0 vi T Sv k = 0 vj T Sv k = 0 (12) Figure 1. The iterations of the EM algorithm and the detected vanishing points. The lines aligned with the principal orthogonal directions are color coded and the final finite estimates of vanishing points are denoted by o. the relationship between image coordinates of a point and its 3D counterpart is as follows: λx = RX + T Multiplying both sides by calibration matrix K we have: λx = KRX + KT (9) where x denotes a pixel coordinate of X. Let s denote the unit vectors associated with the world coordinate frame to be: e i =[1, 0, 0] T, e j =[0, 1, 0] T, e k =[0, 1, 0] T. The vanishing points corresponding to 3D lines parallel to either of these directions are: v i = KRe i v j = KRe j v k = KRe k and note that the coordinates of vanishing points depend only on rotation and internal parameters of the camera. The orthogonality relations between e i,e j,e k readily provide constraints on the calibration matrix K. In particular we have: e T i e j = vi T K T RR T K 1 v j (10) = vi T K T K 1 v j = vi T Sv j (11) where S is the metric associated with the uncalibrated camera introduced earlier: S = K T K 1 = s 1 s 2 s 3 s 2 s 4 s 5 s 3 s 5 s 6 In general symmetric matrix S 3 3 has six degrees of freedom and can be recovered up to a scale, so without additional constraints we can recover the S only up to two parameter family. Other commonly assumed assumption of zero skew and known aspect ratio can also be expressed in terms of constraints on the metric S as proposed earlier in [9]. The zero skew constraint expresses the fact that the image axes are orthogonal can be written as: [1, 0, 0]S[0, 1, 0] T =0 In the presence of zero skew assumption, the known aspect ratio constraint can be expressed as s 1 = s 4. With these two additional constraints we have a sufficient number of constraints and the solution for s =[s 1,s 2,s 3,s 4,s 5,s 6 ] T can be obtained by solving a linear least squares estimation problem. Writing the individual constraints as b T j s = 0 and stacking them into matrix B, s can be obtained by minimizing Bs 2 and corresponds to the eigenvector associated with the smallest eigenvalue of B T B. The calibration matrix K 1 can be obtained from S by Cholesky decomposition. In the case one of the vanishing directions lies close to infinity one of the constraints becomes degenerate and recovered S fails to be positive definite. This situation can be also noticed by checking the condition number of B. In such case we assume that the principal point lies in the center of the image and hence S is parameterized by the focal length only as: S = 1/f /f In such case the focal length can be recovered in a closed form [5] from the remaining constraint vj T Sv k =0. The recovered calibration matrices for the examples outlined in Figure 1 are below: K building = K room = (13) (14) Note that in the above examples the difference in the focal length is due to the difference in the image size. While the 5

6 subsampling affects also the position of the principal point, the above statement assumes that the focal length of the subsampled images is related to the original focal length by the subsampling factor. The quality of the estimates depends on the accuracy of the estimated vanishing points. As the vanishing points approach infinity their estimates become less accurate. This affects in particular the estimate of the principal point, which in case one of the vanishing points is at infinity cannot be uniquely determined unless additional constraints are introduced [9]. In such case we assume that the principal point of the camera lies in the center of the image and estimate the focal length in the closed form, using a single orthogonality constraint between vanishing directions. The estimate of the focal length obtained in this manner is less accurate then if all the constraints are used simultaneously (if available) and the principal point is estimated as well. Figure 2. Original set of correspondences, with putative matches found by from homography computed from relative orientations of two single views Ĥij. Relative orientation Once the vanishing points have been detected and the unknown camera parameters determined by the above procedure, the relative orientation of the camera with respect to the scene can be computed. Note first that since the vanishing directions are projections of the vectors associated with three orthogonal directions i, j, k and depend on rotation only. In particular we can write that: Figure 3. The matched correspondences (left) and detected outliers (right) after RANSAC iterations. Outliers are denoted in red. K 1 v i = Re i K 1 v j = Re j K 1 v k = Re k with each vanishing direction being proportional to the column of the rotation matrix R = [r 1, r 2, r 3 ]. Choosing the two best vanishing directions and properly normalizing them, the third row can be obtained by enforcing the orthogonality constraints as r 3 = r 1 r 2. There is a four way ambiguity in R due to the sign ambiguity in r 1 and r 2. Spurious solutions can be eliminated by considering relative orientation or structure constraints. 2.2 Image alignment The techniques described in the previous paragraph were limited to the single view analysis and enabled us to recover partial calibration of the camera as well as orientation of the camera with respect to the scene. In the case of two uncalibrated views related by rotation only, the image coordinates of corresponding points in two views satisfy: x j Hx i (15) where H = KRK 1 R 3 3 is so called homography with the plane at infinity sometimes denoted as H and x i =[x i,y i, 1] T and x j =[x j,y j, 1] T are the image coordinates of the corresponding points. When the orientation and partial calibration information are estimated from a single view, as described in previous section, the homography between two views can be computed and becomes: Ĥ ij = KR i R T j K 1 Together with the knowledge of the camera calibration matrix K building from equation 13 the rotation angle and the axis can be recovered from the rotation matrix R i Rj T, yielding in this case a rotation of φ ij =14.7 o around y axis. The quality of H ij estimate is often not satisfactory for accurate alignment of the views. However the estimate of H ij can serve as a good starting point for guided feature matching and re-estimation of the homography. Note that in order to estimate H in equation 15 at least 4 corresponding points in two views are needed each providing two constraints on H. In the first step of the refinement we detect additional features from each view using Harris corner detector with the results in Figure 2. Traditional robust techniques for estimation of H require at least 4 corresponding points, in order to bootstrap the robust matching algorithm, such as RANSAC. In the case of large displacements the initialization step however can be quite expensive. Without any prior knowledge about the motion potentially large regions in the image have to be searched for possible matches. Given that 6

7 Figure 4. Image alignment using the refined homography estimates. we can approximately estimate the homography based on single view analysis, we can select candidate feature points in one view and use Ĥij to predict their location in another view. The predicted location of the feature then limits the neighborhood search to few pixels around the predicted location. The displacement was over 300 pixels in the horizontal direction which would make the search for putative correspondences very consuming and the differential techniques would require a large number of iterations in this case. Examples of the initial and refined estimates of the homographies between two views are: H ij = H = (16) If we assume H is the reference homography then the error between the two estimates H Hij H 100 is 34%. Applying the homography refinement and matching techniques to all views we can obtain refined homography estimates, which will enable us the to align all the views to the common reference plane 4. The above alignment can be also achieved without any knowledge of the intrinsic parameters of the camera, using computationally more expensive image matching techniques. Once the homographies between the reference view and additional views have been computed the intrinsic parameters of the camera can be reestimated. In the case of panoramic mosaics, where the views are related by the rotation around single independent axis the matrix S can be determined only up to two parameter family and hence all the intrinsic parameters of the camera cannot be estimated. In the case of panoramic mosaics we assume that the skew is zero, the aspect ratio of the camera is one (known) and the principal point is in the image center. In such case the only unknown is the focal Figure 5. Final mosaic obtained with the refined estimate of the focal length yield correct mapping to the cylindrical coordinates. length f, which can be computed in closed form exploiting properties of rotation matrices. H = h 0 h 1 h 2 h 3 h 4 h 5 r 11 r 12 r 13 /f r 21 r 22 r 23 /f (17) h 6 h 7 1 r 31 r 32 r 33 /f The fact the any two rows (or columns) of the scaled rotation matrix must have the same norm and be orthogonal gives us constraints for computing f. In order to obtain accurate estimates of rotation and camera intrinsic parameters, spherical mosaics are favorable, since they are composed from multiple views related by rotations of around more then two independent axis. In such case the complete intrinsic parameters can be determined and associated rotation matrices computed [17]. For the mosaic construction accurate focal length parameters are necessary in order to establish the mapping between image coordinates [x, y, 1] T and cylindrical coordinates: θ = tan 1 ( X Z )=tan 1 ( x f ) (18) v = Y X2 + Z 2 = y x2 + f 2 (19) where the radius of the cylinder is the focal lenght. The final mosaic obtained from the estimates of homographies and global refinement of the focal length is depicted in figure 5. 3 Conclusions We presented an efficient, completely automated approach for construction of rotational mosaics from an uncalibrated sparse set of views. Along the way the assumptions about the structure of man-made environments, were used towards efficient initialization and grouping of line segments into dominant vanishing directions aligned with the axes of the world coordinate frame. The estimation and grouping problems for vanishing point estimation were addressed simultaneously using the Expectation Maximization algorithm. The single view analysis was instrumental 7

8 for obtaining initial estimates of homographies as well as intrinsic parameters of the camera. These initial estimates enabled us to do efficient matching and final homography computation and view alignment for the case of largely separated views. The presented approach extends currently available techniques for mosaic construction and demonstrates how the constraints of architectural environments can be used efficiently towards this task. References [1] M. Antone and S. Teller. Automatic recovery of relative camera rotations for urban scenes. In IEEE Proceedings of CVPR, [2] R. Collins. Vanishing point calculation as statistical inference on the unit sphere. In Proceedings of International Conference on Computer Vision, pages , [3] Paul E. Debevec, C.J.Taylor, and Jitendra Malik. Modeling and rendering architecture from photographs:a hybrid geometry-and image-based approach. In SIGRAPH, [12] M. Pollefeys R. Koch and L. Van Gool. Multi viewpoint stereo from uncalibrated video sequences. In Springer Verlag, editor, European Conference on Computer Vision, [13] R. Szeliski S. Gotler, R. Grzeszczuk and M. Cohen. The lumingraph. In SIGRAPH, [14] F. Schaffalitzky and A. Zisserman. Viewpoint invariant texture matching and wide baseline stereo. In International Conference on Computer Vision, Vancouver, Canada, [15] R. Szeliski. Video mosaics for virtual environments. In IEEE Computer Graphics and Applications, [16] R. Szeliski and H. Shum. Creating full view panoramic image mosaics and environment maps. In Microsoft Research Technical Report, [17] J. Košecka Y. Ma, S. Soatto and S. Sastry. Reconstruction and reprojection up to subgroups. International Journal of Computer Vision, 38(3):218:29, [4] R. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge University Press, [5] D. Jelinek and C.J. Taylor. Reconstruction of linearly parametrized models from single images with camera of unknown focal lenght. IEEE Transactions of PAMI, pages , July [6] K. Kanatani. Geometric Computation for Machine Vision. Oxford Science Publications, [7] J. Košecka and W. Zhang. Video compass. Technical report, George Mason University, [8] McMillan L. and Bishop G. Plenoptic modellng. an image based rendering system. In In Computer Graphics Annual Conference Series, [9] D. Liebowitz. Combining scene and auto-calibration constraints. In Proceedings of ICCV, [10] G. J. McLachlan and K. E. Basford. Mixture Models: Inference and Applications. Marcel Dekker Inc., N.Y., [11] D.P. Robertson R. Cipolla and E.G. Boye. Photobuilder 3d models of architectural scenes from uncalibrated images. In Proc. IEEE International Conference on Multimedia Computing and Systems, Firenze,

Efficient Computation of Vanishing Points

Efficient Computation of Vanishing Points Efficient Computation of Vanishing Points Wei Zhang and Jana Kosecka 1 Department of Computer Science George Mason University, Fairfax, VA 223, USA {wzhang2, kosecka}@cs.gmu.edu Abstract Man-made environments

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Midterm Exam Solutions

Midterm Exam Solutions Midterm Exam Solutions Computer Vision (J. Košecká) October 27, 2009 HONOR SYSTEM: This examination is strictly individual. You are not allowed to talk, discuss, exchange solutions, etc., with other fellow

More information

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka

CS223b Midterm Exam, Computer Vision. Monday February 25th, Winter 2008, Prof. Jana Kosecka CS223b Midterm Exam, Computer Vision Monday February 25th, Winter 2008, Prof. Jana Kosecka Your name email This exam is 8 pages long including cover page. Make sure your exam is not missing any pages.

More information

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006

Camera Registration in a 3D City Model. Min Ding CS294-6 Final Presentation Dec 13, 2006 Camera Registration in a 3D City Model Min Ding CS294-6 Final Presentation Dec 13, 2006 Goal: Reconstruct 3D city model usable for virtual walk- and fly-throughs Virtual reality Urban planning Simulation

More information

Two-View Geometry (Course 23, Lecture D)

Two-View Geometry (Course 23, Lecture D) Two-View Geometry (Course 23, Lecture D) Jana Kosecka Department of Computer Science George Mason University http://www.cs.gmu.edu/~kosecka General Formulation Given two views of the scene recover the

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

Metric Rectification for Perspective Images of Planes

Metric Rectification for Perspective Images of Planes 789139-3 University of California Santa Barbara Department of Electrical and Computer Engineering CS290I Multiple View Geometry in Computer Vision and Computer Graphics Spring 2006 Metric Rectification

More information

Camera Geometry II. COS 429 Princeton University

Camera Geometry II. COS 429 Princeton University Camera Geometry II COS 429 Princeton University Outline Projective geometry Vanishing points Application: camera calibration Application: single-view metrology Epipolar geometry Application: stereo correspondence

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t R 2 3,t 3 Camera 1 Camera

More information

A Factorization Method for Structure from Planar Motion

A Factorization Method for Structure from Planar Motion A Factorization Method for Structure from Planar Motion Jian Li and Rama Chellappa Center for Automation Research (CfAR) and Department of Electrical and Computer Engineering University of Maryland, College

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds

Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds Marta Wilczkowiak Edmond Boyer Peter Sturm Movi Gravir Inria Rhône-Alpes, 655 Avenue de l Europe, 3833 Montbonnot, France

More information

A simple method for interactive 3D reconstruction and camera calibration from a single view

A simple method for interactive 3D reconstruction and camera calibration from a single view A simple method for interactive 3D reconstruction and camera calibration from a single view Akash M Kushal Vikas Bansal Subhashis Banerjee Department of Computer Science and Engineering Indian Institute

More information

3D Reconstruction from Scene Knowledge

3D Reconstruction from Scene Knowledge Multiple-View Reconstruction from Scene Knowledge 3D Reconstruction from Scene Knowledge SYMMETRY & MULTIPLE-VIEW GEOMETRY Fundamental types of symmetry Equivalent views Symmetry based reconstruction MUTIPLE-VIEW

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection

More information

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with

More information

Unit 3 Multiple View Geometry

Unit 3 Multiple View Geometry Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

Compositing a bird's eye view mosaic

Compositing a bird's eye view mosaic Compositing a bird's eye view mosaic Robert Laganiere School of Information Technology and Engineering University of Ottawa Ottawa, Ont KN 6N Abstract This paper describes a method that allows the composition

More information

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches Reminder: Lecture 20: The Eight-Point Algorithm F = -0.00310695-0.0025646 2.96584-0.028094-0.00771621 56.3813 13.1905-29.2007-9999.79 Readings T&V 7.3 and 7.4 Essential/Fundamental Matrix E/F Matrix Summary

More information

Multiple Views Geometry

Multiple Views Geometry Multiple Views Geometry Subhashis Banerjee Dept. Computer Science and Engineering IIT Delhi email: suban@cse.iitd.ac.in January 2, 28 Epipolar geometry Fundamental geometric relationship between two perspective

More information

Computer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman Assignment #1. (Due date: 10/23/2012) x P. = z

Computer Vision I Name : CSE 252A, Fall 2012 Student ID : David Kriegman   Assignment #1. (Due date: 10/23/2012) x P. = z Computer Vision I Name : CSE 252A, Fall 202 Student ID : David Kriegman E-Mail : Assignment (Due date: 0/23/202). Perspective Projection [2pts] Consider a perspective projection where a point = z y x P

More information

Viewpoint Invariant Features from Single Images Using 3D Geometry

Viewpoint Invariant Features from Single Images Using 3D Geometry Viewpoint Invariant Features from Single Images Using 3D Geometry Yanpeng Cao and John McDonald Department of Computer Science National University of Ireland, Maynooth, Ireland {y.cao,johnmcd}@cs.nuim.ie

More information

Camera model and multiple view geometry

Camera model and multiple view geometry Chapter Camera model and multiple view geometry Before discussing how D information can be obtained from images it is important to know how images are formed First the camera model is introduced and then

More information

Computer Vision Projective Geometry and Calibration. Pinhole cameras

Computer Vision Projective Geometry and Calibration. Pinhole cameras Computer Vision Projective Geometry and Calibration Professor Hager http://www.cs.jhu.edu/~hager Jason Corso http://www.cs.jhu.edu/~jcorso. Pinhole cameras Abstract camera model - box with a small hole

More information

Projective geometry for Computer Vision

Projective geometry for Computer Vision Department of Computer Science and Engineering IIT Delhi NIT, Rourkela March 27, 2010 Overview Pin-hole camera Why projective geometry? Reconstruction Computer vision geometry: main problems Correspondence

More information

Two-view geometry Computer Vision Spring 2018, Lecture 10

Two-view geometry Computer Vision Spring 2018, Lecture 10 Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of

More information

Geometric camera models and calibration

Geometric camera models and calibration Geometric camera models and calibration http://graphics.cs.cmu.edu/courses/15-463 15-463, 15-663, 15-862 Computational Photography Fall 2018, Lecture 13 Course announcements Homework 3 is out. - Due October

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection CHAPTER 3 Single-view Geometry When we open an eye or take a photograph, we see only a flattened, two-dimensional projection of the physical underlying scene. The consequences are numerous and startling.

More information

Model Refinement from Planar Parallax

Model Refinement from Planar Parallax Model Refinement from Planar Parallax A. R. Dick R. Cipolla Department of Engineering, University of Cambridge, Cambridge, UK {ard28,cipolla}@eng.cam.ac.uk Abstract This paper presents a system for refining

More information

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry 55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence

More information

Image-based Modeling and Rendering: 8. Image Transformation and Panorama

Image-based Modeling and Rendering: 8. Image Transformation and Panorama Image-based Modeling and Rendering: 8. Image Transformation and Panorama I-Chen Lin, Assistant Professor Dept. of CS, National Chiao Tung Univ, Taiwan Outline Image transformation How to represent the

More information

Motion Tracking and Event Understanding in Video Sequences

Motion Tracking and Event Understanding in Video Sequences Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

1D camera geometry and Its application to circular motion estimation. Creative Commons: Attribution 3.0 Hong Kong License

1D camera geometry and Its application to circular motion estimation. Creative Commons: Attribution 3.0 Hong Kong License Title D camera geometry and Its application to circular motion estimation Author(s Zhang, G; Zhang, H; Wong, KKY Citation The 7th British Machine Vision Conference (BMVC, Edinburgh, U.K., 4-7 September

More information

Rectification and Distortion Correction

Rectification and Distortion Correction Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points

More information

Vision par ordinateur

Vision par ordinateur Epipolar geometry π Vision par ordinateur Underlying structure in set of matches for rigid scenes l T 1 l 2 C1 m1 l1 e1 M L2 L1 e2 Géométrie épipolaire Fundamental matrix (x rank 2 matrix) m2 C2 l2 Frédéric

More information

Stereo Image Rectification for Simple Panoramic Image Generation

Stereo Image Rectification for Simple Panoramic Image Generation Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,

More information

A COMPREHENSIVE TOOL FOR RECOVERING 3D MODELS FROM 2D PHOTOS WITH WIDE BASELINES

A COMPREHENSIVE TOOL FOR RECOVERING 3D MODELS FROM 2D PHOTOS WITH WIDE BASELINES A COMPREHENSIVE TOOL FOR RECOVERING 3D MODELS FROM 2D PHOTOS WITH WIDE BASELINES Yuzhu Lu Shana Smith Virtual Reality Applications Center, Human Computer Interaction Program, Iowa State University, Ames,

More information

Euclidean Reconstruction Independent on Camera Intrinsic Parameters

Euclidean Reconstruction Independent on Camera Intrinsic Parameters Euclidean Reconstruction Independent on Camera Intrinsic Parameters Ezio MALIS I.N.R.I.A. Sophia-Antipolis, FRANCE Adrien BARTOLI INRIA Rhone-Alpes, FRANCE Abstract bundle adjustment techniques for Euclidean

More information

Module 4F12: Computer Vision and Robotics Solutions to Examples Paper 2

Module 4F12: Computer Vision and Robotics Solutions to Examples Paper 2 Engineering Tripos Part IIB FOURTH YEAR Module 4F2: Computer Vision and Robotics Solutions to Examples Paper 2. Perspective projection and vanishing points (a) Consider a line in 3D space, defined in camera-centered

More information

Lecture 3: Camera Calibration, DLT, SVD

Lecture 3: Camera Calibration, DLT, SVD Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P

More information

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert

Week 2: Two-View Geometry. Padua Summer 08 Frank Dellaert Week 2: Two-View Geometry Padua Summer 08 Frank Dellaert Mosaicking Outline 2D Transformation Hierarchy RANSAC Triangulation of 3D Points Cameras Triangulation via SVD Automatic Correspondence Essential

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Literature Survey Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This literature survey compares various methods

More information

Structure from motion

Structure from motion Structure from motion Structure from motion Given a set of corresponding points in two or more images, compute the camera parameters and the 3D point coordinates?? R 1,t 1 R 2,t 2 R 3,t 3 Camera 1 Camera

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

How to Compute the Pose of an Object without a Direct View?

How to Compute the Pose of an Object without a Direct View? How to Compute the Pose of an Object without a Direct View? Peter Sturm and Thomas Bonfort INRIA Rhône-Alpes, 38330 Montbonnot St Martin, France {Peter.Sturm, Thomas.Bonfort}@inrialpes.fr Abstract. We

More information

A Calibration Algorithm for POX-Slits Camera

A Calibration Algorithm for POX-Slits Camera A Calibration Algorithm for POX-Slits Camera N. Martins 1 and H. Araújo 2 1 DEIS, ISEC, Polytechnic Institute of Coimbra, Portugal 2 ISR/DEEC, University of Coimbra, Portugal Abstract Recent developments

More information

Simultaneous Vanishing Point Detection and Camera Calibration from Single Images

Simultaneous Vanishing Point Detection and Camera Calibration from Single Images Simultaneous Vanishing Point Detection and Camera Calibration from Single Images Bo Li, Kun Peng, Xianghua Ying, and Hongbin Zha The Key Lab of Machine Perception (Ministry of Education), Peking University,

More information

A Summary of Projective Geometry

A Summary of Projective Geometry A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies

Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies Feature Transfer and Matching in Disparate Stereo Views through the use of Plane Homographies M. Lourakis, S. Tzurbakis, A. Argyros, S. Orphanoudakis Computer Vision and Robotics Lab (CVRL) Institute of

More information

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1

More information

Factorization with Missing and Noisy Data

Factorization with Missing and Noisy Data Factorization with Missing and Noisy Data Carme Julià, Angel Sappa, Felipe Lumbreras, Joan Serrat, and Antonio López Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona,

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 263 Index 3D reconstruction, 125 5+1-point algorithm, 284 5-point algorithm, 270 7-point algorithm, 265 8-point algorithm, 263 affine point, 45 affine transformation, 57 affine transformation group, 57 affine

More information

Planar homographies. Can we reconstruct another view from one image? vgg/projects/singleview/

Planar homographies. Can we reconstruct another view from one image?   vgg/projects/singleview/ Planar homographies Goal: Introducing 2D Homographies Motivation: What is the relation between a plane in the world and a perspective image of it? Can we reconstruct another view from one image? Readings:

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253

Index. 3D reconstruction, point algorithm, point algorithm, point algorithm, point algorithm, 253 Index 3D reconstruction, 123 5+1-point algorithm, 274 5-point algorithm, 260 7-point algorithm, 255 8-point algorithm, 253 affine point, 43 affine transformation, 55 affine transformation group, 55 affine

More information

Today s lecture. Image Alignment and Stitching. Readings. Motion models

Today s lecture. Image Alignment and Stitching. Readings. Motion models Today s lecture Image Alignment and Stitching Computer Vision CSE576, Spring 2005 Richard Szeliski Image alignment and stitching motion models cylindrical and spherical warping point-based alignment global

More information

Computer Vision I - Appearance-based Matching and Projective Geometry

Computer Vision I - Appearance-based Matching and Projective Geometry Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 05/11/2015 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation

More information

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo --

Computational Optical Imaging - Optique Numerique. -- Multiple View Geometry and Stereo -- Computational Optical Imaging - Optique Numerique -- Multiple View Geometry and Stereo -- Winter 2013 Ivo Ihrke with slides by Thorsten Thormaehlen Feature Detection and Matching Wide-Baseline-Matching

More information

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry

CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from: David Lowe, UBC Jiri Matas, CMP Prague http://cmp.felk.cvut.cz/~matas/papers/presentations/matas_beyondransac_cvprac05.ppt

More information

An Overview of Matchmoving using Structure from Motion Methods

An Overview of Matchmoving using Structure from Motion Methods An Overview of Matchmoving using Structure from Motion Methods Kamyar Haji Allahverdi Pour Department of Computer Engineering Sharif University of Technology Tehran, Iran Email: allahverdi@ce.sharif.edu

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry 55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence

More information

CS231A Midterm Review. Friday 5/6/2016

CS231A Midterm Review. Friday 5/6/2016 CS231A Midterm Review Friday 5/6/2016 Outline General Logistics Camera Models Non-perspective cameras Calibration Single View Metrology Epipolar Geometry Structure from Motion Active Stereo and Volumetric

More information

Visual Recognition: Image Formation

Visual Recognition: Image Formation Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know

More information

Recovering structure from a single view Pinhole perspective projection

Recovering structure from a single view Pinhole perspective projection EPIPOLAR GEOMETRY The slides are from several sources through James Hays (Brown); Silvio Savarese (U. of Michigan); Svetlana Lazebnik (U. Illinois); Bill Freeman and Antonio Torralba (MIT), including their

More information

Multiple View Geometry in Computer Vision Second Edition

Multiple View Geometry in Computer Vision Second Edition Multiple View Geometry in Computer Vision Second Edition Richard Hartley Australian National University, Canberra, Australia Andrew Zisserman University of Oxford, UK CAMBRIDGE UNIVERSITY PRESS Contents

More information

Towards Generic Self-Calibration of Central Cameras

Towards Generic Self-Calibration of Central Cameras Towards Generic Self-Calibration of Central Cameras Srikumar Ramalingam 1&2, Peter Sturm 1, and Suresh K. Lodha 2 1 INRIA Rhône-Alpes, GRAVIR-CNRS, 38330 Montbonnot, France 2 Dept. of Computer Science,

More information

Agenda. Rotations. Camera models. Camera calibration. Homographies

Agenda. Rotations. Camera models. Camera calibration. Homographies Agenda Rotations Camera models Camera calibration Homographies D Rotations R Y = Z r r r r r r r r r Y Z Think of as change of basis where ri = r(i,:) are orthonormal basis vectors r rotated coordinate

More information

AUTO-RECTIFICATION OF USER PHOTOS. Krishnendu Chaudhury, Stephen DiVerdi, Sergey Ioffe.

AUTO-RECTIFICATION OF USER PHOTOS. Krishnendu Chaudhury, Stephen DiVerdi, Sergey Ioffe. AUTO-RECTIFICATION OF USER PHOTOS Krishnendu Chaudhury, Stephen DiVerdi, Sergey Ioffe krish@google.com, diverdi@google.com, sioffe@google.com ABSTRACT The image auto rectification project at Google aims

More information

Rectification for Any Epipolar Geometry

Rectification for Any Epipolar Geometry Rectification for Any Epipolar Geometry Daniel Oram Advanced Interfaces Group Department of Computer Science University of Manchester Mancester, M13, UK oramd@cs.man.ac.uk Abstract This paper proposes

More information

Practical Camera Auto-Calibration Based on Object Appearance and Motion for Traffic Scene Visual Surveillance

Practical Camera Auto-Calibration Based on Object Appearance and Motion for Traffic Scene Visual Surveillance Practical Camera Auto-Calibration Based on Object Appearance and Motion for Traffic Scene Visual Surveillance Zhaoxiang Zhang, Min Li, Kaiqi Huang and Tieniu Tan National Laboratory of Pattern Recognition,

More information

Structure from Motion

Structure from Motion Structure from Motion Outline Bundle Adjustment Ambguities in Reconstruction Affine Factorization Extensions Structure from motion Recover both 3D scene geoemetry and camera positions SLAM: Simultaneous

More information

An Image-Based System for Urban Navigation

An Image-Based System for Urban Navigation An Image-Based System for Urban Navigation Duncan Robertson and Roberto Cipolla Cambridge University Engineering Department Trumpington Street, Cambridge, CB2 1PZ, UK Abstract We describe the prototype

More information

Invariant Features from Interest Point Groups

Invariant Features from Interest Point Groups Invariant Features from Interest Point Groups Matthew Brown and David Lowe {mbrown lowe}@cs.ubc.ca Department of Computer Science, University of British Columbia, Vancouver, Canada. Abstract This paper

More information

Single-view metrology

Single-view metrology Single-view metrology Magritte, Personal Values, 952 Many slides from S. Seitz, D. Hoiem Camera calibration revisited What if world coordinates of reference 3D points are not known? We can use scene features

More information

3D Reconstruction from Two Views

3D Reconstruction from Two Views 3D Reconstruction from Two Views Huy Bui UIUC huybui1@illinois.edu Yiyi Huang UIUC huang85@illinois.edu Abstract In this project, we study a method to reconstruct a 3D scene from two views. First, we extract

More information

Camera Calibration from the Quasi-affine Invariance of Two Parallel Circles

Camera Calibration from the Quasi-affine Invariance of Two Parallel Circles Camera Calibration from the Quasi-affine Invariance of Two Parallel Circles Yihong Wu, Haijiang Zhu, Zhanyi Hu, and Fuchao Wu National Laboratory of Pattern Recognition, Institute of Automation, Chinese

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 10 130221 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Canny Edge Detector Hough Transform Feature-Based

More information

Exponential Maps for Computer Vision

Exponential Maps for Computer Vision Exponential Maps for Computer Vision Nick Birnie School of Informatics University of Edinburgh 1 Introduction In computer vision, the exponential map is the natural generalisation of the ordinary exponential

More information

A Stratified Approach for Camera Calibration Using Spheres

A Stratified Approach for Camera Calibration Using Spheres IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. XX, NO. Y, MONTH YEAR 1 A Stratified Approach for Camera Calibration Using Spheres Kwan-Yee K. Wong, Member, IEEE, Guoqiang Zhang, Student-Member, IEEE and Zhihu

More information

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz Epipolar Geometry Prof. D. Stricker With slides from A. Zisserman, S. Lazebnik, Seitz 1 Outline 1. Short introduction: points and lines 2. Two views geometry: Epipolar geometry Relation point/line in two

More information

A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images

A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images Peter F Sturm and Stephen J Maybank Computational Vision Group, Department of Computer Science The University of

More information

Metric Structure from Motion

Metric Structure from Motion CS443 Final Project Metric Structure from Motion Peng Cheng 1 Objective of the Project Given: 1. A static object with n feature points and unknown shape. 2. A camera with unknown intrinsic parameters takes

More information

Automatic Photo Popup

Automatic Photo Popup Automatic Photo Popup Derek Hoiem Alexei A. Efros Martial Hebert Carnegie Mellon University What Is Automatic Photo Popup Introduction Creating 3D models from images is a complex process Time-consuming

More information

Vision Review: Image Formation. Course web page:

Vision Review: Image Formation. Course web page: Vision Review: Image Formation Course web page: www.cis.udel.edu/~cer/arv September 10, 2002 Announcements Lecture on Thursday will be about Matlab; next Tuesday will be Image Processing The dates some

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image

More information

Agenda. Rotations. Camera calibration. Homography. Ransac

Agenda. Rotations. Camera calibration. Homography. Ransac Agenda Rotations Camera calibration Homography Ransac Geometric Transformations y x Transformation Matrix # DoF Preserves Icon translation rigid (Euclidean) similarity affine projective h I t h R t h sr

More information

DD2429 Computational Photography :00-19:00

DD2429 Computational Photography :00-19:00 . Examination: DD2429 Computational Photography 202-0-8 4:00-9:00 Each problem gives max 5 points. In order to pass you need about 0-5 points. You are allowed to use the lecture notes and standard list

More information