A Framework for 3D Pushbroom Imaging CUCS

Size: px
Start display at page:

Download "A Framework for 3D Pushbroom Imaging CUCS"

Transcription

1 A Framework for 3D Pushbroom Imaging CUCS Naoyuki Ichimura and Shree K. Nayar Information Technology Research Institute, National Institute of Advanced Industrial Science and Technology (AIST) Tsukuba, Ibaraki , Japan Department of Computer Science, Columbia University New York, New York 10027, USA This work was supported in parts by the JSPS Overseas Research Fellowship.

2 Abstract Pushbroom cameras produce one-dimensional images of a scene with high resolution at a high frame-rate. As a result, they provide superior data compared to conventional twodimensional cameras in cases where the scene of interest can be temporally scanned. In this paper, we consider the problem of recovering the structure of a scene using a set of pushbroom cameras. Although pushbroom cameras have been used to recover scene structure in the past, the algorithms for recovery were developed separately for different camera motions such as translation and rotation. In this paper, we present a general framework of structure recovery for pushbroom cameras with 6 degree-of-freedom motion. We analyze the translation and rotation cases using our framework and demonstrate that several previous results are really special cases of our result. Using this framework, we also show that three or more pushbroom cameras can be used to compute scene structure as well as motion of translation or rotation. We conclude with a set of experiments that demonstrate the use of pushbroom imaging to recover structure from unknown motion. 2

3 Contents 1 Introduction 4 2 General Framework for Structure Recovery Imaging Geometry Deriving Geometric Constraints Recovering Structure Cases of Translation and Rotation Translating Pushbroom Cameras Rotating Pushbroom Cameras Structure from Unknown Translation or Unknown Rotation Recovering Velocity Structure from Unknown Velocity Experiments 14 6 Summary 18 A Rotation Matrices 18 B Derivation of Vertical Scale 18 3

4 space 1 Introduction A pushbroom camera, also known as a line scanner/sensor, is an imaging system with a one-dimensional (1D) array of pixels [1]. The use of a 1D detector gives a pushbroom camera several significant advantages over conventional cameras that use two-dimensional (2D) detectors: the pushbroom camera produces data with high resolution and has a high frame-rate. For example, recent off-the-shelf pushbroom cameras have a resolution of 2,048 pixels and a frame-rate of 20 KHz [2]. For these reasons, pushbroom cameras have begun to emerge as a popular alternative to 2D cameras, especially in the fields of remote sensing and visual inspection, where it is convenient to scan the scene of interest [3, 4]. Figure 1(a) shows a scene being scanned by a moving pushbroom camera. At each instant in time, the camera produces a 1D image which represents the brightness of the scene points that intersect the view plane of the camera. By concatenating consecutive 1D images, we obtain a 2D image called a pushbroom panorama [5], such as the one in Fig. 1(b). Note that the columns of this image include just spatial information, i.e., a scene pushbroom camera space and time view plane motion (a) Figure 1: The acquisition of a pushbroom panorama. (a) A moving pushbroom camera sweeps a scene with the view plane. A 1D image, which represents the brightness of the scene points that intersect the view plane, is produced at each instant of time. (b) An example of a pushbroom panorama created by concatenating the 1D images. The columns of this image include spatial information, while the rows include both spatial and temporal information. This embedding of spatial and temporal information can be exploited to recover both scene structure and motion. (b) 4

5 1D perspective projection of the scene, while the rows include both spatial and temporal information, i.e., projections of scene points captured at different instants of time. In short, a moving pushbroom camera encodes both spatial and temporal information within a single 2D image. As we shall see, this feature of a pushbroom camera makes it an interesting device in the context of structure and motion recovery. This paper presents a general framework for the use of pushbroom cameras for recovering the structure of a scene. Before we present this framework, a quick review of previous work is in order. Recently, several investigators have explored structure recovery using pushbroom panoramas. For demonstration purposes, pushbroom cameras were emulated with a conventional 2D camera by extracting two or more lines of pixels. Chai and Shum [6] proposed stereo reconstruction based on parallel projections realized by a camera with 1D and 2D translation. Zhu et. al. [7] considered parallel-perspective stereo mosaics obtained from a camera with 3D translation. Ishiguro [8] et. al., Shum et. al. [9] and Perr et. al. [10] presented approaches using pushbroom panoramas captured by a rotating camera to achieve scene reconstruction. Benosman et al. [11] constructed a stereo system using two rotating pushbroom cameras. It is important to note that all of these previous works are restricted in two ways: (a) they consider only specific motions of the pushbroom cameras, i.e., translation and rotation; (b) all of them require the motion of the pushbroom cameras to be known. In previous work, algorithms for recovering scene structure were developed separately for translation and rotation; there exists no common framework which subsumes the various configurations in which pushbroom cameras can be used for structure recovery. In addition, previous methods rely on prior knowledge of camera motion. This information has been obtained using a variety of devices including gyroscopes [3], GPS and INS [7], and optical encoders [8, 9, 10, 11]. The goal of this paper is to explore the conditions under which a set of pushbroom cameras can be used to recover structure. First, we present a general framework for structure recovery using pushbroom cameras. The key contribution here is a general geometric constraint, called the view plane equation. We have developed an algorithm based on this constraint which is general in that it can handle 6 degree-of-freedom (DOF) motion of the camera. As special cases, we analyze two specific camera motions, i.e., translation and rotation and show that several previous methods for structure recovery are really special cases of our framework. Next, we apply our framework to the problem of structure from unknown translation or rotation. We show that, in these cases, three (or more) pushbroom cameras can be used to recover both scene structure and motion. We conclude the paper with a set of experiments that demonstrates the practical utility of our results. 2 General Framework for Structure Recovery In this section, we derive geometric constraints imposed by a pushbroom camera with 6 DOF motion. Then, we present the fundamental equations for recovering scene structure based on the geometric constraints. 5

6 view ray 1D detector P scene ( uk, vk) Y k f p v R, t k k Y w Z w Z k X k view plane image plane X w Figure 2: Imaging geometry of a pushbroom camera. The world coordinate system is denoted by the X w, Y w and Z w axes. The camera coordinate system at the instant of time kt s is denoted by the X k, Y k and Z k axes. The relationship between the world and camera coordinate systems is given by the rotation matrix R k and translation vector t k. The world and camera coordinates of a 3D point P are p w =(x w, y w, z w ) t and p k =(x k, y k, z k ) t, respectively. The 1D detector (and hence the view plane) lies on the Y k Z k plane. A pushbroom panorama with coordinates (u k,v k ) is created by concatenating 1D images. 2.1 Imaging Geometry Figure 2 depicts the imaging geometry of a pushbroom camera. The world coordinate system is denoted by X w, Y w and Z w. The camera captures a scene at discrete instants of time represented by kt s, where k (k =0, 1, 2,...) is the time index and T s is the sampling interval. The camera motion at the time instant kt s is represented by the rotation matrix R k and translation vector t k, which define the camera coordinate system given by X k, Y k and Z k. The 1D detector of the camera lies on the Y k Z k plane and it produces a 1D image at every time instant. By concatenating successive 1D images, we can produce a pushbroom panorama with coordinates (u k,v k ). The u coordinate represents the discrete time. That is: u k = kt s. (1) The v coordinate represents spatial information. Consider a 3D scene point P whose camera coordinates are p k =(x k, y k, z k ) t. Its v coordinate is determined by a 1D perspective projection as: v k = fy k + p v, z k where, f and p v are the focal length and the image center, respectively. (2) 6

7 2.2 Deriving Geometric Constraints We denote the world coordinates of the 3D point P as p w =(x w, y w, z w ) t. We know that the camera and world coordinates are related to each other as: ( ) pw p k =(R k R k t k ), (3) 1 where, R k = (i k, j k, k k ) t = R x (θ k ) R y (φ k ) R z (ψ k ), (4) t k = (t xk,t yk,t zk ) t. (5) The rows of the rotation matrix, i k, j k and k k, define the directions of the axes of the camera coordinate system, X k, Y k and Z k. The rotation matrices R x (θ k ), R y (φ k ) and R z (ψ k ) represent rotations of θ k, φ k and ψ k about the X k, Y k and Z k axes, respectively. The detailed expressions for the rotation matrices are given in Appendix A. The orientation and position of the camera are represented by the rotation angles and translation vector, respectively. We can express the camera coordinate x k by expanding Eq.(3) as follows: where, x k = i k p w i k t k, (6) i k = (cos φ k cos ψ k, cos φ k sin ψ k, sin φ k ) t. (7) Note that, since the view plane (1D detector) lies on the Y k Z k plane, we have x k =0. Therefore, we have: i k p w = i k t k. (8) The above expression which represents the view plane passing through the 3D point is called the view plane equation. It imposes a geometric constraint on the world coordinates p w and thus it is useful in recovering scene structure. Another constraint is obtained from the perspective projection of the 3D point to the 1D detector given by Eq.(2). Using y k and z k from Eq.(3), we have: where, r k p w = r k t k, (9) r k =(v k p v ) k k fj k, (10) j k = (cos θ k sin ψ k + sin θ k sin φ k cos ψ k, (11) cos θ k cos ψ k sin θ k sin φ k sin ψ k, sin θ k cos φ k ), (12) k k = (sin θ k sin ψ k cos θ k sin φ k cos ψ k, (13) sin θ k cos ψ k + cos θ k sin φ k sin ψ k, cos θ k cos φ k ). (14) In summary, we have two geometric constraints (Eqs. (8) and (9)) on the world coordinates of a scene point from one pushbroom camera. 7

8 2.3 Recovering Structure We need three constraints to recover the three coordinates p w =(x w, y w, z w ) t of a scene point. As one camera yields only two constraints, we need one more camera. The view plane equations for the two cameras can be written as: i k1 p w = i k1 t k1 and i k2 p w = i k2 t k2, (15) where, the time indices k 1 and k 2 represent the instants at which the cameras observe the 3D point. The constraints obtained from the 1D projections are: r k1 p w = r k1 t k1 and r k2 p w = r k2 t k2. (16) Using the above expressions, we can recover the coordinates of the 3D point, because we have four constraints and only three unknowns. For example, a matrix equation which has the following form can be constructed from Eq.(15) and the first expression in Eq.(16): Ap w = b (17) where, A =(i k1, i k2, r k1, ) t isa3 3 matrix and b =(i k1 t k1, i k2 t k2, r k1 t k1 ) t is a 3 1 vector. If the internal parameters and the motion of the cameras are known, then the elements of A and b can be computed and the coordinates (x w,y w,z w ) t can be found. The matrix A must satisfy RankA=3 to have a solution. This means that the two pushbroom cameras must have different motions. When the camera motions are identical then we have a degeneracy as RankA <3 in that case. If we have more cameras, we simply add more geometric constraints (represented by Eqs.(8) and (9)) and use the least mean square method to recover structure. The structure recovery described in this section is general in that it can handle 6 DOF camera motion and multiple cameras. In the subsequent sections, we will demonstrate the generality of this result by analyzing specific camera motions, such as translation and rotation. 3 Cases of Translation and Rotation We consider two specific camera motions, translation and rotation, as special cases of our framework. So far, the analysis of these special cases has been done separately in several previous works [6, 7, 8, 9, 10, 11]. We show that several equations for structure recovery used in the past can be easily derived from our framework. 3.1 Translating Pushbroom Cameras Figure 3 shows pushbroom cameras with 3D translation. As the orientations of the cameras are time independent (fixed), the camera coordinate system at time k is defined The cameras may be rigidly attached to each other but, in that case, they must have different orientations; this naturally results in different motions. 8

9 Y W X W Y k X k Z W t k f Z k H fixation plane d x /2 ( xw, yw, zw) Figure 3: The configuration of a pushbroom camera with 3D translation. The orientation (rotation angles) of the camera is fixed and thus the time dependent motion is only the translation t k. by: R k = R x (θ) R y (φ) R z (ψ), (18) t k = (t xk,t yk,t zk ) t. (19) We can reconstruct scene structure using Eqs.(15) and (16) if we have two or more pushbroom cameras with the different motions represented by Eqs.(18) and (19). We now show that this result can be used to describe previous work on parallel perspective stereo mosaics [6, 7]. Assume that we have two cameras with rotation angles (0,φ 1, 0) and (0,φ 2, 0) which satisfy φ 1 = φ 2 and φ 1 = φ 2 = φ. From Eq.(8), we have the following view plane equation of the first camera: x w cos φ 1 +z w sin φ 1 = t xk1 cos φ 1 + t zk1 sin φ 1. (20) The view plane equation for the second camera can be obtained by replacing φ 1, t xk1 and t zk1 by φ 2, t xk2 and t zk2, respectively. The depth z w can be calculated from the intersection of the two view planes as: z w = t xk2 t xk1 2 tan φ + t zk1 + t zk2. (21) 2 In terms of the notation used in [7], the above expression can be written as: ( z w = H 1+ x ) d x + t zk1 + t zk2 2, (22) where, x = x2 x1, x1 = f t H xk1+ dx, x2 = f t 2 H xk2 dx (dx/2), f is the focal length, tan φ = 2 f and H is the height of the fixation plane [7] (see Fig. 3). 1D parallel stereo [6] is a special case of the 3D case, where t yk1 = t zk1 = 0 and t yk2 = t zk2 = 0. Thus, the first term of Eq.(21) is equivalent to the depth equation for 1D parallel stereo given in [6]. 9

10 Y k ( xw, yw, zw) Z k X k R k Y w Z w H X w Figure 4: The configuration of a pushbroom camera with rotation. The imaging system is rotated about the Y w axis with the angular velocity ω. The elevation of the camera is H. The radius of the rotation circle is R. The rotation angle of the camera at time k is ξ k. view plane Z k normal to circle X k Z w Y w Y k R k X w Figure 5: Definition of tilt angle τ. The angle between the normal to the circle and the Y k Z k plane remains τ, regardless of the rotation angle ξ k of the camera. 3.2 Rotating Pushbroom Cameras Figure 4 shows a pushbroom camera rotating about the Y w axis with the angular velocity ω. The radius of the circle, i.e. the distance between the center of rotation and the origin of the camera coordinate system, is R. The elevation of the camera is H. The rotation angle ξ k of the camera at time k is given by: ξ k = ωkt s + ξ 0. (23) where, ξ 0 is the initial angle which corresponds to the initial position of the camera. The camera coordinate system at time k is defined by the following motion: R k = R x (θ) R y (φ k ) R z (ψ), (24) t k = (R cos ξ k,h,rsin ξ k ) t. (25) 10

11 ( xw, yw, zw) Z k2 Z k1 k2 Z w Y w R k1 X w Figure 6: The camera configuration used by omnidirectional stereo [8] and concentric mosaics [9]. Note that only φ k is time dependent among the three rotation angles. Figure 5 shows the definition of the tilt angle τ which is the angle between the normal to the circle and the direction of the view plane. Since the view plane of the camera is restricted to the Y k Z k plane of the camera coordinate system, the rotation of the imaging system about the Y w axis only changes the angle φ k in Eq. (24). Therefore, φ k can be determined from the tilt angle τ and the rotation of the camera ξ k as: φ k = ξ k (π/2) + τ. (26) We can reconstruct scene structure using two or more pushbroom cameras with different motions given by Eqs.(24) and (25). The above result includes previous works such as omnidirectional stereo and concentric mosaics [8, 9, 10, 11]. To show this, we derive the equations for structure recovery and epipolar geometry used in these previous works using our view plane equation. Figure 6 shows the camera configuration used in omnidirectional stereo and concentric mosaics [8, 9]. Let the two cameras have the following rotation angles and translation vectors: (0,φ k1, 0), (R cos ξ k1,h,rsin ξ k1 ), (0,φ k2, 0) and (R cos ξ k2,h,rsin ξ k2 ), where, φ k1 = ξ k1 (π/2) + τ 1 and φ k2 = ξ k2 (π/2) + τ 2. In Fig. 6, assume that τ 1 = τ 2 and τ 1 = τ 2 = τ. The view plane equation for the first camera is: x w cos φ k1 +z w sin φ k1 = R sin τ 1, (27) φ k1 = ξ k1 (π/2) + τ 1. (28) The view plane equation for the second camera can be obtained by replacing φ k1, ξ k1 and τ 1 by φ k2, ξ k2 and τ 2, respectively. From these equations, the depth of a 3D point can be 11

12 determined as: R sin τ z w = sin (2τ (ξ k2 ξ k1 ))/ (cos φ k1 + cos φ k2 ). (29) Although our notations for orientation and position are different from those used in the previous works, Eq.(29) is essentially equivalent to the depth equation derived in [8]; depth is proportional to the radius and the tilt angle, and it is inversely proportional to the difference between the tilt angle and the rotation angles. We now derive the vertical scale defined in [9]. The vertical scale is the ratio of the image coordinates v k1 and v k2 given by Eq.(2). It is used to determine whether corresponding features lie on the same scan line of two panoramas. In Appendix B, we show that the vertical scale can be written as: S vs = v k1 v k2 = 1 ( R sin τ r 2 1 ( R sin τ r 1 ) 2 R r cos τ 2 ) 2 R r cos τ 1, (30) where, r = x w2 +z w2. In summary, Eqs. (22), (29) and (30) demonstrate that several previous results using translation and rotation are really special cases of our approach to pushbroom imaging. 4 Structure from Unknown Translation or Unknown Rotation As we discussed in Section 2, an imaging system with two pushbroom cameras is sufficient to compute structure when the motion of the system is known. We now show how three pushbroom cameras are sufficient to compute both structure and motion, when the motion is 1D translation or rotation. 4.1 Recovering Velocity In the translation and rotation cases, we can classify motion parameters into two groups: time independent and time dependent. In 1D translation, which is a special case (t yk = t zk = 0) of Section 3.1, the angles θ, φ and ψ are time independent and the translation t xk is time dependent. In the rotation case of Section 3.2, the angles θ, τ and ψ of camera orientation are time independent and the rotation angle ξ k given by Eq.(23) is time dependent. While we can determine the time independent parameters by off-line calibration, we must recover the time dependent parameters. We show that one additional pushbroom camera is sufficient to achieve this. In the 1D translation case, the translation t xk1 and t xk2 of the cameras at the times k 1 and k 2 when the cameras observe the same 3D point are given by: t xk1 = v x k 1 T s + t x01 and t xk2 = v x k 2 T s + t x02, (31) 12

13 Y W X W d Z w k R Z W t xk Y w X w (a) translation (b) rotation Figure 7: Configurations of two cameras for recovering velocity of (a) translation and (b) rotation. The cameras must have the same motion so that they observe the same 3D scene point from the same position. In both cases (translation and rotation), the velocity of the cameras can be found from the difference of a scene point s observation times, which can be found from feature correspondences between the two acquired pushbroom panoramas and the distance d (for translation) or the angle δ (for rotation) between the two cameras. where, v x is the velocity of the translation and t x01 and t x02 are the initial positions of cameras. In the rotation case, the rotation angles ξ k1 and ξ k2 are given by Eq.(23): ξ k1 = ωk 1 T s + ξ 01 and ξ k2 = ωk 2 T s + ξ 02,. (32) We need to know the translations and the rotation angles at the times k 1 and k 2 to recover structure. In Eqs.(31) and (32), the initial positions of the cameras, t x01, t x02, ξ 01 and ξ 02, can be obtained from calibration. The sampling interval T s is known. Note that the indices k 1 and k 2 are column numbers in the two acquired pushbroom panoramas. They can be determined by finding correspondences (feature matching) between the two panoramas. To determine the translations and rotation angles of the cameras, we need only the velocity v x and angular velocity ω. Figure 7 shows two cameras with the same motion and different initial positions for (a) translation and (b) rotation. That is, the two cameras have the same orientation and are rigidly fixed with respect to each other to have different initial positions. The difference between the initial positions of the cameras is d or δ. As we discussed in Section 2.3, we cannot use these configurations to recover structure since the motions of the two cameras are identical; the rank of A in Eq.(17) is 2 in this case. Interestingly, these configurations can be used to recover the linear or angular velocity. Since the view plane equations of the two cameras are identical, they observe the same 3D point from the same position regardless of scene structure. The difference between the observation time is d/v x or δ/ω. Therefore, the u coordinates of the projections of the 3D 13

14 Figure 8: An imaging system with three pushbroom cameras. The orientations of the cameras and the distances between cameras can be adjusted using the positioning stages that the cameras are mounted on. The cameras are synchronized and produce 850 lines of pixels per second and each line has 1024 pixels. The time resolution of this system is very high, 1/850= [sec]. point for the two cameras are: u k1 = k 1 T s, u k2 = k 2 T s = k 1 T s + t, (33) { d/vx (translation) t = (34) δ/ω (rotation). Now, we can recover the velocity as: v x = d (translation), (35) u ω = δ (rotation), (36) u where, u =(k 2 k 1 ) T s. 4.2 Structure from Unknown Velocity In our analysis in Sections 4.1 and 3, we have shown that we can recover both the structure of the scene and the velocity using three or more pushbroom cameras. Two cameras must have the same motion and different initial positions to recover velocity. The other camera(s) must have different motions to recover structure. 5 Experiments Figure 8 shows the imaging system we have configured with three pushbroom cameras (model L103 from Basler). The orientations of the cameras and the distances between cameras can be adjusted and determined using the positioning stages that the cameras are mounted on. The cameras are synchronized and produce 850 lines of pixels per second and each line has 1024 pixels. That is, the time resolution of this system is 1/850= [sec]. 14

15 (a) (b) Figure 9: Comparison between the pushbroom panoramas captured by (a) an NTSC camera and (b) a pushbroom camera for an object with the same velocity. The field of views of the two cameras are identical. The former has only 25 columns due to the slow frame rate while the latter has 708 columns. This example shows the clear advantage of using pushbroom cameras for capturing moving objects. First, we show results of experiments with 1D translation. In the experiments, the cameras Pb1, Pb2 and Pb3 (see Fig. 8) had the orientations (0, φ, 0), (0, φ, 0) and (0,φ,0), where φ = 7[deg]. The distances between Pb1 and Pb2, and Pb2 and Pb3 (d in Fig. 7 (a)) were 5 and 8 inches, respectively. To emphasize the importance of the velocity recovery using pushbroom cameras, we moved an object instead of the imaging system. When we move the imaging system, we can use devices for measuring motion such as gyroscopes and optical encoders. We cannot, however, use such devices for moving objects and thus we need to measure velocity only from captured images. Note that, the principle of structure recovery for the moving object case is the same for the moving camera case discussed in Sections 3.1 and 4.1. The object was moved by a robot (AdeptOne Robot [12]). Figure 9 shows the difference between pushbroom panoramas created by a conventional NTSC camera and a pushbroom camera. The object was moving and we obtained an image with only 25 columns from the NTSC camera while we obtained an image with 708 columns using the pushbroom camera. Figure 10 shows three pushbroom panoramas of the object generated using our system. The size of each image is [pixel] (the images in Fig. 10 were cropped for display purposes). The first and second images, which were generated by Pb2 and Pb3, were used to recover the velocity of the object. We used an image alignment method to find the velocity. We extracted corners in each image using a corner detector [13] and then calculated the parameters of a translational motion model using corresponding corners. The horizontal translation between the two images shows the time difference u in Eq.(35) needed to calculate velocity. Normalized correlation was used to find the correspondences and the RANSAC [14] was applied to remove false matching. The estimated velocity was 434[mm/sec]. This value is exactly the same speed by the robot (ground truth). 15

16 Figure 10: Pushbroom panoramas of a moving object produced by the system in Fig. 8. The size of each image is [pixel] (the images were cropped for display purposes). The left and middle panoramas were used to recover velocity, while the middle and right panoramas were used to recover 3D structure. (a) (b) Figure 11: Recovered structure (unknown linear velocity). (a) A depth image, where brighter pixels correspond to closer scene points. (b) A VRML model of the moving object as seen from a different viewpoint. This result shows that our method can recover fairly accurate scene structure even when the velocity of the moving object is high and unknown. This accuracy is due to the very high time resolution of the imaging system, [sec]. This result shows the advantage of using temporal information provided by pushbroom panoramas for motion estimation. Figure 11 shows the recovered structure of the moving object. The structure was computed from the second and third images in Fig. 10 using an area-based stereo matching algorithm that uses normalized correlation. Figure 11 (a) shows a depth image, where brighter pixels correspond to closer scene points. Figure 11 (b) shows a VRML model of the moving object. The depth image and VRML model demonstrate that our method can recover scene structure from unknown velocity using only pushbroom cameras. We now show results of experiments with rotation. The orientations of the pushbroom cameras were (θ, τ, ψ), (θ, τ, ψ) and (θ, τ,ψ), where θ = 15.0, τ = 20.0 and ψ = 5.0[deg], respectively. Note that unlike the previous works discussed in 3.2, none of the orientation angles are 0. The angle between the cameras (δ in Fig. 7 (b)) were 27.0[deg]. 16

17 Figure 12: Pushbroom panoramas captured by a set of rotating pushbroom cameras. The size of each image is [pixel] (the images were cropped for display purposes). The top and middle panoramas were used to recover angular velocity, while the middle and bottom panoramas were used to recover 3D structure. Figure 13: Recovered structure (unknown angular velocity). Top: A depth panorama, where brighter pixels correspond to closer scene points. Bottom: A VRML model of the scene corresponding to a different viewpoint. This result was obtained by rotating the imaging system with unknown angular velocity. The imaging system was rotated by a DC motor and angular velocity was unknown (no ground truth). Figure 12 shows the three captured pushbroom panoramas. The size of each image is [pixel]. The estimated angular velocity obtained from the first and second images was [deg/sec]. Using the estimated angular velocity, we were able to reconstruct scene structure as shown in Fig. 13, which was computed from the second and third images. 17

18 The results of the experiments demonstrate the usefulness of the method for structure from unknown velocity based on our framework. 6 Summary In this paper, we have presented a framework for structure recovery using pushbroom cameras. We derived a set of general geometric constraints, including the view plane equation, for a pushbroom camera with 6 DOF motion. The method for structure recovery from unknown velocity using three or more pushbroom cameras for the 1D translation and rotation cases was developed based on these constraints. We showed that the results of several previous works can be viewed as special cases of our approach. Experiments were performed to show the effectiveness of our approach. We believe that the framework presented here will facilitate more applications of pushbroom cameras, especially in fields where high spatial resolution and high frame-rate are important. A Rotation Matrices The entities of the rotation matrices used in the transformation between the world and camera coordinate systems in Fig. 2 are as follows: cos φ 0 sin φ cos ψ sin ψ 0 R x (θ) = 0 cos θ sin θ, R y (φ) = 0 1 0, R z (ψ) = sin ψ cos ψ 0. 0 sin θ cos θ sin φ 0 cos φ B Derivation of Vertical Scale The vertical scale which shows vertical parallax is the ratio of image coordinates v k1 and v k2 given by the following equations: v k1 = fy k1 z k1 + p v, v k2 = fy k2 z k2 + p v, (37) where, y k1, z k1, y k2 and z k2 are obtained from Eq.(3) using the orientations and the positions of the cameras. In the special case discussed in Section 3.2, y k1 =y k2 =y w (this can be verified by Eq.(3)). Thus the vertical scale is obtained as follows by removing the common term, image center p v : S vs = z k2. z k1 From Eq.(3), z k1 is obtained as follows: (38) z k1 = x w sin φ k1 +z w cos φ k1 R cos τ 1. (39) We have the following equation from Eq.(39): z k1 = r 1 sin 2 (α ξ k1 τ 1 ) R cos τ 1, (40) 18

19 where, α = tan 1 (z w /x w ) and r = x w2 +z w2. Eq.(27) can be modified using α and r as follows: The view plane equations given by sin (α ξ k1 τ 1 )= R r sin τ 1. (41) By introducing Eq.(41) into Eq.(40), we have the following representation of z k1 : ( R 2 z k1 = r 1 1) r sin τ R cos τ 1. (42) We can do the same manipulation for z k2. Then Eq.(30) is derived using the condition of r 0. References [1] R. Gupta and R. I. Hartley: Linear pushbroom cameras, IEEE Trans. Pattern Anal. & Mach. Intell., Vol.19, No.9, pp , 1997 [2] [3] A. Borner: The optimization of the stereo angle of CCD-line-Scanners, ISPRS Vol.XXXI, Part B1, Commission I, pp.26-30, 1996 [4] A. Gaich, A. Fasching and M. Gruber: Stereoscopic imaging and geological evaluation for geotechnical modeling at the tunnel site, Felsbau/Rock and Soil Engineering, No.1, pp.15-21, 1999 [5] S. M. Seitz and J. Kim: The space of all stereo images, Int. J. Computer Vision, Vol.48, No.1, pp.21-38, 2002 [6] J. X. Chai and H. Y. Shum: Parallel projections for stereo reconstruction, Proc. CVPR 00, Vol.II, pp , 2000 [7] Z. Zhu, E. M. Riseman and A. R. Hanson: Parallel-Perspective Stereo Mosaics, Proc. ICCV 01, Vol.I, pp , 2001 [8] H. Ishiguro, H. Yamamoto and S. Tsuji: Omni-directional Stereo, IEEE Trans. Pattern Anal. & Mach. Intell., Vol.14, pp , 1992 [9] H. Y. Shum and R. Szeliski: Stereo reconstruction from multiperspective panoramas, Proc. ICCV 99, pp.14-21, 1999 [10] P. Peer and F. Solina: Mosaic-based panoramic depth imaging with a single standard camera, Proc. Workshop on Stereo and Multi-Baseline Vision, pp.75-84, 2001 [11] R. Benosman and J. Devars: Calibration of the stereovision panoramic sensor, Panoramic Vision, pp , Springer-Verlag, 2001 [12] 19

20 [13] F. Chabat, G.Z. Yang, and D.M. Hansell: A corner orientation detector, Im. and Vis. Comp., 17, pp , [14] M. A. Fischler and R. C. Bolles: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography, Commun. ACM, Vol.24, No.6, pp ,

Catadioptric camera model with conic mirror

Catadioptric camera model with conic mirror LÓPEZ-NICOLÁS, SAGÜÉS: CATADIOPTRIC CAMERA MODEL WITH CONIC MIRROR Catadioptric camera model with conic mirror G. López-Nicolás gonlopez@unizar.es C. Sagüés csagues@unizar.es Instituto de Investigación

More information

Omni Stereo Vision of Cooperative Mobile Robots

Omni Stereo Vision of Cooperative Mobile Robots Omni Stereo Vision of Cooperative Mobile Robots Zhigang Zhu*, Jizhong Xiao** *Department of Computer Science **Department of Electrical Engineering The City College of the City University of New York (CUNY)

More information

Panoramic 3D Reconstruction Using Rotational Stereo Camera with Simple Epipolar Constraints

Panoramic 3D Reconstruction Using Rotational Stereo Camera with Simple Epipolar Constraints Panoramic 3D Reconstruction Using Rotational Stereo Camera with Simple Epipolar Constraints Wei Jiang Japan Science and Technology Agency 4-1-8, Honcho, Kawaguchi-shi, Saitama, Japan jiang@anken.go.jp

More information

Mathematics of a Multiple Omni-Directional System

Mathematics of a Multiple Omni-Directional System Mathematics of a Multiple Omni-Directional System A. Torii A. Sugimoto A. Imiya, School of Science and National Institute of Institute of Media and Technology, Informatics, Information Technology, Chiba

More information

A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space

A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space A Robust and Efficient Motion Segmentation Based on Orthogonal Projection Matrix of Shape Space Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-4, Umezono, Tsukuba Ibaraki, 35-8568 Japan ichimura@etl.go.jp

More information

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration

Camera Calibration. Schedule. Jesus J Caban. Note: You have until next Monday to let me know. ! Today:! Camera calibration Camera Calibration Jesus J Caban Schedule! Today:! Camera calibration! Wednesday:! Lecture: Motion & Optical Flow! Monday:! Lecture: Medical Imaging! Final presentations:! Nov 29 th : W. Griffin! Dec 1

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Measurement of Pedestrian Groups Using Subtraction Stereo

Measurement of Pedestrian Groups Using Subtraction Stereo Measurement of Pedestrian Groups Using Subtraction Stereo Kenji Terabayashi, Yuki Hashimoto, and Kazunori Umeda Chuo University / CREST, JST, 1-13-27 Kasuga, Bunkyo-ku, Tokyo 112-8551, Japan terabayashi@mech.chuo-u.ac.jp

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

On the Epipolar Geometry of the Crossed-Slits Projection

On the Epipolar Geometry of the Crossed-Slits Projection In Proc. 9th IEEE International Conference of Computer Vision, Nice, October 2003. On the Epipolar Geometry of the Crossed-Slits Projection Doron Feldman Tomas Pajdla Daphna Weinshall School of Computer

More information

Rectification and Distortion Correction

Rectification and Distortion Correction Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification

More information

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR

Laser sensors. Transmitter. Receiver. Basilio Bona ROBOTICA 03CFIOR Mobile & Service Robotics Sensors for Robotics 3 Laser sensors Rays are transmitted and received coaxially The target is illuminated by collimated rays The receiver measures the time of flight (back and

More information

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate

More information

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion

Accurate Motion Estimation and High-Precision 3D Reconstruction by Sensor Fusion 007 IEEE International Conference on Robotics and Automation Roma, Italy, 0-4 April 007 FrE5. Accurate Motion Estimation and High-Precision D Reconstruction by Sensor Fusion Yunsu Bok, Youngbae Hwang,

More information

Dynamic Pushbroom Stereo Mosaics for Moving Target Extraction

Dynamic Pushbroom Stereo Mosaics for Moving Target Extraction Dynamic Pushbroom Stereo Mosaics for Moving Target Extraction Paper ID #413 Abstract Our goal is to rapidly acquire panoramic mosaicing maps with information of all 3D (moving) targets as a light aerial

More information

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy 1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:

More information

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi

More information

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG. Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview

More information

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera

3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera 3D Environment Measurement Using Binocular Stereo and Motion Stereo by Mobile Robot with Omnidirectional Stereo Camera Shinichi GOTO Department of Mechanical Engineering Shizuoka University 3-5-1 Johoku,

More information

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India

A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India A Review of Image- based Rendering Techniques Nisha 1, Vijaya Goel 2 1 Department of computer science, University of Delhi, Delhi, India Keshav Mahavidyalaya, University of Delhi, Delhi, India Abstract

More information

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few... STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own

More information

Computer Vision cmput 428/615

Computer Vision cmput 428/615 Computer Vision cmput 428/615 Basic 2D and 3D geometry and Camera models Martin Jagersand The equation of projection Intuitively: How do we develop a consistent mathematical framework for projection calculations?

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric

More information

A Summary of Projective Geometry

A Summary of Projective Geometry A Summary of Projective Geometry Copyright 22 Acuity Technologies Inc. In the last years a unified approach to creating D models from multiple images has been developed by Beardsley[],Hartley[4,5,9],Torr[,6]

More information

Camera Model and Calibration

Camera Model and Calibration Camera Model and Calibration Lecture-10 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the

More information

Stereo Image Rectification for Simple Panoramic Image Generation

Stereo Image Rectification for Simple Panoramic Image Generation Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,

More information

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry

More information

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz

Epipolar Geometry Prof. D. Stricker. With slides from A. Zisserman, S. Lazebnik, Seitz Epipolar Geometry Prof. D. Stricker With slides from A. Zisserman, S. Lazebnik, Seitz 1 Outline 1. Short introduction: points and lines 2. Two views geometry: Epipolar geometry Relation point/line in two

More information

Self-calibration of a pair of stereo cameras in general position

Self-calibration of a pair of stereo cameras in general position Self-calibration of a pair of stereo cameras in general position Raúl Rojas Institut für Informatik Freie Universität Berlin Takustr. 9, 14195 Berlin, Germany Abstract. This paper shows that it is possible

More information

Motion Segmentation Based on Factorization Method and Discriminant Criterion

Motion Segmentation Based on Factorization Method and Discriminant Criterion Motion Segmentation Based on Factorization Method and Discriminant Criterion Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-, Umezono, Tsukuba Ibaraki, 35-8568 Japan ichimura@etl.go.jp Abstract A motion

More information

Omnivergent Stereo-panoramas with a Fish-eye Lens

Omnivergent Stereo-panoramas with a Fish-eye Lens CENTER FOR MACHINE PERCEPTION CZECH TECHNICAL UNIVERSITY Omnivergent Stereo-panoramas with a Fish-eye Lens (Version 1.) Hynek Bakstein and Tomáš Pajdla bakstein@cmp.felk.cvut.cz, pajdla@cmp.felk.cvut.cz

More information

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: , 3D Sensing and Reconstruction Readings: Ch 12: 12.5-6, Ch 13: 13.1-3, 13.9.4 Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by Space Carving 3D Shape from X means getting 3D coordinates

More information

A Novel Stereo Camera System by a Biprism

A Novel Stereo Camera System by a Biprism 528 IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, VOL. 16, NO. 5, OCTOBER 2000 A Novel Stereo Camera System by a Biprism DooHyun Lee and InSo Kweon, Member, IEEE Abstract In this paper, we propose a novel

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II

Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II Massachusetts Institute of Technology Department of Computer Science and Electrical Engineering 6.801/6.866 Machine Vision QUIZ II Handed out: 001 Nov. 30th Due on: 001 Dec. 10th Problem 1: (a (b Interior

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp

More information

Introduction to Homogeneous coordinates

Introduction to Homogeneous coordinates Last class we considered smooth translations and rotations of the camera coordinate system and the resulting motions of points in the image projection plane. These two transformations were expressed mathematically

More information

Motion Segmentation Based on Factorization Method and Discriminant Criterion

Motion Segmentation Based on Factorization Method and Discriminant Criterion Motion Segmentation Based on Factorization Method and Discriminant Criterion Naoyuki ICHIMURA Electrotechnical Laboratory 1-1-, Umezono, Tsukuba Ibaraki, -6 Japan ichimura@etl.go.jp Abstract A motion segmentation

More information

Homographies and RANSAC

Homographies and RANSAC Homographies and RANSAC Computer vision 6.869 Bill Freeman and Antonio Torralba March 30, 2011 Homographies and RANSAC Homographies RANSAC Building panoramas Phototourism 2 Depth-based ambiguity of position

More information

Error Characteristics of Parallel-Perspective Stereo Mosaics

Error Characteristics of Parallel-Perspective Stereo Mosaics Error Characteristics of Parallel-Perspective Stereo Mosaics Zhigang Zhu, Allen R. anson, oward Schultz, Edward M. Riseman Department of Computer Science, University of Massachusetts at Amherst, MA 01003

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming

Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Visual Odometry for Non-Overlapping Views Using Second-Order Cone Programming Jae-Hak Kim 1, Richard Hartley 1, Jan-Michael Frahm 2 and Marc Pollefeys 2 1 Research School of Information Sciences and Engineering

More information

Compositing a bird's eye view mosaic

Compositing a bird's eye view mosaic Compositing a bird's eye view mosaic Robert Laganiere School of Information Technology and Engineering University of Ottawa Ottawa, Ont KN 6N Abstract This paper describes a method that allows the composition

More information

Visual Tracking of a Hand-eye Robot for a Moving Target Object with Multiple Feature Points: Translational Motion Compensation Approach

Visual Tracking of a Hand-eye Robot for a Moving Target Object with Multiple Feature Points: Translational Motion Compensation Approach Visual Tracking of a Hand-eye Robot for a Moving Target Object with Multiple Feature Points: Translational Motion Compensation Approach Masahide Ito Masaaki Shibata Department of Electrical and Mechanical

More information

Summary Page Robust 6DOF Motion Estimation for Non-Overlapping, Multi-Camera Systems

Summary Page Robust 6DOF Motion Estimation for Non-Overlapping, Multi-Camera Systems Summary Page Robust 6DOF Motion Estimation for Non-Overlapping, Multi-Camera Systems Is this a system paper or a regular paper? This is a regular paper. What is the main contribution in terms of theory,

More information

Jump Stitch Metadata & Depth Maps Version 1.1

Jump Stitch Metadata & Depth Maps Version 1.1 Jump Stitch Metadata & Depth Maps Version 1.1 jump-help@google.com Contents 1. Introduction 1 2. Stitch Metadata File Format 2 3. Coverage Near the Poles 4 4. Coordinate Systems 6 5. Camera Model 6 6.

More information

A COMPREHENSIVE TOOL FOR RECOVERING 3D MODELS FROM 2D PHOTOS WITH WIDE BASELINES

A COMPREHENSIVE TOOL FOR RECOVERING 3D MODELS FROM 2D PHOTOS WITH WIDE BASELINES A COMPREHENSIVE TOOL FOR RECOVERING 3D MODELS FROM 2D PHOTOS WITH WIDE BASELINES Yuzhu Lu Shana Smith Virtual Reality Applications Center, Human Computer Interaction Program, Iowa State University, Ames,

More information

COMP30019 Graphics and Interaction Three-dimensional transformation geometry and perspective

COMP30019 Graphics and Interaction Three-dimensional transformation geometry and perspective COMP30019 Graphics and Interaction Three-dimensional transformation geometry and perspective Department of Computing and Information Systems The Lecture outline Introduction Rotation about artibrary axis

More information

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10

Structure from Motion. Introduction to Computer Vision CSE 152 Lecture 10 Structure from Motion CSE 152 Lecture 10 Announcements Homework 3 is due May 9, 11:59 PM Reading: Chapter 8: Structure from Motion Optional: Multiple View Geometry in Computer Vision, 2nd edition, Hartley

More information

Computer Vision I - Appearance-based Matching and Projective Geometry

Computer Vision I - Appearance-based Matching and Projective Geometry Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 01/11/2016 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

3D Modeling using multiple images Exam January 2008

3D Modeling using multiple images Exam January 2008 3D Modeling using multiple images Exam January 2008 All documents are allowed. Answers should be justified. The different sections below are independant. 1 3D Reconstruction A Robust Approche Consider

More information

Comments on Consistent Depth Maps Recovery from a Video Sequence

Comments on Consistent Depth Maps Recovery from a Video Sequence Comments on Consistent Depth Maps Recovery from a Video Sequence N.P. van der Aa D.S. Grootendorst B.F. Böggemann R.T. Tan Technical Report UU-CS-2011-014 May 2011 Department of Information and Computing

More information

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion

More information

Computer Vision I - Appearance-based Matching and Projective Geometry

Computer Vision I - Appearance-based Matching and Projective Geometry Computer Vision I - Appearance-based Matching and Projective Geometry Carsten Rother 05/11/2015 Computer Vision I: Image Formation Process Roadmap for next four lectures Computer Vision I: Image Formation

More information

Task selection for control of active vision systems

Task selection for control of active vision systems The 29 IEEE/RSJ International Conference on Intelligent Robots and Systems October -5, 29 St. Louis, USA Task selection for control of active vision systems Yasushi Iwatani Abstract This paper discusses

More information

How to Compute the Pose of an Object without a Direct View?

How to Compute the Pose of an Object without a Direct View? How to Compute the Pose of an Object without a Direct View? Peter Sturm and Thomas Bonfort INRIA Rhône-Alpes, 38330 Montbonnot St Martin, France {Peter.Sturm, Thomas.Bonfort}@inrialpes.fr Abstract. We

More information

Camera Model and Calibration. Lecture-12

Camera Model and Calibration. Lecture-12 Camera Model and Calibration Lecture-12 Camera Calibration Determine extrinsic and intrinsic parameters of camera Extrinsic 3D location and orientation of camera Intrinsic Focal length The size of the

More information

LUMS Mine Detector Project

LUMS Mine Detector Project LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines

More information

Lecture 9: Epipolar Geometry

Lecture 9: Epipolar Geometry Lecture 9: Epipolar Geometry Professor Fei Fei Li Stanford Vision Lab 1 What we will learn today? Why is stereo useful? Epipolar constraints Essential and fundamental matrix Estimating F (Problem Set 2

More information

MAPI Computer Vision. Multiple View Geometry

MAPI Computer Vision. Multiple View Geometry MAPI Computer Vision Multiple View Geometry Geometry o Multiple Views 2- and 3- view geometry p p Kpˆ [ K R t]p Geometry o Multiple Views 2- and 3- view geometry Epipolar Geometry The epipolar geometry

More information

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction

Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Computer Vision I - Algorithms and Applications: Multi-View 3D reconstruction Carsten Rother 09/12/2013 Computer Vision I: Multi-View 3D reconstruction Roadmap this lecture Computer Vision I: Multi-View

More information

Visualization 2D-to-3D Photo Rendering for 3D Displays

Visualization 2D-to-3D Photo Rendering for 3D Displays Visualization 2D-to-3D Photo Rendering for 3D Displays Sumit K Chauhan 1, Divyesh R Bajpai 2, Vatsal H Shah 3 1 Information Technology, Birla Vishvakarma mahavidhyalaya,sumitskc51@gmail.com 2 Information

More information

Minimal Solutions for Generic Imaging Models

Minimal Solutions for Generic Imaging Models Minimal Solutions for Generic Imaging Models Srikumar Ramalingam Peter Sturm Oxford Brookes University, UK INRIA Grenoble Rhône-Alpes, France Abstract A generic imaging model refers to a non-parametric

More information

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing

Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing Visual Tracking of Unknown Moving Object by Adaptive Binocular Visual Servoing Minoru Asada, Takamaro Tanaka, and Koh Hosoda Adaptive Machine Systems Graduate School of Engineering Osaka University, Suita,

More information

Degeneracy of the Linear Seventeen-Point Algorithm for Generalized Essential Matrix

Degeneracy of the Linear Seventeen-Point Algorithm for Generalized Essential Matrix J Math Imaging Vis 00 37: 40-48 DOI 0007/s085-00-09-9 Authors s version The final publication is available at wwwspringerlinkcom Degeneracy of the Linear Seventeen-Point Algorithm for Generalized Essential

More information

A Fast Linear Registration Framework for Multi-Camera GIS Coordination

A Fast Linear Registration Framework for Multi-Camera GIS Coordination A Fast Linear Registration Framework for Multi-Camera GIS Coordination Karthik Sankaranarayanan James W. Davis Dept. of Computer Science and Engineering Ohio State University Columbus, OH 4320 USA {sankaran,jwdavis}@cse.ohio-state.edu

More information

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with

More information

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches

Reminder: Lecture 20: The Eight-Point Algorithm. Essential/Fundamental Matrix. E/F Matrix Summary. Computing F. Computing F from Point Matches Reminder: Lecture 20: The Eight-Point Algorithm F = -0.00310695-0.0025646 2.96584-0.028094-0.00771621 56.3813 13.1905-29.2007-9999.79 Readings T&V 7.3 and 7.4 Essential/Fundamental Matrix E/F Matrix Summary

More information

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives

More information

Today s lecture. Image Alignment and Stitching. Readings. Motion models

Today s lecture. Image Alignment and Stitching. Readings. Motion models Today s lecture Image Alignment and Stitching Computer Vision CSE576, Spring 2005 Richard Szeliski Image alignment and stitching motion models cylindrical and spherical warping point-based alignment global

More information

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments

Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Proc. 2001 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems pp. 31-36, Maui, Hawaii, Oct./Nov. 2001. Realtime Omnidirectional Stereo for Obstacle Detection and Tracking in Dynamic Environments Hiroshi

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Extrinsic camera calibration method and its performance evaluation Jacek Komorowski 1 and Przemyslaw Rokita 2 arxiv:1809.11073v1 [cs.cv] 28 Sep 2018 1 Maria Curie Sklodowska University Lublin, Poland jacek.komorowski@gmail.com

More information

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences

Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Camera Parameters Estimation from Hand-labelled Sun Sositions in Image Sequences Jean-François Lalonde, Srinivasa G. Narasimhan and Alexei A. Efros {jlalonde,srinivas,efros}@cs.cmu.edu CMU-RI-TR-8-32 July

More information

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW

ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW ROBUST LINE-BASED CALIBRATION OF LENS DISTORTION FROM A SINGLE VIEW Thorsten Thormählen, Hellward Broszio, Ingolf Wassermann thormae@tnt.uni-hannover.de University of Hannover, Information Technology Laboratory,

More information

Epipolar Geometry in Stereo, Motion and Object Recognition

Epipolar Geometry in Stereo, Motion and Object Recognition Epipolar Geometry in Stereo, Motion and Object Recognition A Unified Approach by GangXu Department of Computer Science, Ritsumeikan University, Kusatsu, Japan and Zhengyou Zhang INRIA Sophia-Antipolis,

More information

C / 35. C18 Computer Vision. David Murray. dwm/courses/4cv.

C / 35. C18 Computer Vision. David Murray.   dwm/courses/4cv. C18 2015 1 / 35 C18 Computer Vision David Murray david.murray@eng.ox.ac.uk www.robots.ox.ac.uk/ dwm/courses/4cv Michaelmas 2015 C18 2015 2 / 35 Computer Vision: This time... 1. Introduction; imaging geometry;

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

Unit 3 Multiple View Geometry

Unit 3 Multiple View Geometry Unit 3 Multiple View Geometry Relations between images of a scene Recovering the cameras Recovering the scene structure http://www.robots.ox.ac.uk/~vgg/hzbook/hzbook1.html 3D structure from images Recover

More information

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication

DD2423 Image Analysis and Computer Vision IMAGE FORMATION. Computational Vision and Active Perception School of Computer Science and Communication DD2423 Image Analysis and Computer Vision IMAGE FORMATION Mårten Björkman Computational Vision and Active Perception School of Computer Science and Communication November 8, 2013 1 Image formation Goal:

More information

Precise Omnidirectional Camera Calibration

Precise Omnidirectional Camera Calibration Precise Omnidirectional Camera Calibration Dennis Strelow, Jeffrey Mishler, David Koes, and Sanjiv Singh Carnegie Mellon University {dstrelow, jmishler, dkoes, ssingh}@cs.cmu.edu Abstract Recent omnidirectional

More information

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry 55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence

More information

A simple method for interactive 3D reconstruction and camera calibration from a single view

A simple method for interactive 3D reconstruction and camera calibration from a single view A simple method for interactive 3D reconstruction and camera calibration from a single view Akash M Kushal Vikas Bansal Subhashis Banerjee Department of Computer Science and Engineering Indian Institute

More information

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES

FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES FAST REGISTRATION OF TERRESTRIAL LIDAR POINT CLOUD AND SEQUENCE IMAGES Jie Shao a, Wuming Zhang a, Yaqiao Zhu b, Aojie Shen a a State Key Laboratory of Remote Sensing Science, Institute of Remote Sensing

More information

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection

CHAPTER 3. Single-view Geometry. 1. Consequences of Projection CHAPTER 3 Single-view Geometry When we open an eye or take a photograph, we see only a flattened, two-dimensional projection of the physical underlying scene. The consequences are numerous and startling.

More information

Midterm Exam Solutions

Midterm Exam Solutions Midterm Exam Solutions Computer Vision (J. Košecká) October 27, 2009 HONOR SYSTEM: This examination is strictly individual. You are not allowed to talk, discuss, exchange solutions, etc., with other fellow

More information

arxiv: v1 [cs.cv] 18 Sep 2017

arxiv: v1 [cs.cv] 18 Sep 2017 Direct Pose Estimation with a Monocular Camera Darius Burschka and Elmar Mair arxiv:1709.05815v1 [cs.cv] 18 Sep 2017 Department of Informatics Technische Universität München, Germany {burschka elmar.mair}@mytum.de

More information

Synchronized Ego-Motion Recovery of Two Face-to-Face Cameras

Synchronized Ego-Motion Recovery of Two Face-to-Face Cameras Synchronized Ego-Motion Recovery of Two Face-to-Face Cameras Jinshi Cui, Yasushi Yagi, Hongbin Zha, Yasuhiro Mukaigawa, and Kazuaki Kondo State Key Lab on Machine Perception, Peking University, China {cjs,zha}@cis.pku.edu.cn

More information

CV: 3D to 2D mathematics. Perspective transformation; camera calibration; stereo computation; and more

CV: 3D to 2D mathematics. Perspective transformation; camera calibration; stereo computation; and more CV: 3D to 2D mathematics Perspective transformation; camera calibration; stereo computation; and more Roadmap of topics n Review perspective transformation n Camera calibration n Stereo methods n Structured

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester

More information

Announcements. Motion. Structure-from-Motion (SFM) Motion. Discrete Motion: Some Counting

Announcements. Motion. Structure-from-Motion (SFM) Motion. Discrete Motion: Some Counting Announcements Motion HW 4 due Friday Final Exam: Tuesday, 6/7 at 8:00-11:00 Fill out your CAPES Introduction to Computer Vision CSE 152 Lecture 20 Motion Some problems of motion 1. Correspondence: Where

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week

More information

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003

CS 664 Slides #9 Multi-Camera Geometry. Prof. Dan Huttenlocher Fall 2003 CS 664 Slides #9 Multi-Camera Geometry Prof. Dan Huttenlocher Fall 2003 Pinhole Camera Geometric model of camera projection Image plane I, which rays intersect Camera center C, through which all rays pass

More information

Camera Calibration Using Line Correspondences

Camera Calibration Using Line Correspondences Camera Calibration Using Line Correspondences Richard I. Hartley G.E. CRD, Schenectady, NY, 12301. Ph: (518)-387-7333 Fax: (518)-387-6845 Email : hartley@crd.ge.com Abstract In this paper, a method of

More information

Measurement and Precision Analysis of Exterior Orientation Element Based on Landmark Point Auxiliary Orientation

Measurement and Precision Analysis of Exterior Orientation Element Based on Landmark Point Auxiliary Orientation 2016 rd International Conference on Engineering Technology and Application (ICETA 2016) ISBN: 978-1-60595-8-0 Measurement and Precision Analysis of Exterior Orientation Element Based on Landmark Point

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry Martin Quinn with a lot of slides stolen from Steve Seitz and Jianbo Shi 15-463: Computational Photography Alexei Efros, CMU, Fall 2007 Our Goal The Plenoptic Function P(θ,φ,λ,t,V

More information