Mathematical Analysis for Visual Tracking Assuming Perspective Projection

Size: px
Start display at page:

Download "Mathematical Analysis for Visual Tracking Assuming Perspective Projection"

Transcription

1 Mathematical Analsis for Visual Tracking Assuming Perspective Projection João P. arreto, Paulo Peioto, Jorge atista, Helder Araujo Institute of Sstems and Robotics Dept. of Electrical Engineering Polo II - Universit of Coimbra 33 Coimbra Portugal Abstract. Applications of visual control of motion require that the relationships between motion in the scene and image motion be established. In the case of active tracking of moving targets these relationships become more comple due to camera motion. This work derives the position and velocit equations that relate image motion, camera motion and target 3D motion. Perspective projection is assumed. oth monocular and binocular tracking sstems are analed. The epressions obtained are simplified to decouple the control of the different mechanical degrees of freedom. The simplification errors are quantified and characteried. This stud contributes for the understanding of the tracking process, for establishing the control laws of the different camera motions,for deriving test signals to evaluate the sstem performance and for developing egomotion compensation algorithms. Introduction Fig..: The Active Vision Sstems at ISR. Right: The modular camera pan and tilt unit MVS. Left: The ISR-MDOF robot head. Visual control of motion is a widel studied subject. Several papers have addressed the problem of controlling motion using visual information [, 2, 3, 4]. In these applications, the relationships between image motion and 3D motion in the scene have to be established. In active tracking applications these relationships become more comple due to camera motion. Works like [5] and [6] address mainl the control aspects of tracking. Cameras are modeled as a constant gain and the effects of perspective projection are not considered. Affine models are assumed in [7] and approimations in the perspective model are considered in [8]. Man of these works rel on closed-loop control strategies to reduce the influence of modeling simplification. However some visual behaviors such the saccadic motion tpicall use open-loop configuration. {jpbar,peioto,batista,helder}@isr.uc.pt

2 In this paper we derive the mathematical relationships between image motion, camera motion and target 3D motion in the scene, both for monocular and binocular tracking applications. Perspective projection is considered b assuming a camera pinhole model. oth position and velocit equations are derived. The results obtained allow a better understanding of the tracking process as a regulation control problem. Strategies for both monocular and binocular tracking are developed. The simplifications of the equations derived to decouple the degrees of freedom of the vision sstem are discussed. Control laws for different visual behaviors (smooth pursuit, vergence and saccade) are established and approimation errors are studied and characteried. We are using the results of the present work to derive test signals to characterie the performance and robustness of the active tracking algorithms. Egomotion compensation techniques are being studied as well. The knowledge of the relationship between velocit in image, camera motion and targets 3D velocit can also be useful in the development of high-level visual behaviors such as target segmentation in an environment with multiple moving targets. 2 Monocular Tracking Fig. 2.: Image formation process in monocular tracking. The 3D scheme, ehibiting the variables and reference frames used in mathematical analsis, is shown from two different points of view. Fig. (R) depicts our MVS unit. The camera has two degrees of freedom: pan and tilt. oth rotation aes go through the optical center. Camera undergoes pure rotation motion. The tpical goal for a tracking application is to move the camera in such a wa that target is projected in the center of image. In 3D space it means that the optical ais must be aligned with the target. This section derives the mathematical relationship between 3D target motion, camera motion and projection in image. Camera motion is tpicall known b using motor encoders. The scheme of figure 2 describes the image formation process. Assume a standard pinhole model where camera performs a perfect perspective transform with center O (camera optic center) at a distance f (focal length) of the retinal plane. Vector M = (X, Y, Z) t represents target cartesian coordinates in R i (O, i, i, i ), (inertial reference frame). Target 3D position can also be represented b spherical coordinates (ρ, θ, φ). R c (O, c, c, c ) is the referential frame attached to the camera where c is aligned with the optical ais and c and c are aligned with horiontal and vertical aes in the image. The retinal plane is perpendicular to the optical ais. Target is projected at point (, ) in image plane. This point is represented as an homogeneous 3D vector λm = λ(,, ) corresponding to a line of a given direction passing through the optical center (2D projective space). Camera first rotates in pan and then in tilt (Fick model). Camera pan and tilt positions are given b α p and α t (the rotation angles around and c ). R p and R t are the pan and tilt rotation matrices s s ψ u F A = s v = F ()

3 A is the matri of the intrinsic parameters in the projection (see equation ). s and s stand for the scaling along the horiontal and vertical aes of the image plane, s ψ gives the skew between the aes and (u, v ) are the principal point coordinates. Notice that A is an upper triangular matri, thus it is alwas invertible. It is assumed the aes are orthogonal (s ψ = ), the optical ais intersects the retinal plane at the origin of the image referential (u = v = ) and s = s = F. 2. Position equations m = AR t t (α t)r t p (α p)m (2) Equation 2 epresses the relationship between target 3D coordinates M in the inertial frame and projective coordinates m in camera reference frame. Notice is target Z coordinate in camera referential, m = (,, ) t the projective coordinates and (, ) target coordinates in image plane. R p( α p )R t ( α t )R t (α t )A m = R t (α t )A m (3) Camera moves α p in pan and α t in tilt. New camera rotation angles are (α p + α p, α t + α t ). Equation 3 establishes the relationship between m, the actual target projective coordinates, and m, the projective coordinates before camera motion. Consider α p and α t are the tracking angular position errors in pand and tilt. If target spherical coordinates are (ρ, θ, φ) (see Fig. 2) then α p = θ α p and α t = φ α t. m = = (F C(α t ) S(α t )) tan( α p ) F 2(S(αt)2 C( α p)+c(α t) 2 ) tan( α t) S(2α t)(c( α p) ) 2(S(α t) 2 +C(α t) 2 C( α p)) S(2α t)(c( α p) )) tan( α t) (4) Whenever camera moves to compensate the angular position errors, the target is projected in the center of the image and m becomes equal to (,, ) t. Equation 4 is derived from 3 making m = (,, ) t. It establishes the relationship between camera tilt position α t, target position in image (, ) and angular position errors ( α p, α t ). F cos(α t ) tan( α p ) m = ỹ = F tan( α t ) (5) oth and epressions are comple and highl non-linear. If the target is being tracked b the active vision sstem, it is reasonable to assume that most of time the target image is nearl in the center. This assumption is used to derive equation 5, a simplification of 4. Using 5 pan and tilt control can be decoupled. position is onl related with pan error and position is related with tilt error. E E pos = pos = (6) ỹ Φ pos = (σ pos ) 2 φ pos φ pos (σpos) 2 = E pos µ N 2 µ pos = pos = E pos (i)p (i) (7) µ pos N 2 i= (E pos (i) µ pos )(E pos (i) µ pos ) t P (i) (8) i= Camera angular position (α p, α t ) is obtained b reading motor encoders. Visual processing determines target position in image (, ). The goal is to move the camera to compensate for the angular errors ( α p, α t ). Equation 4 performs the eact computation of ( α p, α t ) given (, ). E pos is the error in approimating equation 4 b 5 (see equation 6). E pos is a function of α t, α p and α t. Consider α t, α p, α t [ 45, 45 ] and the interval discretied in N samples uniforml spaced. For each camera tilt position α t there are N 2 possible combinations for ( α p, α t ). The pan and tilt angular errors are assumed to be statisticall independent. P (i) is the joint probabilit function of ( α p, α t ). Fiate α t, for each pair ( α p, α t ) there is a corresponding error vector E pos. Equation 7 computes the average error µ pos for a certain camera tilt position α t. Epression 8 computes the covariance matri Φ pos.

4 .5 Average Error.25 Error Standard Deviation and Covariance..2 Std. Dev. X.5.5 Std. Dev. Y X ais Y ais Covariance Camera tilt position (deg) Camera tilt position (deg) Fig. 3.: Quantitative analsis of error function E pos and E pos assuming a normal probabilit distribution for the angular errors in pan and tilt. Above: average error in X (-) and Y (- -). elow: error standard deviation in X (-) and Y (- -). If the sstem is tracking the target then, most of time, target image is not far from the center. It is reasonable to assume that both pan and tilt errors ( α p and α t ) are statisticall described b a normal distribution of average. P (i), in equations 7 and 8, is a bidimensional probabilit densit function for a normal distribution with ero average and standard deviation of 2 in pan and 8 in tilt. Fig. 3 depicts E pos average, standard deviation and covariance as a function of α t. The average error in X and the covariance φ pos are ero and µ pos is an odd function. oth σpos and σpos (standard deviation in X and Y) increase with the module of camera tilt angle. The data depicted in Fig. 4 is useful to understand the behavior of the error function E pos. Different camera tilt positions were studied. First column is for α t = 23, the second for α t = and the third for α t = 23. Whenever α t is known target position in image depends on the angular pan and tilt errors ( α p, α t ). First row depicts the eact and approimated target positions in image for different tilt angles. The second row ehibits the corresponding error X coordinate Epos. The third row shows the error in Y ais Epos. Observe the eact and estimated positions of target projection in image (first row). Assume the angular pan error α p constant. In 3D space target is positioned somewhere in a vertical plane going trough the origin O of the inertial referential frame (see Fig. 2). The plane is projected in a line in the image. If α t = the line is vertical and if α t the line has a slope whose module is inversel proportional to the module of camera tilt angle. The approimation of equation 5 alwas projects the plane in a vertical line in the image. Thus, as depicted in Fig. 4, = whenever α t = or =. The approimation error in X ais Epos is a function of α t, α p and α t. Notice that: E pos(, α p, α t ) = E pos(α t, α p, α t ) = E (α t, α p, α t ) E pos(α t, α p, α t ) = E ( α t, α p, α t ) Consider the angular tilt error α t constant. In 3D space the target is somewhere in a conic surface whose vertice is the origin O of the inertial referential frame (see Fig. 2). The image projection of these surfaces are the hiperbolic lines depicted in Fig. 4 (first row). Whenever α t = α t the conic surface degenerates in the OXZ plane which is projected in an horiontal line. The approimation of equation 5 generates horiontal lines in image. Therefore ỹ = whenever α t = alpha t or α p =. Some properties of E pos, observable in Fig. 4, are itemied below: E pos(α t,, α t ) = E pos(α t, α p, α t ) = E pos(α t, α p, α t ) = E (α t, α p, α t ) E pos(α t, α p, α t ) = E ( α t, α p, α t ) The approimation error E pos = [E pos, E pos] t increases with both angular position errors ( α p, α t ) and camera tilt position. If camera tilt angle α t take a great value the approimation error becomes signif-

5 Camera tilt position 23º.5 Camera tilt position º.5 Camera tilt position 23º Te=45º Te=45º Te=45º.5 Te=23º.5 Te=23º.5 Te=23º Pe= 45º Pe= 23º Pe=º Pe=23º Pe=45º Pe= 45º Pe= 23º Pe=º Pe=23º Pe=45º Image Y ais Te=º Te= 23º Image Y ais Te=º Te= 23º Image Y ais Te=º Pe= 45º Pe= 23º Pe=º Pe=23º Pe=45º Te= 23º Te= 45º Pe=Pan Error Te=Tilt Error Te= 45º Pe=Pan Error Te=Tilt Error Te= 45º Pe=Pan Error Te=Tilt Error.5 Image X ais Image X ais Image X ais Error in X. Camera tilt angle 23º Error in X. Camera tilt angle º Error in X. Camera tilt angle 23º Error in Y. Camera tilt angle 23º Error in Y. Camera tilt angle º Error in Y. Camera tilt angle 23º Fig. 4.: Approimation error E pos (E pos, E pos) assuming F =. Each column corresponds to a certain camera tilt position: first column α t = 23, second column α t = and third column α t = 23. First row: target projection in image (eact (o) and approimated (*) position). Second row: E pos as a function of ( α p, α t ). Third row: E pos as a function of ( α p, α t ). icant even if the target is projected near the center of the image. Therefore to approimate equation 4 b equation 5, the operation range of tilt degree of freedom can not be large. This last result ields important conclusions reagarding the mechanical construction of an active tracking sstem. We have been assuming a Fick model of rotation (pan and tilt). Our sstems have been designed for tracking targets moving in the ground plane (see Fig ). The pan operating range is much larger than tilt operating range. The values of α t are alwas less than 23 (the mechanical limit for the tilt degree of freedom) and the errors in assuming the approimations of equation 5 are small. Consider a sstem, with the same purposes, designed with the camera first rotating in tilt and then in pan (Helmholt model). Assuming this α t would be replaced b α p in the derived equations. The errors in using the approimations would be more significant because the operation range of α p would be larger. Moreover the sstem mechanical construction would involve additional difficulties. It is alwas easier to achieve large operation ranges in the independent rotation than in the dependent one. For these reasons it is advisable to design the sstem with the camera rotating first around the angle where larger operation ranges are requested. 2.2 Velocit Equations This section derives the mathematical epressions for the target velocit in the image. Velocit in the image depends on camera motion and target 3D velocit. Camera motion induces velocit in the image even when the scene is static. This self-induced motion is called in the literature egomotion. Target motion in 3D space also induces motion in image. Tracking is achieved when camera moves in such a wa that egomotion cancels

6 out the velocit induced b motion in the scene and target is kept in the same position in successive frames. In this paper it is assumed that camera intrinsic parameters are kept constant along time (Ȧ = ). żm + ṁ = AR t t(α t)rp t (α p)m + ARt t(α t) R p t (α p)m + ARt t(α t)rp t (α p)ṁ (9) K =. () Equation 9 is derived b differentiating 2. ṁ = [ẋ, ẏ, ] t where [ẋ, ẏ] t is the velocit vector in image. ż is target velocit along Z ais in camera reference frame. In order to derive image velocit vector, matri K is defined in such a wa that (ẋ, ẏ) t = K(żm + ṁ). ẋ = KAP ẏ t A mα t + KARt t(α t)p p R t (α t )A mα p + KARt t (α t)rp t (α p)ṁ () Equation is derived from 9. P p and P t are the diferential generators of the abelean groups R t p() and R t t(). Velocit in image (ẋ, ẏ) t both depends on camera velocit of motion ( α t ) (egomotion) and 3D velocit in the scene Ṁ. The egomotion terms do not depend on scene 3D coordinates because camera describes pure rotation motion. ẋ ẏ αp α t [ = W ego ] θ + W tgt φ The target velocit in 3D space can be described in rectangular (Ẋ, Ẏ, Ż) or spherical ( ρ, θ, φ) coordinates. When the target motion is described in rectangular coordinates, velocit in the image depends on Ẋ, Ẏ and Ż. On the other hand, whenever spherical coordinates are used velocit in the image depends onl on θ and φ. This happens because ρ is the velocit component along the projective ra that can not be observed in the image (for a point target). Spherical coordinates simplif the derivation and understanding of velocit equations. Equation 2 ields the target velocit in the image. The first term refers to egomotion, showing the contribution of camera motion to image velocit. Matri W ego is a 22 weighting matri that depends on the image point coordinates and camera tilt angle. The second term refers to the velocit in the image due to the target motion in the scene. W tgt is a 22 matri which is a function of, and α t. oth W ego and W tgt are easil computed from equation. Fig. 5 depicts the velocit fields induced in image. The first two rows show the velocities due to camera pan and tilt motion (egomotion). The third and fourth rows ehibit the velocit fields induced b target motion in space. αp θ = W α ego W tgt = (F S(α t)+c(α t)) (S(α t) F C(α t)) 2 +(F C(α t) S(α t)) 2 θ (3) t φ 2 +(F C(α t) S(α t)) 2 φ S(α t) F C(α t) (2) The goal in a tracking application is to move the camera in a such a wa that egomotion compensate the image velocit induced b motion in the scene. Assuming that the target velocit is ( θ, φ), perfect tracking is achieved whenever camera pan and tilt velocities are computed b equation 3. W ego is a singular matri F tan(α t) onl when =. This happens whenever target projection las on an horiontal line that contains the intersection point of the pan rotation ais with the image plane. That is the onl case in which perfect velocit regulation can not be achieved. Notice that if φ =, perfect tracking is achieved b making α p = θ and α t =. The velocit induced in the image b the target motion can be compensated for b camera pan rotation. However if α t the camera must move in pan and tilt to keep target position in image. For this case, the residual velocit b making α p = and α t = φ can be observed in Fig. 5 (last row)). [ ẋ F cos(αt )( α p θ) ] ẏ F ( α t φ) (4) Similarl to what was done for position analsis, equation 2 is simplified to decouple pan and tilt control. Equation 2 is obtained from equation 4 assuming (, ) = (, ). The image velocit in Y ais onl depends on camera tilt velocit and φ and perfect tracking is achieved whenever α t = φ. The same happens for the X ais, camera pan velocit and θ.

7 Velocit in Image (pan motion; camera tilt angle 23º) 5 Velocit in Image (pan motion; camera tilt angle º) 5 Velocit in Image (pan motion; camera tilt angle 23º) Velocit in Image (tilt motion; camera tilt angle 23º) 5 Velocit in Image (tilt motion; camera tilt angle º) 5 Velocit in Image (tilt motion; camera tilt angle 23º) Velocit in Image (target pan motion; camera tilt angle 23º) 5 Velocit in Image (target pan motion; camera tilt angle º) 5 Velocit in Image (target pan motion; camera tilt angle 23º) Velocit in Image (target tilt motion; camera tilt angle 23º) 5 Velocit in Image (target tilt motion; camera tilt angle º) 5 Velocit in Image (target tilt motion; camera tilt angle 23º) Velocit in Image (target+camera tilt motion; camera tilt angle 23º) 5 Velocit in Image (target+camera tilt motion; camera tilt angle º) 5 Velocit in Image (target+camera tilt motion; camera tilt angle 23º) Fig. 5.: Velocit field in image. Each column corresponds to a certain camera tilt position: first column α t = 23, second column α t = and third column α t = 23. First row: ( α t ) = (, )(rad/s) and ( θ, φ) = (, )(rad/s). Second row: ( α t ) = (, )(rad/s) and ( θ, φ) = (, )(rad/s). Third row: ( α t ) = (, )(rad/s) and ( θ, φ) = (, )(rad/s). Fourth row: ( α t ) = (, )(rad/s) and ( θ, φ) = (, )(rad/s). Fifth row: ( α t ) = (, )(rad/s) and ( θ, φ) = (, )(rad/s). Large rectangle (-): image with a field of view of Small rectangle (:): image with a field of view of 24 8.

8 i r r (X,Y,Z) α t r α p c c i O c l l i Right Camera Cclopean Camera Left Camera l TOP VIEW r r c c (X,Y,Z) - β β Right Camera Cclopean Camera Left Camera l l Fig. 6.: inocular Sstem. Left:Neck pan and tilt. Right: Vergence Control 2.3 inocular Tracking In previous section the position and velocit equations for monocular tracking were derived. In this section we focus on the equations for binocular tracking. As depicted in figure 6 two cameras are mounted in a moving platform with length 2 (the baseline). The platform is able to rotate in pan (α p angle) and tilt (α t angle). Each ee has and additional pan degree of freedom called vergence (β L and β R angles). It is assumed that vergence is smmetric (β L = β R = β). inocular tracking is achieved b controlling the three independent degrees of freedom: pan (α p ), tilt (α t ) and vergence (β). Imagine a third camera positioned in the center of the platform called the cclopean ee. Neck pan and tilt degrees of freedom align cclopean Z (forward-looking) ais with target. Vergence control adjusts both camera positions so that target images are projected in the corresponding image centers. Assuming that target is foveated with smmetric vergence angles, these onl depend on target motion along the cclopean Z (forward-looking) ais i.e. on the cclopean depth of the object. Therefore binocular tracking can be split in two sub-problems: cclopean ee control and vergence control. Tracking the target with the cclopean ee is essentiall a monocular tracking problem as studied in last section. To perform such a tracking it is necessar to transfer visual information from both retinas to the cclopean camera. Vergence control is achieved using retinal flow disparit. 2.4 Pan and Tilt Control { l m l = AR t p (β)rt t (α t)r t p (α p)m + AR t p (β)t r m r = AR t p ( β)rt t (α t)r t p (α p)m AR t p ( β)t (5) Vector M represents target cartesian coordinates in the cclopean inertial frame. Target is projected in the left image at ( l, l ) and in the right image at ( r, r ). The corresponding homogeneous vectors are λm l and λm r. l and r are target Z coordinates in left and right camera reference frames. Equation 5 establishes the relationship between m l, m r and M. α p and α t are platform pan and tilt angles, β is the smmetric vergence angle and t = (,, ) t where 2 is the baseline. { l m l = AR t p (β)a m + AR t p (β)t r m r = AR t p ( β)a m AR t p ( β)t (6) Assume (, ) are target projection coordinates in the cclopean ee, λm is the corresponding homogeneous vector and represents target cclopean depth. Equation 6 is derived from 5 using 2. The goal of the neck pan and tilt control is to align cclopean Z ais (forward-looking) with the moving target. This task is similar to the monocular tracking problem. It is necessar to compute target position and velocit in the cclopean ee from the position and velocit measurements in left and right image. = [ F 2 ( l + r) 2 S(β) 2 l r F 2 C(β) 2 +F S(β)C(β)( l r) F 2 C(β)( l + r)+f S(β)( l r r l ) 2 S(β) 2 l r F 2 C(β) 2 +F S(β)C(β)( l r) ] (7)

9 Equation 6 is a sstem of si non-linear equations. Assume target image coordinates are known in both left and right camera (( l, l ) and ( r, r )). There are five unknowns in 6: target position in the cclopean image (, ) and target Z coordinate in left, right and cclopean camera referential ( l, r and ). These five unknowns can be determined b solving the sstem of equations. The solution for (, ) is given in 7. ẋ = ±K ẏ AP prp(±β)a t m β + K ARp(±β)P t tr p(±β)a m α t +K ARp(±β)R t t(α t t)p pr t(α t)r p(±β)a m α p K AR t p(±β)p ttα t K AR t p(±β)rt(α t t)p pr t(α t)tα p + K AR t p(±β)rt(α t t)rp(α t p)ṁ (8) Equation 8 computes the velocit in left and right retina. It is obtained b differentiating equation 5. The first five terms correspond to velocit induced b camera motion (egomotion). Notice that fourth and fifth term depend on scene depth because camera performs translational motion. The last term correspond to the image velocit induced b target motion in space. β ẋego = W ẏ ego rot α t + Wego tra αt (9) ego α p α p Wego tra = [ (F S(β) ± C(β))C(α t ) ±(F S(β) + C(β)C(α t )) Equation 9 computes the image velocit component due to camera motion. Wego rot is a 23 weighting matri corresponding to egomotion induced b camera rotation motion. Wego tra is a 22 weighting matri that is related with egomotion component due to camera translation (see equation 2). is the distance from the camera to neck pan and tilt joints. The camera describes translational motion if > and α p. Whenever is ero or the neck pan degree of freedom is stopped, the second term of equation 9 is ero and egomotion is onl due to camera rotation. The velocit in image induced b camera translation is proportional to the ratio. Usuall target distance to sstem is much larger than baseline length. The ratio is nearl ero and the egomotion induced b camera translation can be neglected. ẋegol /r ± C(αt )( + F S(β)) β ẏ egol /r C(β) ±S(α t )(S(β) + α ) t (2) Assuming that target is nearl in the center of image, equation 9 simplifies to equation 2 b making (, ) = (, ). ẋtgtl = ẏ l K l ARp t (β)a X tgt [ tgtl ] (22) ẋr = r K r ARp t ( β)a X tgt ẏ r Consider (ẋ tgt, ẏ tgt ) t as being the velocit induced in cclopean image b target motion in scene. This velocit vector is (ẋ tgt, ẏ tgt ) t = KX tgt, where X tgt is a 3 vector such as X tgt = AR t t(α t )R t p(α p )Ṁ (see equation ). Assume (ẋ tgtl, ẏ tgtl ) t and (ẋ tgtr, ẏ tgtr ) t are image velocities induced b target motion in left and right retina. 22 is sstem of equations derived from the last term of 8. This sstem is used to obtain an analtical solution for (ẋ tgt, ẏ tgt ) t. [ ẋtgtl +ẋ tgtr ] ẋtgt 2 cos(β) 2 ẏ ẏ tgtl +ẏ tgtr (23) tgt 2 cos(β) Assuming that target is projected in the center of both retinas, the velocit induced b target motion in the cclopean image is given b equation 23. ] α p (2)

10 2.5 Vergence Control Ee pan is called vergence. Figure 6 ehibits a schematic of vergence control. The vergence point is the intersection of left and right optical ais. Assuming smmetric vergence this point is alwas in the cclopean Z ais. The sstem is verged whenever tan(β) = where is target Z coordinate in the cclopean reference frame. This section performs the mathematical stud for vergence control. β = arctan( F cos(β)( l r ) + 2 sin(β) l r F (2F cos(β) sin(β)( l r )) ) (24) Assume the vergence angle is β and β is the correction angle that verges the sstem in the target (tan(β + β) = ). β is called the vergence error in position and is given b equation 24. ẋ = ± K AP p Rp t (±β)(a m ± t) β + K ARp t (±β)a (żm + ṁ) (25) ẏ Equation 25 is derived b differentiating equation 6. It gives the velocit in left and right retina in order to velocities in the cclopean ee. The first term corresponds to egomotion induced b vergence degree of freedom. While neck pan and tilt are used to keep the target aligned with cclopean Z ais, the goal of vergence control is to compensate for target motion along the cclopean optical ais. 2F β ẋ l ẋ r = 2F sin(β)2 ż (26) β = ẋtgt l ẋ tgtr (27) 2F Assume that cclopean ee is performing perfect tracking ((, ) = (, ) and (ẋ, ẏ) = (, )) and that target is projected in the center of both retinas (tan(β) = ). Equation 26 is obtained appling these assumptions to 25. It gives the velocit disparit in X ais that is the sum of an egomotion term and target velocit along cclopean Z ais. To achieve perfect tracking this sum must be ero. If ẋ tgtl ẋ tgtr is the disparit due to target motion in scene, equation 27 gives the vergence velocit to achieve perfect tracking. References. J.L. Crowle, J.M. edrune, M. ekker, and M. Schneider. Integration and control of reactive visual processes. In Third European Conference in Computer Vision, volume 2, pages 47 58, Stockolm,Sweden, Ma J. L. Crowle. Coordination of action and perception in a surveillance robot. IEEE Epert, 2, November Espiau, F. Chaumette, and P. Rives. A new approach to visual servoing in robotics. IEEE Trans. on Robot. and Automat., 8(3):33 326, June H. I. Christensen, J. Horstmann, and T. Rasmussen. A control theoretic approach to active vision. In Asian Conference on Computer Vision, pages 2 2, Singapore, December P.I. Corke. Visual Control of Robots: High-Peformance Visual Servoing. Mechatronics. John Wile, P. Krautgartner and Vince M. Performance evaluation of vision-based control tasks. In ICRA98 IEEE Int. Conf. on Robotics and Automation, pages , Leuven, elgium, P. Nordlund and Uhlin T. Closing the loop: detection and pursuit of a moving object b a moving observer. Image and Vision Computing, 4: , D. Murra, P. McLauchlan, I. Reid, and P. Sharke. Reactions to peripheral image motion using a head/ee plataform. In Proc. of IEEE International Conference on Computer Vision, pages 43 4, 993. This article was processed using the TEX macro package with SIRS2 stle

Tracking Multiple Objects in 3D. Coimbra 3030 Coimbra target projection in the same image position (usually

Tracking Multiple Objects in 3D. Coimbra 3030 Coimbra target projection in the same image position (usually Tracking Multiple Objects in 3D Jo~ao P. Barreto Paulo Peioto Coimbra 33 Coimbra 33 Jorge Batista Helder Araujo Coimbra 33 Coimbra 33 Abstract In this paper a system for tracking multiple targets in 3D

More information

Multibody Motion Estimation and Segmentation from Multiple Central Panoramic Views

Multibody Motion Estimation and Segmentation from Multiple Central Panoramic Views Multibod Motion Estimation and Segmentation from Multiple Central Panoramic Views Omid Shakernia René Vidal Shankar Sastr Department of Electrical Engineering & Computer Sciences Universit of California

More information

Perspective Projection Transformation

Perspective Projection Transformation Perspective Projection Transformation Where does a point of a scene appear in an image?? p p Transformation in 3 steps:. scene coordinates => camera coordinates. projection of camera coordinates into image

More information

NATIONAL UNIVERSITY OF SINGAPORE. (Semester I: 1999/2000) EE4304/ME ROBOTICS. October/November Time Allowed: 2 Hours

NATIONAL UNIVERSITY OF SINGAPORE. (Semester I: 1999/2000) EE4304/ME ROBOTICS. October/November Time Allowed: 2 Hours NATIONAL UNIVERSITY OF SINGAPORE EXAMINATION FOR THE DEGREE OF B.ENG. (Semester I: 1999/000) EE4304/ME445 - ROBOTICS October/November 1999 - Time Allowed: Hours INSTRUCTIONS TO CANDIDATES: 1. This paper

More information

Performance Characterization of Visual Behaviors in an Active Vision System

Performance Characterization of Visual Behaviors in an Active Vision System Performance Characterization of Visual Behaviors in an Active Vision System João P Barreto, Paulo Peixoto, Jorge Batista, Helder Araujo Institute of Systems and Robotics Institute of Systems and Robotics

More information

Announcements. Equation of Perspective Projection. Image Formation and Cameras

Announcements. Equation of Perspective Projection. Image Formation and Cameras Announcements Image ormation and Cameras Introduction to Computer Vision CSE 52 Lecture 4 Read Trucco & Verri: pp. 22-4 Irfanview: http://www.irfanview.com/ is a good Windows utilit for manipulating images.

More information

Visual Tracking of a Hand-eye Robot for a Moving Target Object with Multiple Feature Points: Translational Motion Compensation Approach

Visual Tracking of a Hand-eye Robot for a Moving Target Object with Multiple Feature Points: Translational Motion Compensation Approach Visual Tracking of a Hand-eye Robot for a Moving Target Object with Multiple Feature Points: Translational Motion Compensation Approach Masahide Ito Masaaki Shibata Department of Electrical and Mechanical

More information

FEEDBACK CONTROL ANALYSIS OF A STEREO ACTIVE VISION SYSTEM

FEEDBACK CONTROL ANALYSIS OF A STEREO ACTIVE VISION SYSTEM Bulletin of the ransilvania Universit of Braşov Vol. 8 (57) No. 2-2015 Series I: Engineering Sciences FEEDBACK CONROL ANALYSIS OF A SEREO ACIVE VISION SYSEM G. MĂCEŞANU 1 S.M. GRIGORESCU 1 Abstract: he

More information

MEM380 Applied Autonomous Robots Winter Robot Kinematics

MEM380 Applied Autonomous Robots Winter Robot Kinematics MEM38 Applied Autonomous obots Winter obot Kinematics Coordinate Transformations Motivation Ultimatel, we are interested in the motion of the robot with respect to a global or inertial navigation frame

More information

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017

CS 2770: Intro to Computer Vision. Multiple Views. Prof. Adriana Kovashka University of Pittsburgh March 14, 2017 CS 277: Intro to Computer Vision Multiple Views Prof. Adriana Kovashka Universit of Pittsburgh March 4, 27 Plan for toda Affine and projective image transformations Homographies and image mosaics Stereo

More information

Camera model and multiple view geometry

Camera model and multiple view geometry Chapter Camera model and multiple view geometry Before discussing how D information can be obtained from images it is important to know how images are formed First the camera model is introduced and then

More information

3D Geometry and Camera Calibration

3D Geometry and Camera Calibration 3D Geometr and Camera Calibration 3D Coordinate Sstems Right-handed vs. left-handed 2D Coordinate Sstems ais up vs. ais down Origin at center vs. corner Will often write (u, v) for image coordinates v

More information

Task selection for control of active vision systems

Task selection for control of active vision systems The 29 IEEE/RSJ International Conference on Intelligent Robots and Systems October -5, 29 St. Louis, USA Task selection for control of active vision systems Yasushi Iwatani Abstract This paper discusses

More information

Modeling Transformations

Modeling Transformations Modeling Transformations Michael Kazhdan (601.457/657) HB Ch. 5 FvDFH Ch. 5 Overview Ra-Tracing so far Modeling transformations Ra Tracing Image RaTrace(Camera camera, Scene scene, int width, int heigh,

More information

MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration

MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration MAN-522: COMPUTER VISION SET-2 Projections and Camera Calibration Image formation How are objects in the world captured in an image? Phsical parameters of image formation Geometric Tpe of projection Camera

More information

CMSC 425: Lecture 10 Basics of Skeletal Animation and Kinematics

CMSC 425: Lecture 10 Basics of Skeletal Animation and Kinematics : Lecture Basics of Skeletal Animation and Kinematics Reading: Chapt of Gregor, Game Engine Architecture. The material on kinematics is a simplification of similar concepts developed in the field of robotics,

More information

A Tool for Kinematic Error Analysis of Robots/Active Vision Systems

A Tool for Kinematic Error Analysis of Robots/Active Vision Systems A Tool for Kinematic Error Analysis of Robots/Active Vision Systems Kanglin Xu and George F. Luger Department of Computer Science University of New Mexico Albuquerque, NM 87131 {klxu,luger}@cs.unm.edu

More information

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3

Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Unit 2: Locomotion Kinematics of Wheeled Robots: Part 3 Computer Science 4766/6778 Department of Computer Science Memorial University of Newfoundland January 28, 2014 COMP 4766/6778 (MUN) Kinematics of

More information

Modeling Transformations

Modeling Transformations Modeling Transformations Michael Kazhdan (601.457/657) HB Ch. 5 FvDFH Ch. 5 Announcement Assignment 2 has been posted: Due: 10/24 ASAP: Download the code and make sure it compiles» On windows: just build

More information

Visual Recognition: Image Formation

Visual Recognition: Image Formation Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know

More information

A rigid body free to move in a reference frame will, in the general case, have complex motion, which is simultaneously a combination of rotation and

A rigid body free to move in a reference frame will, in the general case, have complex motion, which is simultaneously a combination of rotation and 050389 - Analtical Elements of Mechanisms Introduction. Degrees of Freedom he number of degrees of freedom (DOF) of a sstem is equal to the number of independent parameters (measurements) that are needed

More information

D-Calib: Calibration Software for Multiple Cameras System

D-Calib: Calibration Software for Multiple Cameras System D-Calib: Calibration Software for Multiple Cameras Sstem uko Uematsu Tomoaki Teshima Hideo Saito Keio Universit okohama Japan {u-ko tomoaki saito}@ozawa.ics.keio.ac.jp Cao Honghua Librar Inc. Japan cao@librar-inc.co.jp

More information

A Full Analytical Solution to the Direct and Inverse Kinematics of the Pentaxis Robot Manipulator

A Full Analytical Solution to the Direct and Inverse Kinematics of the Pentaxis Robot Manipulator A Full Analtical Solution to the Direct and Inverse Kinematics of the Pentais Robot Manipulator Moisés Estrada Castañeda, Luis Tupak Aguilar Bustos, Luis A. Gonále Hernánde Instituto Politécnico Nacional

More information

[ ] [ ] Orthogonal Transformation of Cartesian Coordinates in 2D & 3D. φ = cos 1 1/ φ = tan 1 [ 2 /1]

[ ] [ ] Orthogonal Transformation of Cartesian Coordinates in 2D & 3D. φ = cos 1 1/ φ = tan 1 [ 2 /1] Orthogonal Transformation of Cartesian Coordinates in 2D & 3D A vector is specified b its coordinates, so it is defined relative to a reference frame. The same vector will have different coordinates in

More information

Shortest Path Homography-Based Visual Control for Differential Drive Robots

Shortest Path Homography-Based Visual Control for Differential Drive Robots Citation: G. López-Nicolás, C. Sagüés, and J. J. Guerrero. Vision Systems, Shortest Path Homography-Based Visual Control for Differential Drive Robots, chapter 3, pp. 583 596. Edited by Goro Obinata and

More information

Assignment 2 : Projection and Homography

Assignment 2 : Projection and Homography TECHNISCHE UNIVERSITÄT DRESDEN EINFÜHRUNGSPRAKTIKUM COMPUTER VISION Assignment 2 : Projection and Homography Hassan Abu Alhaija November 7,204 INTRODUCTION In this exercise session we will get a hands-on

More information

Pin Hole Cameras & Warp Functions

Pin Hole Cameras & Warp Functions Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Example of SLAM for AR Taken from:

More information

Abstract. Euler Equations Solutions for Incompressible Fluid Flow. Copyright A. A. Frempong

Abstract. Euler Equations Solutions for Incompressible Fluid Flow. Copyright A. A. Frempong Copright A. A. Frempong Euler Equations Solutions for Incompressible Fluid Flow Abstract The simplest solution is usuall the best solution---albert Einstein This paper covers the solutions of the Euler

More information

A New Concept on Automatic Parking of an Electric Vehicle

A New Concept on Automatic Parking of an Electric Vehicle A New Concept on Automatic Parking of an Electric Vehicle C. CAMUS P. COELHO J.C. QUADRADO Instituto Superior de Engenharia de Lisboa Rua Conselheiro Emídio Navarro PORTUGAL Abstract: - A solution to perform

More information

Announcements. The equation of projection. Image Formation and Cameras

Announcements. The equation of projection. Image Formation and Cameras Announcements Image ormation and Cameras Introduction to Computer Vision CSE 52 Lecture 4 Read Trucco & Verri: pp. 5-4 HW will be on web site tomorrow or Saturda. Irfanview: http://www.irfanview.com/ is

More information

Computer Graphics. Geometric Transformations

Computer Graphics. Geometric Transformations Contents coordinate sstems scalar values, points, vectors, matrices right-handed and left-handed coordinate sstems mathematical foundations transformations mathematical descriptions of geometric changes,

More information

Closure Polynomials for Strips of Tetrahedra

Closure Polynomials for Strips of Tetrahedra Closure Polnomials for Strips of Tetrahedra Federico Thomas and Josep M. Porta Abstract A tetrahedral strip is a tetrahedron-tetrahedron truss where an tetrahedron has two neighbors ecept those in the

More information

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism

Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Behavior Learning for a Mobile Robot with Omnidirectional Vision Enhanced by an Active Zoom Mechanism Sho ji Suzuki, Tatsunori Kato, Minoru Asada, and Koh Hosoda Dept. of Adaptive Machine Systems, Graduate

More information

Computer Graphics. P04 Transformations. Aleksandra Pizurica Ghent University

Computer Graphics. P04 Transformations. Aleksandra Pizurica Ghent University Computer Graphics P4 Transformations Aleksandra Pizurica Ghent Universit Telecommunications and Information Processing Image Processing and Interpretation Group Transformations in computer graphics Goal:

More information

Computer Graphics. Geometric Transformations

Computer Graphics. Geometric Transformations Computer Graphics Geometric Transformations Contents coordinate sstems scalar values, points, vectors, matrices right-handed and left-handed coordinate sstems mathematical foundations transformations mathematical

More information

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives

More information

MEAM 520. Mobile Robots

MEAM 520. Mobile Robots MEAM 520 Mobile Robots Katherine J. Kuchenbecker, Ph.D. General Robotics, Automation, Sensing, and Perception Lab (GRASP) MEAM Department, SEAS, Universit of Pennslvania Lecture 22: December 6, 2012 T

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan

Dept. of Adaptive Machine Systems, Graduate School of Engineering Osaka University, Suita, Osaka , Japan An Application of Vision-Based Learning for a Real Robot in RoboCup - A Goal Keeping Behavior for a Robot with an Omnidirectional Vision and an Embedded Servoing - Sho ji Suzuki 1, Tatsunori Kato 1, Hiroshi

More information

Uses of Transformations. 2D transformations Homogeneous coordinates. Transformations. Transformations. Transformations. Transformations and matrices

Uses of Transformations. 2D transformations Homogeneous coordinates. Transformations. Transformations. Transformations. Transformations and matrices Uses of Transformations 2D transformations Homogeneous coordinates odeling: position and resie parts of a comple model; Viewing: define and position the virtual camera Animation: define how objects move/change

More information

Keeping features in the camera s field of view: a visual servoing strategy

Keeping features in the camera s field of view: a visual servoing strategy Keeping features in the camera s field of view: a visual servoing strategy G. Chesi, K. Hashimoto,D.Prattichizzo,A.Vicino Department of Information Engineering, University of Siena Via Roma 6, 3 Siena,

More information

Lecture 3: Camera Calibration, DLT, SVD

Lecture 3: Camera Calibration, DLT, SVD Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P

More information

3-Dimensional Viewing

3-Dimensional Viewing CHAPTER 6 3-Dimensional Vieing Vieing and projection Objects in orld coordinates are projected on to the vie plane, hich is defined perpendicular to the vieing direction along the v -ais. The to main tpes

More information

Robert Collins CSE486, Penn State. Robert Collins CSE486, Penn State. Image Point Y. O.Camps, PSU. Robert Collins CSE486, Penn State.

Robert Collins CSE486, Penn State. Robert Collins CSE486, Penn State. Image Point Y. O.Camps, PSU. Robert Collins CSE486, Penn State. Stereo Vision Inerring depth rom images taken at the same time b two or more s. Lecture 08: Introduction to Stereo Reading: T&V Section 7.1 Scene Point Image Point p = (,,) O Basic Perspective Projection

More information

On the Kinematics of Undulator Girder Motion

On the Kinematics of Undulator Girder Motion LCLS-TN-09-02 On the Kinematics of Undulator Girder Motion J. Welch March 23, 2010 The theor of rigid bod kinematics is used to derive equations that govern the control and measurement of the position

More information

Stability and performance of image based visual servo control using first order spherical image moments

Stability and performance of image based visual servo control using first order spherical image moments Stabilit and performance of image based visual servo control using first order spherical image moments Odile Bourquarde, Robert Mahon, Tarek Hamel and François Chaumette IRISA - CNRS and INRIA Rennes,

More information

Homography-Based Multi-Robot Control with a Flying Camera

Homography-Based Multi-Robot Control with a Flying Camera Citation: G. López-Nicolás, Y. Mezouar, and C. Sagüés. Homography-based multi-robot control with a flying camera. In IEEE International Conference on Robotics and Automation, pp. 4492-4497, Shangai, China,

More information

LUMS Mine Detector Project

LUMS Mine Detector Project LUMS Mine Detector Project Using visual information to control a robot (Hutchinson et al. 1996). Vision may or may not be used in the feedback loop. Visual (image based) features such as points, lines

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

442 IEEE TRANSACTIONS ON ROBOTICS, VOL. 26, NO. 3, JUNE Active Stereo Tracking of N 3 Targets Using Line Scan Cameras

442 IEEE TRANSACTIONS ON ROBOTICS, VOL. 26, NO. 3, JUNE Active Stereo Tracking of N 3 Targets Using Line Scan Cameras 442 IEEE TRANSACTIONS ON ROBOTICS, VOL. 26, NO. 3, JUNE 200 Active Stereo Tracking of N 3 Targets Using Line Scan Cameras Joao P. Barreto, Member, IEEE, Luis Perdigoto, Student Member, IEEE, Rui Caseiro,

More information

Real-Time Visual Behaviors with a Binocular Active Vision System. Jorge Batista Paulo Peixoto Helder Araujo. ISR - Institute of Systems and Robotics

Real-Time Visual Behaviors with a Binocular Active Vision System. Jorge Batista Paulo Peixoto Helder Araujo. ISR - Institute of Systems and Robotics Real-Time Visual Behaviors with a Binocular Active Vision System Jorge Batista Paulo Peixoto Helder Araujo ISR - Institute of Systems and Robotics Dep. of Electrical Engineering - University of oimbra

More information

Lecture 4: Viewing. Topics:

Lecture 4: Viewing. Topics: Lecture 4: Viewing Topics: 1. Classical viewing 2. Positioning the camera 3. Perspective and orthogonal projections 4. Perspective and orthogonal projections in OpenGL 5. Perspective and orthogonal projection

More information

Pin Hole Cameras & Warp Functions

Pin Hole Cameras & Warp Functions Pin Hole Cameras & Warp Functions Instructor - Simon Lucey 16-423 - Designing Computer Vision Apps Today Pinhole Camera. Homogenous Coordinates. Planar Warp Functions. Motivation Taken from: http://img.gawkerassets.com/img/18w7i1umpzoa9jpg/original.jpg

More information

Short on camera geometry and camera calibration

Short on camera geometry and camera calibration Short on camera geometry and camera calibration Maria Magnusson, maria.magnusson@liu.se Computer Vision Laboratory, Department of Electrical Engineering, Linköping University, Sweden Report No: LiTH-ISY-R-3070

More information

New shortest-path approaches to visual servoing

New shortest-path approaches to visual servoing New shortest-path approaches to visual servoing Ville Laboratory of Information rocessing Lappeenranta University of Technology Lappeenranta, Finland kyrki@lut.fi Danica Kragic and Henrik I. Christensen

More information

Singularity Handling on Puma in Operational Space Formulation

Singularity Handling on Puma in Operational Space Formulation Singularity Handling on Puma in Operational Space Formulation Denny Oetomo, Marcelo Ang Jr. National University of Singapore Singapore d oetomo@yahoo.com mpeangh@nus.edu.sg Ser Yong Lim Gintic Institute

More information

3D Computer Vision II. Reminder Projective Geometry, Transformations. Nassir Navab. October 27, 2009

3D Computer Vision II. Reminder Projective Geometry, Transformations. Nassir Navab. October 27, 2009 3D Computer Vision II Reminder Projective Geometr, Transformations Nassir Navab based on a course given at UNC b Marc Pollefes & the book Multiple View Geometr b Hartle & Zisserman October 27, 29 2D Transformations

More information

A Practical Camera Calibration System on Mobile Phones

A Practical Camera Calibration System on Mobile Phones Advanced Science and echnolog Letters Vol.7 (Culture and Contents echnolog 0), pp.6-0 http://dx.doi.org/0.57/astl.0.7. A Practical Camera Calibration Sstem on Mobile Phones Lu Bo, aegkeun hangbo Department

More information

Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the

Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the Camera Projection Models We will introduce different camera projection models that relate the location of an image point to the coordinates of the corresponding 3D points. The projection models include:

More information

Zoom control to compensate camera translation within a robot egomotion estimation approach

Zoom control to compensate camera translation within a robot egomotion estimation approach Zoom control to compensate camera translation within a robot egomotion estimation approach Guillem Alenyà and Carme Torras Institut de Robòtica i Informàtica Industrial (CSIC-UPC), Llorens i Artigas 4-6,

More information

Geometry of a single camera. Odilon Redon, Cyclops, 1914

Geometry of a single camera. Odilon Redon, Cyclops, 1914 Geometr o a single camera Odilon Redon, Cclops, 94 Our goal: Recover o 3D structure Recover o structure rom one image is inherentl ambiguous??? Single-view ambiguit Single-view ambiguit Rashad Alakbarov

More information

Centre for Autonomous Systems

Centre for Autonomous Systems Robot Henrik I Centre for Autonomous Systems Kungl Tekniska Högskolan hic@kth.se 27th April 2005 Outline 1 duction 2 Kinematic and Constraints 3 Mobile Robot 4 Mobile Robot 5 Beyond Basic 6 Kinematic 7

More information

Robot Vision Control of robot motion from video. M. Jagersand

Robot Vision Control of robot motion from video. M. Jagersand Robot Vision Control of robot motion from video M. Jagersand Vision-Based Control (Visual Servoing) Initial Image User Desired Image Vision-Based Control (Visual Servoing) : Current Image Features : Desired

More information

CIS 580, Machine Perception, Spring 2016 Homework 2 Due: :59AM

CIS 580, Machine Perception, Spring 2016 Homework 2 Due: :59AM CIS 580, Machine Perception, Spring 2016 Homework 2 Due: 2015.02.24. 11:59AM Instructions. Submit your answers in PDF form to Canvas. This is an individual assignment. 1 Recover camera orientation By observing

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Transformations II. Week 2, Wed Jan 17

Transformations II. Week 2, Wed Jan 17 Universit of British Columbia CPSC 34 Computer Graphics Jan-Apr 27 Tamara Munzner Transformations II Week 2, Wed Jan 7 http://www.ugrad.cs.ubc.ca/~cs34/vjan27 Readings for Jan 5-22 FCG Chap 6 Transformation

More information

arxiv: v1 [cs.cv] 18 Sep 2017

arxiv: v1 [cs.cv] 18 Sep 2017 Direct Pose Estimation with a Monocular Camera Darius Burschka and Elmar Mair arxiv:1709.05815v1 [cs.cv] 18 Sep 2017 Department of Informatics Technische Universität München, Germany {burschka elmar.mair}@mytum.de

More information

CS770/870 Spring 2017 Transformations

CS770/870 Spring 2017 Transformations CS770/870 Spring 2017 Transformations Coordinate sstems 2D Transformations Homogeneous coordinates Matrices, vectors, points Coordinate Sstems Coordinate sstems used in graphics Screen coordinates: the

More information

Position-Based Navigation Using Multiple Homographies

Position-Based Navigation Using Multiple Homographies Position-Based Navigation Using Multiple Homographies Eduardo Montijano DIIS - IA University of Zaragoza, Spain emonti@unizar.es Carlos Sagues DIIS - IA University of Zaragoza, Spain csagues@unizar.es

More information

Geometry of image formation

Geometry of image formation Geometr of image formation Tomáš Svoboda, svoboda@cmp.felk.cvut.c ech Technical Universit in Prague, enter for Machine Perception http://cmp.felk.cvut.c Last update: November 0, 2008 Talk Outline Pinhole

More information

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG.

Computer Vision. Coordinates. Prof. Flávio Cardeal DECOM / CEFET- MG. Computer Vision Coordinates Prof. Flávio Cardeal DECOM / CEFET- MG cardeal@decom.cefetmg.br Abstract This lecture discusses world coordinates and homogeneous coordinates, as well as provides an overview

More information

Epipolar Constraint. Epipolar Lines. Epipolar Geometry. Another look (with math).

Epipolar Constraint. Epipolar Lines. Epipolar Geometry. Another look (with math). Epipolar Constraint Epipolar Lines Potential 3d points Red point - fied => Blue point lies on a line There are 3 degrees of freedom in the position of a point in space; there are four DOF for image points

More information

Chap 7, 2009 Spring Yeong Gil Shin

Chap 7, 2009 Spring Yeong Gil Shin Three-Dimensional i Viewingi Chap 7, 29 Spring Yeong Gil Shin Viewing i Pipeline H d fi i d? How to define a window? How to project onto the window? Rendering "Create a picture (in a snthetic camera) Specification

More information

Introduction to Homogeneous Transformations & Robot Kinematics

Introduction to Homogeneous Transformations & Robot Kinematics Introduction to Homogeneous Transformations & Robot Kinematics Jennifer Ka, Rowan Universit Computer Science Department Januar 25. Drawing Dimensional Frames in 2 Dimensions We will be working in -D coordinates,

More information

CS F-07 Objects in 2D 1

CS F-07 Objects in 2D 1 CS420-2010F-07 Objects in 2D 1 07-0: Representing Polgons We want to represent a simple polgon Triangle, rectangle, square, etc Assume for the moment our game onl uses these simple shapes No curves for

More information

Optical flow Estimation using Fractional Quaternion Wavelet Transform

Optical flow Estimation using Fractional Quaternion Wavelet Transform 2012 International Conference on Industrial and Intelligent Information (ICIII 2012) IPCSIT vol.31 (2012) (2012) IACSIT Press, Singapore Optical flow Estimation using Fractional Quaternion Wavelet Transform

More information

Improving Vision-Based Distance Measurements using Reference Objects

Improving Vision-Based Distance Measurements using Reference Objects Improving Vision-Based Distance Measurements using Reference Objects Matthias Jüngel, Heinrich Mellmann, and Michael Spranger Humboldt-Universität zu Berlin, Künstliche Intelligenz Unter den Linden 6,

More information

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang

CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS. Cha Zhang and Zhengyou Zhang CALIBRATION BETWEEN DEPTH AND COLOR SENSORS FOR COMMODITY DEPTH CAMERAS Cha Zhang and Zhengyou Zhang Communication and Collaboration Systems Group, Microsoft Research {chazhang, zhang}@microsoft.com ABSTRACT

More information

1. We ll look at: Types of geometrical transformation. Vector and matrix representations

1. We ll look at: Types of geometrical transformation. Vector and matrix representations Tob Howard COMP272 Computer Graphics and Image Processing 3: Transformations Tob.Howard@manchester.ac.uk Introduction We ll look at: Tpes of geometrical transformation Vector and matri representations

More information

What and Why Transformations?

What and Why Transformations? 2D transformations What and Wh Transformations? What? : The geometrical changes of an object from a current state to modified state. Changing an object s position (translation), orientation (rotation)

More information

The Graph of an Equation

The Graph of an Equation 60_0P0.qd //0 :6 PM Page CHAPTER P Preparation for Calculus Archive Photos Section P. RENÉ DESCARTES (96 60) Descartes made man contributions to philosoph, science, and mathematics. The idea of representing

More information

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy 1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Epipolar Geometry and the Essential Matrix

Epipolar Geometry and the Essential Matrix Epipolar Geometry and the Essential Matrix Carlo Tomasi The epipolar geometry of a pair of cameras expresses the fundamental relationship between any two corresponding points in the two image planes, and

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Henrik Malm, Anders Heyden Centre for Mathematical Sciences, Lund University Box 118, SE-221 00 Lund, Sweden email: henrik,heyden@maths.lth.se Abstract. In this

More information

Dynamic Stability of Off-Road Vehicles: Quasi-3D Analysis

Dynamic Stability of Off-Road Vehicles: Quasi-3D Analysis 28 IEEE International Conference on Robotics and Automation Pasadena, CA, USA, Ma 9-23, 28 Dnamic Stabilit of Off-Road Vehicles: Quasi-3D Analsis Moshe Mann and Zvi Shiller Abstract This paper presents

More information

Visualisation Pipeline : The Virtual Camera

Visualisation Pipeline : The Virtual Camera Visualisation Pipeline : The Virtual Camera The Graphics Pipeline 3D Pipeline The Virtual Camera The Camera is defined by using a parallelepiped as a view volume with two of the walls used as the near

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

Quaternion-Based Tracking of Multiple Objects in Synchronized Videos

Quaternion-Based Tracking of Multiple Objects in Synchronized Videos Quaternion-Based Tracking of Multiple Objects in Synchronized Videos Quming Zhou 1, Jihun Park 2, and J.K. Aggarwal 1 1 Department of Electrical and Computer Engineering The University of Texas at Austin

More information

Introduction to Homogeneous Transformations & Robot Kinematics

Introduction to Homogeneous Transformations & Robot Kinematics Introduction to Homogeneous Transformations & Robot Kinematics Jennifer Ka Rowan Universit Computer Science Department. Drawing Dimensional Frames in 2 Dimensions We will be working in -D coordinates,

More information

Lecture 12 Date:

Lecture 12 Date: Information Technolog Delhi Lecture 2 Date: 3.09.204 The Signal Flow Graph Information Technolog Delhi Signal Flow Graph Q: Using individual device scattering parameters to analze a comple microwave network

More information

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors We are IntechOpen, the world s leading publisher of Open Access books Built b scientists, for scientists 3,8 116, 12M Open access books available International authors and editors Downloads Our authors

More information

Approaches to Simulate the Operation of the Bending Machine

Approaches to Simulate the Operation of the Bending Machine Approaches to Simulate the Operation of the Bending Machine Ornwadee Ratanapinunchai Chanunchida Theerasilp Teerut Suksawai Project Advisors: Dr. Krit Chongsrid and Dr. Waree Kongprawechnon Department

More information

A Robust Two Feature Points Based Depth Estimation Method 1)

A Robust Two Feature Points Based Depth Estimation Method 1) Vol.31, No.5 ACTA AUTOMATICA SINICA September, 2005 A Robust Two Feature Points Based Depth Estimation Method 1) ZHONG Zhi-Guang YI Jian-Qiang ZHAO Dong-Bin (Laboratory of Complex Systems and Intelligence

More information

3D Sensing. Translation and Scaling in 3D. Rotation about Arbitrary Axis. Rotation in 3D is about an axis

3D Sensing. Translation and Scaling in 3D. Rotation about Arbitrary Axis. Rotation in 3D is about an axis 3D Sensing Camera Model: Recall there are 5 Different Frames of Reference c Camera Model and 3D Transformations Camera Calibration (Tsai s Method) Depth from General Stereo (overview) Pose Estimation from

More information

25 Hz. 25 Hz. X M(z) u c + u X. C(z) u ff KF. Θ, α. V(z) Θ m. 166 Hz Θ. Θ mot. C(z) Θ ref. C(z) V(z) 25 Hz. V(z) Θ, α. Interpolator.

25 Hz. 25 Hz. X M(z) u c + u X. C(z) u ff KF. Θ, α. V(z) Θ m. 166 Hz Θ. Θ mot. C(z) Θ ref. C(z) V(z) 25 Hz. V(z) Θ, α. Interpolator. Comparison of Control Structures for Visual Servoing S. Chroust y, J.P. Barreto z H. Araújo z, M. Vincze y y Vienna University of Technology Institute of Flexible Automation Gusshausstr. 27-20 /361 fsc,vmg@infa.tuwien.ac.at

More information

Module 4F12: Computer Vision and Robotics Solutions to Examples Paper 2

Module 4F12: Computer Vision and Robotics Solutions to Examples Paper 2 Engineering Tripos Part IIB FOURTH YEAR Module 4F2: Computer Vision and Robotics Solutions to Examples Paper 2. Perspective projection and vanishing points (a) Consider a line in 3D space, defined in camera-centered

More information

6-dof Eye-Vergence Visual Servoing by 1-step GA Pose Tracking

6-dof Eye-Vergence Visual Servoing by 1-step GA Pose Tracking International Journal of Applied Electromagnetics and Mechanics 41 (214) 1 1 IOS Press 6-dof Ee-Vergence Visual Servoing b 1-step GA Pose Tracking Yu Cui a, Kenta Nishimura a, Yusuke Sunami a, Mamoru Minami

More information

COSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor

COSC579: Scene Geometry. Jeremy Bolton, PhD Assistant Teaching Professor COSC579: Scene Geometry Jeremy Bolton, PhD Assistant Teaching Professor Overview Linear Algebra Review Homogeneous vs non-homogeneous representations Projections and Transformations Scene Geometry The

More information