3D Face Reconstruction from Stereo Images

Size: px
Start display at page:

Download "3D Face Reconstruction from Stereo Images"

Transcription

1 3D Face Reconstruction from Stereo Images Stefano Arca, Elena Casiraghi, Giuseppe Lipori, Gabriele Lombardi Università degli Studi di Milano, Dipartimento di Scienze dell Informazione, Via Comelico 39/41, Milano {arca, casiraghi, lipori, Abstract In this paper we propose a system for 3D face reconstruction from 2D stereo images. Two different stereo matching algorithms are presented; the first is a local method that finds the correspondences of 15 most characteristic facial fiducial points, whose 3D reconstruction allows us to deform a parameterized sparse model. The second method employs a snake grid algorithm which minimizes a multi-resolution energy function; it defines a dense disparity map and determines a dense 3D face model. Besides we investigate how 3D features affect the recognition performance of a face recognition system working on 2D images. 1. Introduction In the last few decades a great deal of research work has been devoted to the development of algorithms for image rectification, 3D modelling and reconstruction. In [1] we presented a face recognition system working on 2D images, whose recognition performance might be improved by the highly informative biometric measures that can be extrapolated by the 3D face representation; based on this consideration we present in this paper a 3D face reconstruction system working on 2D images. The main problem of 3D reconstruction from 2D stereo images is to find the correspondences, that is the points in the images that are the projection of the same point in the scene. The problem of finding correspondences in stereo images, has long been the focus of much research in computer vision and it is still an open problem; it has been addressed and described in the survey presented by Brown in [4] who classifies the employed techniques as local and global. Local methods [12, 7, 11] find correspondences by looking at the local neighborhoods of the points, while global methods [3, 6, 15] exploit constraints on global features in order to reduce sensitivity to those local regions in the stereo images that fail to match due to occlusion or uniform texture. In this paper we describe two 3D reconstruction techniques; the first creates a sparse model, while the second determines a dense model. In section 3 we present a local method to find the correspondences of the 15 most characteristic face fiducial points 1, extracted as described in [1]. The stereo matching is based on the analysis of the grey level intensities in the neighborhoods of the points, by means of the SSD and a Gabor filter bank. A 3D face model is then obtained, by using the 3D reconstructed face fiducial points as control points to deform the parameterized 3D face model (Candide model) defined in [13]. Section 4 presents a different stereo matching method that deploys global and local techniques to define a dense disparity map; it employs a multi-resolution energy minimizing snake grid algorithm [10]. The grid ensures uniqueness, continuity and ordering, while the minimum of the global energy function brings to a good correspondence quality and smoothness. 2. Image preprocessing The first and fundamental step to obtain a good 3D reconstruction is a precise camera calibration, which provides the intrinsic, extrinsic and distortion parameters of the cameras used for the stereo images acquisition. Good calibration results can be obtained using the very well known algorithm, presented by Zhang [16]; the weakness of this method is that it needs an accurate initialization based on the precise corner positions of a calibration pattern (e.g. a chessboard). To initialize the camera calibration algorithm we use a chessboard corner detector [2] which finds the chessboard corners with sub-pixel precision, making no assumption about the scale and the orientation of the chessboard, and working under very different illumination con- 1 The face fiducial points are: the eyebrow vertexes, the nose tip, the eye and the lip corners, the upper and the lower middle points of the lips and the eyes. 1

2 ditions. The camera parameters, determined by the camera calibration algorithm, are used by the image rectification algorithm presented by Fusiello et al. [9]. In the rectified images, the relative rotation among the original images is removed to simplify the task of the following steps for 3D reconstruction. Image rectification enables an efficient and reliable stereo matching since the search of conjugate points (or matching patterns in general) can be executed along the same horizontal lines of the rectified images. Our stereo matching algorithm works on pairs of stereo images (a frontal and two lateral views) of the same subject, and defines the stereo correspondences by matching the points in the lateral views to those in the frontal view. 3. 3D reconstruction of face Fiducial Points In this section we present our method for 3D facial fiducial point reconstruction, using three stereo views (left, right and frontal view). For each view, the 15 2D fiducial points are at first identified as described in [1]; we define as P L (i), P F (i) and P R (i) (i = 1,.., 15) the sets of fiducial points located respectively in the left, frontal and right view (figure 1). Based on the assumption that the fiducial points in the frontal view are more precisely detected than those in the lateral views, we consider the P F (i) as the elements that must be matched to determine the conjugate pairs. To find the corresponding points of P F (i) we employ a stereo matching procedure in the neighborhood of PL (i) and P R (i) since they are very close to the conjugate of P F (i). The rectified images are scaled so that the distance between the fiducial points located on the external eye corners in the frontal view is 150 pixels. The conjugate point corresponding to the fiducial point P F (i) in the left (right) side of the face is searched in the left (right) view. The nose tip, the upper and the lower lip middle points are searched in both the views. For each point P F (i) we search for the correspondent, P L (i), in a rectangular area, SA, which dimensions are 21 7 pixels. It is vertically centered on the epipolar line corresponding to P F (i) 2, and horizontally centered on the point P L (i). Our method uses the multiple window approach proposed in [8]; the distance metric analyzed to find the matching is the Sum of Square Difference (SSD). For each fiducial point P F (i) we extract 9 template sub-images from the frontal view. The template models are shown in figure 2: they are characterized by varying positions of the pixel P F (i) in the template. The algorithm starts by scanning the search area SA to search, for each template, its best matching point. This is found by taking the point which minimizes the SSD of the grey level intensities. In this 2 This is a reasonable choice since we work on rectified images. way we find nine candidate points C k (k = 1,.., 9). To select the best match we characterize the point P F (i) and the nine candidates C k in terms of their surrounding grey level portion of image. A vector (Jet) J( x) of 40 coefficients is calculated convolving the portion of grey level image around the point x with the following bank of 40 Gabor kernels (we consider 5 frequencies and 8 orientations): ( ) [ ( )] ψ j ( x) = k2 j σ 2 exp k2 j x2 2σ 2 exp(i k j x) exp σ2 2 The best candidate selected is the one that maximizes a similarity measure between Jets: z S(J(P F (i)), J(C k )) = J z(p F (i))j z (C k ) z (J z(p F (i))) 2 z (J z(c k )) 2 where z = 0,..., 39. To improve the robustness of the method this procedure is repeated three times, by varying the dimension of the template windows (11 11, and 31 31). We choose as correspondent of P F (i) the one with the highest similarity measure. Given the set of conjugate pairs the algorithm for the 3D point reconstruction can be applied. At first, in order to correct the effect of the lens distortion, the undistorted positions of the fiducial points are computed. For each pair of conjugate points their optical rays may be computed and expressed in the same reference frame [14]. Their intersection defines the 3D reconstruction of the point represented in the different views 3. Due to the calibration and the stereo matching errors the rays may not intersect; this is the reason why the reconstructed 3D point is the mid point of the segment minimizing the distance between the two rays [14]. In this way we obtain the set of the 15 3D face fiducial points. These points are used as the control points to deform the parameterized 3D face model (see figure 3-a). It is composed of 64 vertices and 114 triangles. The model deformation is obtained by two steps: the global rigid alignment scales and aligns the Candide model to the control points, while the local deformation brings its vertices closer to the 3D fiducial points of a specific face. This second step replaces the 3D coordinates of the model fiducial points with the ones of the 3D fiducial points, and determines the position of the other vertices according to these displacements. Although this model is quite simple and can be deformed in a low computational time, its simplicity does not allow 3 Note that for the nose tip, the upper and the lower lip middle points we have two conjugate pairs; their final 3D reconstruction is the mid point of the segment connecting the two 3D points obtained by each pair.

3 Figure 1. A stereo image set: left, frontal and right view; the fiducial points are highlighted. Figure 2. The nine template windows. The pixel P F (i) is highlighted. significative deformation, and the method creates very similar models for different faces. To obtain a model that could better represent different faces we infer from the 15 face fiducial points 72 additional points on the images (see figure 3-b,c). A stereo matching algorithm is then used to find the correspondences of these points in the three images, and their 3D reconstructions allow us to build a 3D model with 87 (72+15) vertices and 156 triangles (see figure 3-d). Since the 72 inferred points are often located in low textured or occluded regions, the conjugate pairs found are often inaccurate, leading to an imprecise 3D model (see figure 4). Therefore we tried to obtain a more precise model by using a dense stereo matching technique, described in the following section. 4. Stereo matching by multi-resolution snake grid model. The stereo matching technique described in this section defines a dense disparity map by employing a significantly modified version of the energy minimizing snake grid algorithm presented in [10]. It is applied separately on two pairs of stereo images (left-frontal, right-frontal), and the dense models obtained are integrated in an unique face model. At first the eye centers are localized in the stereo images with the algorithm described in [5]. Their position allows to define the bounding boxes strictly including the face. Since we work on rectified images the bounding box in the lateral (left or right) image is translated and scaled to have the same vertical position and the same dimensions of the bounding box detected in the frontal view. To obtain a more precise reconstruction, the snake grid is applied at different scales in a coarse to fine fashion. To create a multi scale pyramid the two stereo sub-images (referred as images in the following) are iteratively halfweighted; the first level of the pyramid, L 0, contains the original stereo images, and the lowest level, L h, contains two images whose biggest dimension is less than 2 7. In this way we obtain a number of levels h = lg 2 X 7, where X is the maximum dimension of the original images. The snake grid algorithm is at first applied at the lowest level of the pyramid and initialized by means of two grids, G F and G L, composed of N M grid points, and created on the frontal and the lateral views. We choose N = 8 and M = 12. G F is kept fixed and G L is iteratively distorted to

4 a b c d Figure 3. a: Candide face model. b,c: 72 inferred points. d: New 3D model. Figure 4. Example of 3D reconstruction with the new model. find the best matches of the grid points in G F, by minimizing a global energy function (see figure 5). When the snake grid, applied to the generic level L k of the pyramid, has stabilized to the minimum global energy, the two grids are scaled to initialize the grids in the stereo images at the above level of the pyramid, L k 1. Using bilinear interpolation additional points are added to those grids and the energy minimizing algorithm is applied to find more and more correspondences. This iterative process is repeated until the first level L 0 is reached. At the generic level, L k, the number of grid points, G L and G F, is given by the equation: G L (k) = G F (k) = (2 (h k) (N 1)+1)(2 (h k) (M 1)+1) For each level L k, 0 k h, the energy minimizing snake-grid algorithm can be described as follows; let P F (i, j) (x,y), P L (i, j) (x,y) be the grid points in position (i, j) in G F and G L, the superscript (x, y) refers to the horizontal and vertical image coordinates of these points, let E n be the global energy of the grid at the n th iteration. Figure 5. Distortion of the grid stabilized to the minimum global energy.

5 E n 1 := ; E n :=0; while E n E n 1 > ε(k) do begin E n 1 := E n ; For each point,p F (i, j) 4, in G F do begin Define the neighborhood A of P L (i, j) where the best match of P F (i, j) must be searched: it is the horizontal segment 5 composed of the points p (x,y) L having a vertical coordinate p y L = P L(i, j) y, and a horizontal coordinate P L (i 1, j) x < p x L < P L(i + 1, j) x ; For each point p L A compute the energy function E( p L, P F (i, j)); (see equation 1); Choose as best matching point the p L with the minimum energy; Set the new position of P L (i, j) (x,y) := p (x,y) L ; Define the energy of the grid point (i, j) as E j i :=E( p L, P F (i, j)); end; Compute the global energy of the grid as E n := i,j Ej i ; I C L = 1 t l t u=1 v=1 l ĨL C (u, v) where ĨL(u, v) are the pixel grey levels in the portion of image around p L, and (t, l) are the dimensions of the portion of image used for the matching. The energy parameter α is proportional to the gradient in the portion of image I F : α = t u=1 v=1 l ( ) I F (u, v) u + I F (u, v) v thus, in image regions with high gradient, which correlate well and unambiguously, the grid points are moved mostly relying on the external energy. On the other hand, in low gradient regions where false matches may occur, the most significative term is the internal energy. It penalizes those points p L that are far away from the centroid of the eight surrounding grid points, P L (see figure 6), assuming that the best match must be located somewhere close to it; it also ensures the smoothness of the disparity map. end; The energy function E( p L, P F (i, j)), is used to evaluate the match between the two points p L and P F (i, j); it is computed as a linear combination of an external energy, E ext, and an internal energy, E int. More precisely: E( p L, P F (i, j)) = E int αe ext (1) where α is an energy parameter. Based on the assumption that a portion of image, I F, around P F (i, j) (x,y) remains almost unchanged in the lateral image we choose as a measure of the external energy, E ext, the median of the cross correlation value (CC) calculated over the nine templates described in [8]. CC is defined by the following equation and includes the correlation of all the three color channels in the same formula: t,l C=R,G,B u,v=1 CC = t,l C=R,G,B u,v=1 I C F = 1 t l ( ) ) I C F (u, v) IC F (ĨC L (u, v) IC L ( ) 2 IF C(u, v) t,l IC F (ĨC L (u, v) I C L u,v=1 t u=1 v=1 l IF C (u, v) 4 The P F (i, j) are chosen in random order. 5 A is chosen according to the fact that we work on rectified images. ) 2 Figure 6. The point p L and its 8-neighbors. More precisely: P (x,y) L = 1 8 t,l= 1,0,1; (t,l) (0,0) P (i + t, j + l) (x,y) and the internal energy is composed of two terms as follows: E int = P exp(a; p 1, s 1) + P exp(b; p 2, s 2) (2) { s z if z < p P exp(z; p, s) = s p e z p 1 (3) otherwise where p 1 = 0.5, s 1 = 1, p 2 = 0.2, s 2 = 2 are experimentally set parameters and the function P exp(z; p, s) is defined in 1 < z < 1 (see figure 7). The first term in equation 2 is a function of the distance, a, between the horizontal position of p L and the horizontal

6 Figure 7. The functions P exp(a; 0.5, 1) and P exp(b; 0.2, 2) Figure 8. The 3D reconstruction obtained with the snake grid algorithm. position of P L ; a is defined as: p x L P x a = ( PL max (i + 1, j) x P x PL, (i 1, j) x P x ) In the second term, b is a function of the slope of the segments connecting p L to P L (i, j 1) and P L (i, j +1), and it is defined as follows, where the quantities x and y are shown in figure 6: x1 b = 1 2 y1 x2 y2 ( y1 + y2) The iterative snake algorithm running on the grids at the level L k of the pyramid, is stopped when the difference between the global energies of two consecutive iterations is GL(k) E n E n 1 < ε(k). We set ε(k) = Once the dense disparity maps are computed on the two pairs of stereo images, two 3D models are built by a ray intersection algorithm. An example of 3D model obtained is shown in figure 8. The original face sub-image has a dimension of , the multi-scale pyramid has 4 levels and the integration of the two 3D models provides a final 3D face model composed of = points. 5. Conclusions and future works In this paper we present two methods for 3D face reconstruction from 2D stereo images. We introduce two stereo matching algorithms used to create a sparse and a dense model. The sparse 3D model is mainly based on the reconstruction of the 15 most characteristic facial fiducial points. In order to verify the usefulness of 3D information for face recognition, we have computed 3D biometric measures of the face such as the inter-ocular distance, the width of the eyes and the mouth, and the distance between the eyes corners and the nose tip. Preliminary experiments show that these measures do not increase the performance of the face recognition system [1], working on 2D images only. This is probably due to the small number of features that can be extracted by the 15 3D fiducial points. We believe that a greater number of points may help defining a set of more discriminative features and, based on this assumption, we developed a more detailed dense model. Future works are focused on the definition of a feature set extracted on the dense model; the features may be obtained by applying 3D filters or by inferring measures on characteristic face curves. References [1] S. Arca, P. Campadelli, and R. Lanzarotti. A face recognition system based on automatically determined face fiducial points. Pattern Recognition, 39: , March [2] S. Arca, E. Casiraghi, P. Campadelli, and G. Lombardi. A fully automatic system for chessboard corner localization. Submitted to Pattern Recognition Letters, August, [3] A. Benshair, P. Miche, and R. Debrie. Fast and automatic stereo vision matching algorithm based on dynamic programming method. Pattern Recognition Letters, 17: , [4] M. Brown, B. Burschka, and G. Hager. Advances in computational stereo. IEEE Transaction On Pattern Analysis and Machine Intelligence, 25: , [5] P. Campadelli, R. Lanzarotti, and G. Lipori. Eye localization and face recognition. RAIRO, to be published, 2006.

7 [6] I. Cox, S. Hingorani, S. Rao, and B. Maggs. A maximum likelihood stereo algorithm. Computer Vision and Image Understanding, 63: , [7] F. et al. Realtime correlationbased stereo: algorithm, implementation and applications. (2013), [8] A. Fusiello and V. Roberto. Efficient stereo with multiple windowing. IEEE 19/97, pages , [9] A. Fusiello, E. Trucco, and A. Verri. A compact algorithm for rectification of stereo pairs. Machine Vision and Applications, pages 16 22, [10] S. Huq, B. Abidi, A. Goshtasby, and M. Abidi. Stereo matching with energy minimizing snake grid for 3d face modeling. Proceedings of SPIE, Conference on Biometric Technology for Human Identification, 5404: , [11] S. S. Intille and A. F. Bobick. Disparityspace images and large occlusion stereo. European Conference on Computer Vision, pages , [12] Y. Ohta and T. Kanade. Stereo by intra and interscanline search using dynamic programming. Pattern Analysis and Machine Intelligence, (7(2)): , [13] M. Rydfalk. Candide, a parameterized face. Report No. LiTH-ISY-I-866, Dept. of Electrical Engineering, Linkping University, Sweden, [14] E. Trucco and A. Verri. Introductory Techinques for 3D Computer Vision. Prentice Hall, [15] H. Zao. Global optimal surface from stereo. In Proc. of International Conference on Pattern Recognition, 1: , [16] Z. Zhang. Flexible camera calibration by viewing a plane from inknown orientations. In Proc of International Conference on Computer Vision (ICCV 99), Corfu, Greece, pages , 1999.

A face recognition system based on local feature analysis

A face recognition system based on local feature analysis A face recognition system based on local feature analysis Stefano Arca, Paola Campadelli, Raffaella Lanzarotti Dipartimento di Scienze dell Informazione Università degli Studi di Milano Via Comelico, 39/41

More information

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes

More information

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri] Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri] Introduction Disparity: Informally: difference between two pictures Allows us to gain a strong

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Lecture 14: Basic Multi-View Geometry

Lecture 14: Basic Multi-View Geometry Lecture 14: Basic Multi-View Geometry Stereo If I needed to find out how far point is away from me, I could use triangulation and two views scene point image plane optical center (Graphic from Khurram

More information

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Eye Detection by Haar wavelets and cascaded Support Vector Machine Eye Detection by Haar wavelets and cascaded Support Vector Machine Vishal Agrawal B.Tech 4th Year Guide: Simant Dubey / Amitabha Mukherjee Dept of Computer Science and Engineering IIT Kanpur - 208 016

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17

Stereo Wrap + Motion. Computer Vision I. CSE252A Lecture 17 Stereo Wrap + Motion CSE252A Lecture 17 Some Issues Ambiguity Window size Window shape Lighting Half occluded regions Problem of Occlusion Stereo Constraints CONSTRAINT BRIEF DESCRIPTION 1-D Epipolar Search

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence

CS4495/6495 Introduction to Computer Vision. 3B-L3 Stereo correspondence CS4495/6495 Introduction to Computer Vision 3B-L3 Stereo correspondence For now assume parallel image planes Assume parallel (co-planar) image planes Assume same focal lengths Assume epipolar lines are

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Rectification and Distortion Correction

Rectification and Distortion Correction Rectification and Distortion Correction Hagen Spies March 12, 2003 Computer Vision Laboratory Department of Electrical Engineering Linköping University, Sweden Contents Distortion Correction Rectification

More information

Chaplin, Modern Times, 1936

Chaplin, Modern Times, 1936 Chaplin, Modern Times, 1936 [A Bucket of Water and a Glass Matte: Special Effects in Modern Times; bonus feature on The Criterion Collection set] Multi-view geometry problems Structure: Given projections

More information

Using Geometric Blur for Point Correspondence

Using Geometric Blur for Point Correspondence 1 Using Geometric Blur for Point Correspondence Nisarg Vyas Electrical and Computer Engineering Department, Carnegie Mellon University, Pittsburgh, PA Abstract In computer vision applications, point correspondence

More information

Local Image Registration: An Adaptive Filtering Framework

Local Image Registration: An Adaptive Filtering Framework Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,

More information

Detection of Facial Landmarks of North Eastern Indian People

Detection of Facial Landmarks of North Eastern Indian People International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 9 (2013), pp. 953-962 International Research Publications House http://www. irphouse.com /ijict.htm Detection

More information

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera Tomokazu Sato, Masayuki Kanbara and Naokazu Yokoya Graduate School of Information Science, Nara Institute

More information

Stereo imaging ideal geometry

Stereo imaging ideal geometry Stereo imaging ideal geometry (X,Y,Z) Z f (x L,y L ) f (x R,y R ) Optical axes are parallel Optical axes separated by baseline, b. Line connecting lens centers is perpendicular to the optical axis, and

More information

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision

Stereo. Outline. Multiple views 3/29/2017. Thurs Mar 30 Kristen Grauman UT Austin. Multi-view geometry, matching, invariant features, stereo vision Stereo Thurs Mar 30 Kristen Grauman UT Austin Outline Last time: Human stereopsis Epipolar geometry and the epipolar constraint Case example with parallel optical axes General case with calibrated cameras

More information

Capturing, Modeling, Rendering 3D Structures

Capturing, Modeling, Rendering 3D Structures Computer Vision Approach Capturing, Modeling, Rendering 3D Structures Calculate pixel correspondences and extract geometry Not robust Difficult to acquire illumination effects, e.g. specular highlights

More information

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1)

Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1) Image Rectification (Stereo) (New book: 7.2.1, old book: 11.1) Guido Gerig CS 6320 Spring 2013 Credits: Prof. Mubarak Shah, Course notes modified from: http://www.cs.ucf.edu/courses/cap6411/cap5415/, Lecture

More information

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman Stereo 11/02/2012 CS129, Brown James Hays Slides by Kristen Grauman Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman Why multiple views? Structure

More information

Computer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16

Computer Vision I. Dense Stereo Correspondences. Anita Sellent 1/15/16 Computer Vision I Dense Stereo Correspondences Anita Sellent Stereo Two Cameras Overlapping field of view Known transformation between cameras From disparity compute depth [ Bradski, Kaehler: Learning

More information

Step-by-Step Model Buidling

Step-by-Step Model Buidling Step-by-Step Model Buidling Review Feature selection Feature selection Feature correspondence Camera Calibration Euclidean Reconstruction Landing Augmented Reality Vision Based Control Sparse Structure

More information

Multiple View Geometry

Multiple View Geometry Multiple View Geometry CS 6320, Spring 2013 Guest Lecture Marcel Prastawa adapted from Pollefeys, Shah, and Zisserman Single view computer vision Projective actions of cameras Camera callibration Photometric

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Epipolar Geometry in Stereo, Motion and Object Recognition

Epipolar Geometry in Stereo, Motion and Object Recognition Epipolar Geometry in Stereo, Motion and Object Recognition A Unified Approach by GangXu Department of Computer Science, Ritsumeikan University, Kusatsu, Japan and Zhengyou Zhang INRIA Sophia-Antipolis,

More information

Implementing the Scale Invariant Feature Transform(SIFT) Method

Implementing the Scale Invariant Feature Transform(SIFT) Method Implementing the Scale Invariant Feature Transform(SIFT) Method YU MENG and Dr. Bernard Tiddeman(supervisor) Department of Computer Science University of St. Andrews yumeng@dcs.st-and.ac.uk Abstract The

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe

Structured Light. Tobias Nöll Thanks to Marc Pollefeys, David Nister and David Lowe Structured Light Tobias Nöll tobias.noell@dfki.de Thanks to Marc Pollefeys, David Nister and David Lowe Introduction Previous lecture: Dense reconstruction Dense matching of non-feature pixels Patch-based

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix

More information

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC

Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Accurate and Dense Wide-Baseline Stereo Matching Using SW-POC Shuji Sakai, Koichi Ito, Takafumi Aoki Graduate School of Information Sciences, Tohoku University, Sendai, 980 8579, Japan Email: sakai@aoki.ecei.tohoku.ac.jp

More information

Flexible Calibration of a Portable Structured Light System through Surface Plane

Flexible Calibration of a Portable Structured Light System through Surface Plane Vol. 34, No. 11 ACTA AUTOMATICA SINICA November, 2008 Flexible Calibration of a Portable Structured Light System through Surface Plane GAO Wei 1 WANG Liang 1 HU Zhan-Yi 1 Abstract For a portable structured

More information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week

More information

Displacement estimation

Displacement estimation Displacement estimation Displacement estimation by block matching" l Search strategies" l Subpixel estimation" Gradient-based displacement estimation ( optical flow )" l Lukas-Kanade" l Multi-scale coarse-to-fine"

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Ball detection and predictive ball following based on a stereoscopic vision system

Ball detection and predictive ball following based on a stereoscopic vision system Research Collection Conference Paper Ball detection and predictive ball following based on a stereoscopic vision system Author(s): Scaramuzza, Davide; Pagnottelli, Stefano; Valigi, Paolo Publication Date:

More information

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision What Happened Last Time? Human 3D perception (3D cinema) Computational stereo Intuitive explanation of what is meant by disparity Stereo matching

More information

What have we leaned so far?

What have we leaned so far? What have we leaned so far? Camera structure Eye structure Project 1: High Dynamic Range Imaging What have we learned so far? Image Filtering Image Warping Camera Projection Model Project 2: Panoramic

More information

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar.

Matching. Compare region of image to region of image. Today, simplest kind of matching. Intensities similar. Matching Compare region of image to region of image. We talked about this for stereo. Important for motion. Epipolar constraint unknown. But motion small. Recognition Find object in image. Recognize object.

More information

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model TAE IN SEOL*, SUN-TAE CHUNG*, SUNHO KI**, SEONGWON CHO**, YUN-KWANG HONG*** *School of Electronic Engineering

More information

Occluded Facial Expression Tracking

Occluded Facial Expression Tracking Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy

Machine vision. Summary # 11: Stereo vision and epipolar geometry. u l = λx. v l = λy 1 Machine vision Summary # 11: Stereo vision and epipolar geometry STEREO VISION The goal of stereo vision is to use two cameras to capture 3D scenes. There are two important problems in stereo vision:

More information

Depth from two cameras: stereopsis

Depth from two cameras: stereopsis Depth from two cameras: stereopsis Epipolar Geometry Canonical Configuration Correspondence Matching School of Computer Science & Statistics Trinity College Dublin Dublin 2 Ireland www.scss.tcd.ie Lecture

More information

Final Exam Study Guide

Final Exam Study Guide Final Exam Study Guide Exam Window: 28th April, 12:00am EST to 30th April, 11:59pm EST Description As indicated in class the goal of the exam is to encourage you to review the material from the course.

More information

Gaze interaction (2): models and technologies

Gaze interaction (2): models and technologies Gaze interaction (2): models and technologies Corso di Interazione uomo-macchina II Prof. Giuseppe Boccignone Dipartimento di Scienze dell Informazione Università di Milano boccignone@dsi.unimi.it http://homes.dsi.unimi.it/~boccignone/l

More information

Local Feature Detectors

Local Feature Detectors Local Feature Detectors Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Slides adapted from Cordelia Schmid and David Lowe, CVPR 2003 Tutorial, Matthew Brown,

More information

A MULTI-RESOLUTION APPROACH TO DEPTH FIELD ESTIMATION IN DENSE IMAGE ARRAYS F. Battisti, M. Brizzi, M. Carli, A. Neri

A MULTI-RESOLUTION APPROACH TO DEPTH FIELD ESTIMATION IN DENSE IMAGE ARRAYS F. Battisti, M. Brizzi, M. Carli, A. Neri A MULTI-RESOLUTION APPROACH TO DEPTH FIELD ESTIMATION IN DENSE IMAGE ARRAYS F. Battisti, M. Brizzi, M. Carli, A. Neri Università degli Studi Roma TRE, Roma, Italy 2 nd Workshop on Light Fields for Computer

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 12 130228 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Panoramas, Mosaics, Stitching Two View Geometry

More information

Omni-directional Multi-baseline Stereo without Similarity Measures

Omni-directional Multi-baseline Stereo without Similarity Measures Omni-directional Multi-baseline Stereo without Similarity Measures Tomokazu Sato and Naokazu Yokoya Graduate School of Information Science, Nara Institute of Science and Technology 8916-5 Takayama, Ikoma,

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy BSB663 Image Processing Pinar Duygulu Slides are adapted from Selim Aksoy Image matching Image matching is a fundamental aspect of many problems in computer vision. Object or scene recognition Solving

More information

Stereo Image Rectification for Simple Panoramic Image Generation

Stereo Image Rectification for Simple Panoramic Image Generation Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,

More information

Announcements. Stereo Vision Wrapup & Intro Recognition

Announcements. Stereo Vision Wrapup & Intro Recognition Announcements Stereo Vision Wrapup & Intro Introduction to Computer Vision CSE 152 Lecture 17 HW3 due date postpone to Thursday HW4 to posted by Thursday, due next Friday. Order of material we ll first

More information

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix F Fundamental

More information

Critique: Efficient Iris Recognition by Characterizing Key Local Variations

Critique: Efficient Iris Recognition by Characterizing Key Local Variations Critique: Efficient Iris Recognition by Characterizing Key Local Variations Authors: L. Ma, T. Tan, Y. Wang, D. Zhang Published: IEEE Transactions on Image Processing, Vol. 13, No. 6 Critique By: Christopher

More information

GOPRO CAMERAS MATRIX AND DEPTH MAP IN COMPUTER VISION

GOPRO CAMERAS MATRIX AND DEPTH MAP IN COMPUTER VISION Tutors : Mr. Yannick Berthoumieu Mrs. Mireille El Gheche GOPRO CAMERAS MATRIX AND DEPTH MAP IN COMPUTER VISION Delmi Elias Kangou Ngoma Joseph Le Goff Baptiste Naji Mohammed Hamza Maamri Kenza Randriamanga

More information

Stereo vision. Many slides adapted from Steve Seitz

Stereo vision. Many slides adapted from Steve Seitz Stereo vision Many slides adapted from Steve Seitz What is stereo vision? Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape What is

More information

Computer Vision, Lecture 11

Computer Vision, Lecture 11 Computer Vision, Lecture 11 Professor Hager http://www.cs.jhu.edu/~hager Computational Stereo Much of geometric vision is based on information from (or more) camera locations hard to recover 3D information

More information

Stereo matching. Francesco Isgrò. 3D Reconstruction and Stereo p.1/21

Stereo matching. Francesco Isgrò. 3D Reconstruction and Stereo p.1/21 Stereo matching Francesco Isgrò 3D Reconstruction and Stereo p.1/21 Structure of a stereo vision system Extract interesting point from each image Determine a set of matching points Compute the fundamental

More information

Quasi-Euclidean Uncalibrated Epipolar Rectification

Quasi-Euclidean Uncalibrated Epipolar Rectification Dipartimento di Informatica Università degli Studi di Verona Rapporto di ricerca Research report September 2006 RR 43/2006 Quasi-Euclidean Uncalibrated Epipolar Rectification L. Irsara A. Fusiello Questo

More information

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few...

There are many cues in monocular vision which suggests that vision in stereo starts very early from two similar 2D images. Lets see a few... STEREO VISION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Bill Freeman and Antonio Torralba (MIT), including their own

More information

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Image Formation. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania Image Formation Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 18/03/2014 Outline Introduction; Geometric Primitives

More information

Depth from two cameras: stereopsis

Depth from two cameras: stereopsis Depth from two cameras: stereopsis Epipolar Geometry Canonical Configuration Correspondence Matching School of Computer Science & Statistics Trinity College Dublin Dublin 2 Ireland www.scss.tcd.ie Lecture

More information

High Accuracy Depth Measurement using Multi-view Stereo

High Accuracy Depth Measurement using Multi-view Stereo High Accuracy Depth Measurement using Multi-view Stereo Trina D. Russ and Anthony P. Reeves School of Electrical Engineering Cornell University Ithaca, New York 14850 tdr3@cornell.edu Abstract A novel

More information

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera Tomokazu Satoy, Masayuki Kanbaray, Naokazu Yokoyay and Haruo Takemuraz ygraduate School of Information

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Stereo Vision 2 Inferring 3D from 2D Model based pose estimation single (calibrated) camera Stereo

More information

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923 Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923 Teesta suspension bridge-darjeeling, India Mark Twain at Pool Table", no date, UCR Museum of Photography Woman getting eye exam during

More information

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views? Recap: Features and filters Epipolar geometry & stereo vision Tuesday, Oct 21 Kristen Grauman UT-Austin Transforming and describing images; textures, colors, edges Recap: Grouping & fitting Now: Multiple

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania.

Visual Tracking. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania. Image Processing Laboratory Dipartimento di Matematica e Informatica Università degli studi di Catania 1 What is visual tracking? estimation of the target location over time 2 applications Six main areas:

More information

Robert Collins CSE486, Penn State. Lecture 09: Stereo Algorithms

Robert Collins CSE486, Penn State. Lecture 09: Stereo Algorithms Lecture 09: Stereo Algorithms left camera located at (0,0,0) Recall: Simple Stereo System Y y Image coords of point (X,Y,Z) Left Camera: x T x z (, ) y Z (, ) x (X,Y,Z) z X right camera located at (T x,0,0)

More information

Final project bits and pieces

Final project bits and pieces Final project bits and pieces The project is expected to take four weeks of time for up to four people. At 12 hours per week per person that comes out to: ~192 hours of work for a four person team. Capstone:

More information

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition Berk Gökberk, M.O. İrfanoğlu, Lale Akarun, and Ethem Alpaydın Boğaziçi University, Department of Computer

More information

Lecture 16: Computer Vision

Lecture 16: Computer Vision CS4442/9542b: Artificial Intelligence II Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field

More information

Lecture 16: Computer Vision

Lecture 16: Computer Vision CS442/542b: Artificial ntelligence Prof. Olga Veksler Lecture 16: Computer Vision Motion Slides are from Steve Seitz (UW), David Jacobs (UMD) Outline Motion Estimation Motion Field Optical Flow Field Methods

More information

Real time facial expression recognition from image sequences using Support Vector Machines

Real time facial expression recognition from image sequences using Support Vector Machines Real time facial expression recognition from image sequences using Support Vector Machines I. Kotsia a and I. Pitas a a Aristotle University of Thessaloniki, Department of Informatics, Box 451, 54124 Thessaloniki,

More information

Lecture 19: Motion. Effect of window size 11/20/2007. Sources of error in correspondences. Review Problem set 3. Tuesday, Nov 20

Lecture 19: Motion. Effect of window size 11/20/2007. Sources of error in correspondences. Review Problem set 3. Tuesday, Nov 20 Lecture 19: Motion Review Problem set 3 Dense stereo matching Sparse stereo matching Indexing scenes Tuesda, Nov 0 Effect of window size W = 3 W = 0 Want window large enough to have sufficient intensit

More information

Stereo Video Processing for Depth Map

Stereo Video Processing for Depth Map Stereo Video Processing for Depth Map Harlan Hile and Colin Zheng University of Washington Abstract This paper describes the implementation of a stereo depth measurement algorithm in hardware on Field-Programmable

More information

A novel point matching method for stereovision measurement using RANSAC affine transformation

A novel point matching method for stereovision measurement using RANSAC affine transformation A novel point matching method for stereovision measurement using RANSAC affine transformation Naiguang Lu, Peng Sun, Wenyi Deng, Lianqing Zhu, Xiaoping Lou School of Optoelectronic Information & Telecommunication

More information

The SIFT (Scale Invariant Feature

The SIFT (Scale Invariant Feature The SIFT (Scale Invariant Feature Transform) Detector and Descriptor developed by David Lowe University of British Columbia Initial paper ICCV 1999 Newer journal paper IJCV 2004 Review: Matt Brown s Canonical

More information

STEREO BY TWO-LEVEL DYNAMIC PROGRAMMING

STEREO BY TWO-LEVEL DYNAMIC PROGRAMMING STEREO BY TWO-LEVEL DYNAMIC PROGRAMMING Yuichi Ohta Institute of Information Sciences and Electronics University of Tsukuba IBARAKI, 305, JAPAN Takeo Kanade Computer Science Department Carnegie-Mellon

More information

Head Pose Estimation by using Morphological Property of Disparity Map

Head Pose Estimation by using Morphological Property of Disparity Map Head Pose Estimation by using Morphological Property of Disparity Map Sewoong Jun*, Sung-Kee Park* and Moonkey Lee** *Intelligent Robotics Research Center, Korea Institute Science and Technology, Seoul,

More information

FACIAL RECOGNITION BASED ON THE LOCAL BINARY PATTERNS MECHANISM

FACIAL RECOGNITION BASED ON THE LOCAL BINARY PATTERNS MECHANISM FACIAL RECOGNITION BASED ON THE LOCAL BINARY PATTERNS MECHANISM ABSTRACT Alexandru Blanda 1 This work presents a method of facial recognition, based on Local Binary Models. The idea of using this algorithm

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014 SIFT SIFT: Scale Invariant Feature Transform; transform image

More information

Stereo. Many slides adapted from Steve Seitz

Stereo. Many slides adapted from Steve Seitz Stereo Many slides adapted from Steve Seitz Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1 image 2 Dense depth map Binocular stereo Given a calibrated

More information

Precise eye localization through a general-to-specific model definition

Precise eye localization through a general-to-specific model definition Precise eye localization through a general-to-specific model definition Paola Campadelli, Raffaella Lanzarotti, Giuseppe Lipori Dipartimento di Scienze dell Informazione Università degli Studi di Milano

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

Rectification and Disparity

Rectification and Disparity Rectification and Disparity Nassir Navab Slides prepared by Christian Unger What is Stereo Vision? Introduction A technique aimed at inferring dense depth measurements efficiently using two cameras. Wide

More information

3D Computer Vision. Structured Light I. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light I. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light I Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

FACE RECOGNITION USING INDEPENDENT COMPONENT

FACE RECOGNITION USING INDEPENDENT COMPONENT Chapter 5 FACE RECOGNITION USING INDEPENDENT COMPONENT ANALYSIS OF GABORJET (GABORJET-ICA) 5.1 INTRODUCTION PCA is probably the most widely used subspace projection technique for face recognition. A major

More information

Visual Tracking. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania

Visual Tracking. Antonino Furnari. Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania Visual Tracking Antonino Furnari Image Processing Lab Dipartimento di Matematica e Informatica Università degli Studi di Catania furnari@dmi.unict.it 11 giugno 2015 What is visual tracking? estimation

More information

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography

Today. Stereo (two view) reconstruction. Multiview geometry. Today. Multiview geometry. Computational Photography Computational Photography Matthias Zwicker University of Bern Fall 2009 Today From 2D to 3D using multiple views Introduction Geometry of two views Stereo matching Other applications Multiview geometry

More information

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, School of Computer Science and Communication, KTH Danica Kragic EXAM SOLUTIONS Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006, 14.00 19.00 Grade table 0-25 U 26-35 3 36-45

More information

Inexpensive Construction of a 3D Face Model from Stereo Images

Inexpensive Construction of a 3D Face Model from Stereo Images Inexpensive Construction of a 3D Face Model from Stereo Images M. Shahriar Hossain, Monika Akbar and J. Denbigh Starkey Department of Computer Science, Montana State University, Bozeman, MT 59717, USA.

More information

Flow Estimation. Min Bai. February 8, University of Toronto. Min Bai (UofT) Flow Estimation February 8, / 47

Flow Estimation. Min Bai. February 8, University of Toronto. Min Bai (UofT) Flow Estimation February 8, / 47 Flow Estimation Min Bai University of Toronto February 8, 2016 Min Bai (UofT) Flow Estimation February 8, 2016 1 / 47 Outline Optical Flow - Continued Min Bai (UofT) Flow Estimation February 8, 2016 2

More information