Head Detection and Tracking by 2-D and 3-D Ellipsoid Fitting.

Size: px
Start display at page:

Download "Head Detection and Tracking by 2-D and 3-D Ellipsoid Fitting."

Transcription

1 Head Detection and Tracking by 2-D and 3-D Ellipsoid Fitting. Nikos Grammalidis and Michael G.Strintzis Department of Electrical Engineering, University of Thessaloniki Thessaloniki , GREECE Abstract A novel procedure for segmenting a set of scattered 3- D data obtained from a head and shoulders multiview sequence is presented. The procedure consists of two steps. In the first step, two ellipses corresponding to the head and the body of the person are identified based on ellipse fitting of the outline of the person in each image. The fitting is based on a fast direct least squares method using the constraint that forces a general conic to be an ellipse. In order to achieve head/body segmentation, a K-means algorithm is used to minimise the fitting error between the points and the two ellipsoids. In the second step, a 3-D ellipsoid model corresponding to the head of the person is identified using an extension of the above method. Robustness and outlier removal can be achieved if 3-D ellipsoid model estimation technique is used in conjunction with the Median of Least Squares (MedLS) technique, which minimises the median of the errors corresponding to each 3-D point. An interesting application of the proposed method is the combination of the 3-D ellipsoid model with a generic face model which is adapted to the face images to provide information only for the high-detail front part of the head, while the 3-D ellipsoid is used for the back of the head, which is usually not visible. 1. Introduction Estimation of the position, shape and motion of the head from image sequences has been the focus of a significant amount of recent research [11, 8, 1]. In [11], a face segmentation and identification algorithm, utilising the elliptical structure of the human head is presented. An ellipse is fitted to a properly preprocessed edge map, in order to mark the boundary between the head and the background regions. A similar approach is used in [2], where a 2-D ellipse is fitted onto binary edge data and postprocessing techniques are used to eliminate detection errors. However, such approaches work well only with when the background is homogeneous and lightly to moderately cluttered. n [8], a technique is proposed for facial detection and coding. The detection of the face, as well as the facial features (eyes, nose, mouth, etc) is based on a parametric image model obtained using a Karhunen-Loève decomposition. The detected face image is then normalised by preprocessing, aiming to align the detected facial features with those of a standard model and also to compensate lighting and contrast variations. The normalised image is then coded by projecting it (by inner product) to a set of eigenfaces [13, 14], i.e. a subset of the eigenvectors obtained using a large database of normalised face images. If only a small number of primary eigenfaces are used, this technique results to successful face detection and coding using extremely low bitrates. Furthermore, the resulting face representations may additionally be used for face recognition applications. In [1], a 3-D ellipsoidal model is used for robust tracking of the rigid motion of the human head from a video sequence. This approach is based on the interpretation of the optical flow in terms of 3-D motion of the model. The method is seen to be robust to large head movements and to provide better results than those obtained using a simpler planar head model. Initialisation is based on predicting the shape of the ellipsoid from the positions of the face features (eyes, nose, mouth) which were previously estimated as in [8]. In [17], a hybrid system is presented using visual and sound cues for face localisation. The system exploits mainly luminance and motion information for locating and tracking the head edges from videoconferencing image sequences. In the present paper we shall assume a typical videoconferencing scene, with one person in front of the cameras. The initial set of 3-D points may be produced using depth estimation from stereoscopic or multiple views, provided that the camera geometry and calibration parameters are known. Furthermore, we assume that the outline of the person is available using some preprocessing technique. In the test

2 sequences, used for our experimental results, this outline is easily available since the background is homogeneous. A K-means algorithm, which is able to identify ellipses in the image, was then used for simultaneous ellipse fitting and segmentation of the outline. Specifically, a least squares fit of the data on a conic is performed, using an additional constraint, forcing the conic to be an ellipse. The algorithm is applied to estimate two ellipses, each corresponding to the head and body of the person. The basic ellipse fitting technique was also extended to the three dimensional case. This 3-D ellipsoid fitting technique was applied so as to estimate 3-D ellipsoids for the head of a person. Specifically, 3-D points that are projected inside each of the estimated ellipses in the previous step are used to estimate the 3-D ellipsoid model. Because of its high computational efficiency, this technique can be used for fast tracking of the head and body position and orientation. Since the segmentation of the nodes is only performed at the first time instant, only the estimation of the 3-D ellipsoid from the available data points has to be determined at subsequent instants. Finally, an interesting application of the proposed method is the combination of the 3-D ellipsoid model with a generic face model which is adapted to the face images to provide information only for the high-detail front part of the head, while the 3-D ellipsoid is used for the back of the head, which is usually not visible. 2. Estimating the initial set of 3-D points In order to estimate the initial 3-D point set, a multiocular system with N fully calibrated views is assumed. Each camera c k k =1 ::: N is modeled using a pinhole camera model based on perspective projection and we assume that accurate calibration information is available. A number of methods have been developed for depth estimation in stereoscopic [16, 15] and multiview [6, 4] image sequences. In this paper, we use the depth estimation and from trinocular image sequences developed in the ACTS 092 PANORAMA project. More specifically, depth is estimated from a fully calibrated trinocular image pair using the depth estimation algorithm in [4]. Furthermore, a reliability value is also estimated for depth estimate. A set of 3-D points where depth estimation with high reliability is then identified by selecting the 3-D points whose reliability exceeds a certain threshold. 3. Approximation of a 3-D data set using an 3- D Ellipsoid Model In [5], a new efficient method was presented for fitting ellipses to scattered 2-D data. The method is based on fitting the general equation of a conic to the given data, subject to a constraint which forces the conic to be an ellipse. In this paper, we develop an segmentation and estimation scheme using this technique in a K-means algorithm. Furthermore, we extend this technique for 3-D data, so that it can be used for approximating a general set of 3-D points using a 3-D ellipsoid. The general equation of a conic in the N-D space is: F (A b c x) =x T Ax + b T x + c =0 (1) where A is a symmetric matrix. In order for this general conic to be an ellipsoid in N-dimensions, the matrix A must be either positive or negative definite. As a result, the problem of fitting an ellipsoid into N data points x i can be solved by minimising the sum of squares of the algebraic distances NX d(a b c)= F (A b c x) 2 (2) i=1 subject to the constraint that A is either positive or negative definite. In general, this constrained problem is very difficult to solve since the Kuhn-Tucker conditions [9] do not guarantee a solution [5]. For the 2-D case, a direct solution to the problem was presented in [5]. The equation of the 2-D conic is F (a x) =ax = ax 2 + bxy + cy 2 + dx + ey + f =0 (3) where a = [a bcdef], x = [x 2 xy y 2 xy1] and the constraint to lead to an ellipse is 4ac ; b 2 > 0. Sincea is a free parameter, the equality constraint 4ac ; b 2 =1may be imposed instead. This in turn may be be written as ; a T Ca = a 26 T a =1 (4) 75 Using the above constraint, the following constrained fitting problem is formulated [5]: P N Problem 1 Minimise E = i=1 F 2 (a x i )=jjdajj 2 subject to the constraint a T Ca =1 where D = [x1 x2 x N ] T. This problem can be directly solved using a Lagrange multiplier, providing the following solution: a is the generalised eigenvector of D T Da = Ca (5) corresponding to the smallest positive eigenvalue.

3 In [5], it is proved that exactly one eigenvalue of eq. (5) is positive, thus a unique eigenvector a s is always found using the above method, which is non-iterative and thus extremely efficient. Even though this approach is not easily extended into the general N-D problem, a generalisation suitable for the 3-D case will now be formulated. The equation of a 3-D conic is: F (a x) =a x = = ax 2 +bxy +cy 2 +dxz +fz 2 +gz+hy +kz+l =0 (6) where a x are 10-dimensional vectors, can be written in the form of eq. (1), where A = U = 2 4 a b=2 d=2 b=2 c e=2 d=2 e=2 f a b=2 b=2 c 3 5 U v = v T f v =[d=2 e=2] T : The following Lemma provides the constraint forcing a general 3-D conic to be an ellipsoid. Lemma 1 A is positive definite or positive negative, provided that i. det(u) > 0 () 4ac ; b 2 > 0 ii. (a + c) det(a) > 0 Proof Let x =[xyz] T, u =[xy] T and u c = ;U ;1 vz. Then, x T Ax = u T Uu +2u T vz + fz 2 = (u ; u c ) T U(u ; u c )+z 2 (f ; v T U ;1 v)= (u ; u c ) T U(u ; u c )+ det(u) z2 (f det(u) ;v T c ;b=2 ;b=2 a v) = (u ; u c ) T U(u ; u c )+ z2 det(u) (acf ; fb2 =4 ;cd 2 =4 ; ae 2 =4+bde=4) = (u ; u c ) T U(u ; u c )+z 2 det(a) det(u) Assuming that U is positive definite, (u ; u c ) T U(u ; u c ) > 0 when x 6= 0. Furthermore, det(u) > 0 and tr(u) =a + c>0, thus if det(a) > 0, then eq. (7) yields x T Ax > 0 for x 6= 0, thus A is positive definite. Similarly, if U is negative definite, (u ; u c ) T U(u ; u c ) < 0, det(u) > 0 and tr(u) =a + c<0. Therefore, if det(a) < 0 then eq. (7) yields x T Ax < 0 for x 6= 0, thus A is negative definite, q.e.d. The first condition of Lemma 1 is very similar to the constraint 4ac ; b 2 > 0 used in the 2-D case. Therefore, we (7) can setup the problem of minimising the error E, asbefore using the constraint a T Ca =1,where: C = ; The solution is again obtained by solving Problem 1 using eq. (5) with the new C of (8). However, in this case, after obtaining the solution, we have to check whether the second condition of Lemma 1 is satisfied. If we select the sign of the solution eigenvector so that a + c>0, then the second condition is simplified to det(a) > Ellipse estimation and segmentation using K-means algorithm A significant advantage of the method described in the previous Section is its low computational requirements. This feature may be exploited by designing an algorithm for simultaneous ellipse estimation and segmentation of a set of 2-D points into ellipses. A simple algorithm that can be used for this purpose is a modified K-means algorithm, which is described below. In this version, we consider an algorithm using K =2,which is suitable for identifying the head and the body of a person from the person s outline in one of the available views (8) i. Initialisation: Divide the available points into K=2 sets, based on subsidiary information (e.g. relative position of head and body). ii. Fit an ellipse to each set, using the procedure described in the previous Section. iii. Reassign each point to the ellipse for which the distance F 2 (a x i ) in eq. (3) between the point x i and the ellipsoid is minimised. Points with distances larger than a given threshold are not assigned to any set. iv. End of the estimation procedure if the change in the estimated ellipsoid parameters is below a threshold else return to step ii.

4 5. Robustness and removal of outliers using Median of Least Squares When the initial 3-D point estimates are noisy, or contain outliers (e.g. points from the neck or the body of the person), the above method result to erroneous estimates of the 3-D ellipsoid parameters. Robustness can be achieved by using the MedLS technique which minimises the median of the errors corresponding to each 3-D point. Furthermore, this scheme allows the integration of the tests for the validity of the conditions forcing the estimated conic forms to be 3-D ellipsoids. More precisely, the procedure consists of the following steps: 1. Set E min to a large number. Randomly select a set of N 3-D points from the initial 3-D point set. For these points: 2. Estimate the ellipsoid parameters A b c. If the resulting conic is not an ellipsoid, discard the solution and return to Step Compute the Median of the Least Squares Fitting Error (E MedLS ) for each of point in the initial 3-D point set, using the estimated ellipsoid parameters. If E MedLS <E min,sete min = E MedLS and the corresponding parameters as the optimal parameters. 4. Iterate Steps 1-3 until E min is below a threshold or a maximum number of iterations is reached. 6 Combining 3-D Face Models with the 3-D Ellipsoid Another important application of the proposed methods is to combine a generic face model with a 3-D ellipsoid model estimated using the method described in Section 3. Various methods for adapting generic 3-D face models to face images have been proposed [12, 7, 3, 10]. In this paper, we have used the face adaptation method proposed in [10], where a generic 3D face model is transformed in a way that it adapts to the characteristics of an actual face observed in its front and profile image views. Assuming that a set of feature points have been detected on both views, the adaptation procedure initialises with a rigid transformation of the model aiming to minimise the distances of the 3D model feature nodes from the calculated 3D coordinates of the 2D feature points. Then, a non-rigid transformation ensures that the feature nodes are displaced optimally close to their exact calculated positions, dragging their neighbours in a way that does not deform the facial model in an unnatural way. In theory, any 3D model of the human face may be used. In practice however, the closer the characteristics of the model are to the characteristics of the image (e.g. model of female face for a female face) the fewer feature points will be needed for accurate adaptation. The estimated 3-D ellipsoid is then used to approximate the back of the head, while the adapted generic 3-D wireframe provides information for the high-detail front part of the head. In order to identify the part of the ellipsoid that corresponds to the back of the head, a check is performed as to whether the line that connects each node with the ellipsoid center intersects with any triangle of the wireframe corresponding to the front(face) wireframe. If so, the corresponding node is removed from the ellipsoid wireframe. Furthermore, each triangle in the ellipsoid wireframe containing this node is also removed. At this stage, two wireframe models corresponding to the front and the back part of the head have been generated. Since these parts are not connected, an algorithm was developed to automatically connect them. This is achieved by adding additional triangles between the list of nodes that define the largest wireframe boundary in each wireframe model. The largest wireframe boundary is simply the set of nodes corresponding to the largest path that can be defined using wireframe edges (edges belonging to only one triangle). 7. Experimental Results Results were obtained using the 3-view sequence Ludo The set of initial 3-D points is illustrated in Figure 1 (a triangulation was performed for display purposes). The approximation of the outline of the person with two ellipses corresponding to the head and the body area is shown in Figure 2. In this case a very simple initialisation was used: points lying higher (lower) than the mean height of the outline are attributed to the head (body) region. Convergence is achieved within a small number of iterations, typically 3, as shown in Figure 3, each requiring less than 0.5 sec at an SGI R4400 workstation. Then, the nodes of the 3-D wireframe of the head that are projected inside the head outline are used to estimate a 3-D ellipsoid model for the head. This model is illustrated in Figure 4. Table 1 shows the improvement at the performance of the proposed 3-D estimation technique when the MedLS technique is used to achieve robustness to noise. The comparison criterion used was the value of a fitting ratio r defined as follows: r = C intersection (9) C union where C intersection,c union are the number of pixels belonging to the intersection and the union respectively of the head area (defined manually after sketching the head contour) and the estimated ellipse obtained by projecting the 1 This sequence was provided by THOMPSON BROADCAST SYSTEMS for the the ACTS 092 project PANORAMA.

5 0.015 Error E Iterations Figure 3. Convergence of mean error E Figure 1. The initial 3-D wireframe Figure 2. The outline of the person (left view) and the two estimated ellipses using the K- means algorithm 3-D ellipsoid on the image plane. Larger values of the ratio r 2 (0 1] indicate better approximation of the head area with the ellipsoid. All methods were applied to 10 consecutive frames from the Ludo sequence (left view), and the average fitting ratio was calculated and is presented in Table 1. Two views of generated head models produced when combining a generic face model with the estimated 3-D ellipsoid model, as described in Section 6, are presented in Figures 5(a-b)(front and rear view respectively). 8. Conclusions An efficient technique to segment the head and body parts obtained from a head and shoulders multiview sequence and to estimate a 3-D ellipsoid model corresponding to the head was presented. A coarse approximation of an initial wireframe requiring minimal bitrate was thus obtained. Robustness and outlier removal was achieved by augmenting the above basic methods with either the Me- Figure 4. Estimated 3-D ellipsoid model superimposed on the head of Ludo (left view) dian of Least Squares (MedLS) technique which minimises the median of the errors corresponding to each 3-D point. Also, an application of the proposed method for combining of the 3-D ellipsoid model with a generic face model which is adapted to the face images was examined. 9. Acknowledgement This work was supported by the EU project ACTS VI- DAS, the GSRT PENED project and the GSRT AMEA Project: LIP-PHONE - A Videophone for the Deaf. Method fitting ratio r One-step 3-D parameter estimation MedLS Table 1. Improvement of fitting ratio r when using the MedLS robust estimation technique.

6 References [1] S. Basu, I. Essa, and A. Pentland. Motion Regularization for Model-based Head Tracking. In Proceedings, International Conference on Pattern Recognition, Vienna, Austria, August [2] A. Eleftheriadis and A. Jacquin. Automatic face location detection for model-assisted rate control in h.261 compatible coding of video. Signal Processing : Image Communication, 7: , November [3] M. Escher and N. Magnenat-Thalmann. Automatic 3d cloning and real-time animation of a human face. In Proceedings Computer Animation (CA97), Geneva, Switzerland, [4] L. Falkenhagen. Block-Based Depth Estimation from Image Triples with Unrestricted Camera Setup. In IEEE Workshop on Multimedia Image Processing, Princeton, NJ, June [5] A. W. Fitzgibbon, M. Pilu, and R. Fisher. Direct Least Squares Fitting of Ellipses. In Proc. International Conference on Pattern Recognition, Vienna, Austria, August [6] N. Grammalidis and M. G. Strintzis. Disparity and Occlusion Estimation in Multiocular Systems and their Coding for the Communication of Multiview Image Sequences. IEEE Trans. Circuits and Systems for Video Technology, 8(3): , June [7] N. Magnenat-Thalmann, P.Kalra, and M. Escher. Face to virtual face. Proc. IEEE, 86(5), [8] B. Moghaddam and A. Pentland. An Automatic System for Model-Based Coding of Faces. In IEEE Data Compression Conference, Snowbird, Utah, March [9] S. Rao. Optimization:Theory and Applications. Wiley Eastern, [10] N. Sarris and M. G. Strintzis. Three dimensional facial model adaptation. In submitted to ICIP 00, Vancouver, Canada, [11] S. Sirohey. Human Face Segmentation and Identification. Master s thesis, CV Laboratory, University of Maryland, College Park, MD, November [12] D. Terzopoulos and K. Waters. Physically-based facial modelling, analysis, and animation. The Journal of Visualization and Computer Animation, 1:73 80, [13] M. Turk and A. Pentland. Eigenfaces for recognition. Journal of Cognitive Neuro Science, 3(1):71 86, [14] M. A. Turk and A. P. Pentland. Face recognition using eigenfaces. In Proceedings IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages , Hawai, June [15] D. Tzovaras, N. Grammalidis, and M. G. Strintzis. Object- Based Coding of Stereo Image Sequences Using Joint 3D motion/disparity Compensation. IEEE Trans. on Ciscuits and Systems for Video Technology, 7(2): , April [16] D. Tzovaras, N. Grammalidis, and M. G. Strintzis. Disparity field and Depth Map Coding for Multiview 3D Image Generation. Signal Processing : Image Communication, 11(3): , January [17] C. Wang and M. Brandstein. A hybrid real-time face tracking system. In Proc. IEEE Int. Conf. Acoust., Speech and Signal Processing, Seattle, WA, May (a) (b) Figure 5. The combination of the generic face model and the 3-D ellipsoid model as viewed from two different viewpoints a) front view b) rear view.

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press, ISSN

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press,   ISSN Hopeld Network for Stereo Correspondence Using Block-Matching Techniques Dimitrios Tzovaras and Michael G. Strintzis Information Processing Laboratory, Electrical and Computer Engineering Department, Aristotle

More information

Multiview Image Compression using Algebraic Constraints

Multiview Image Compression using Algebraic Constraints Multiview Image Compression using Algebraic Constraints Chaitanya Kamisetty and C. V. Jawahar Centre for Visual Information Technology, International Institute of Information Technology, Hyderabad, INDIA-500019

More information

Exploring Curve Fitting for Fingers in Egocentric Images

Exploring Curve Fitting for Fingers in Egocentric Images Exploring Curve Fitting for Fingers in Egocentric Images Akanksha Saran Robotics Institute, Carnegie Mellon University 16-811: Math Fundamentals for Robotics Final Project Report Email: asaran@andrew.cmu.edu

More information

Sprite Generation and Coding in Multiview Image Sequences

Sprite Generation and Coding in Multiview Image Sequences 302 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 10, NO. 2, MARCH 2000 Sprite Generation and Coding in Multiview Image Sequences Nikos Grammalidis, Student Member, IEEE, Dimitris

More information

Structure from Motion. Prof. Marco Marcon

Structure from Motion. Prof. Marco Marcon Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)

More information

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet:

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet: Local qualitative shape from stereo without detailed correspondence Extended Abstract Shimon Edelman Center for Biological Information Processing MIT E25-201, Cambridge MA 02139 Internet: edelman@ai.mit.edu

More information

NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING

NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING Nicole Atzpadin 1, Serap Askar, Peter Kauff, Oliver Schreer Fraunhofer Institut für Nachrichtentechnik, Heinrich-Hertz-Institut,

More information

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Final Report Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This report describes a method to align two videos.

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational

More information

Speech Driven Synthesis of Talking Head Sequences

Speech Driven Synthesis of Talking Head Sequences 3D Image Analysis and Synthesis, pp. 5-56, Erlangen, November 997. Speech Driven Synthesis of Talking Head Sequences Peter Eisert, Subhasis Chaudhuri,andBerndGirod Telecommunications Laboratory, University

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Image Processing and Image Representations for Face Recognition

Image Processing and Image Representations for Face Recognition Image Processing and Image Representations for Face Recognition 1 Introduction Face recognition is an active area of research in image processing and pattern recognition. Since the general topic of face

More information

On Modeling Variations for Face Authentication

On Modeling Variations for Face Authentication On Modeling Variations for Face Authentication Xiaoming Liu Tsuhan Chen B.V.K. Vijaya Kumar Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213 xiaoming@andrew.cmu.edu

More information

Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17

Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17 Recognition (Part 4) CSE 152 Lecture 17 Announcements Homework 5 is due June 9, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying Images Chapter 17: Detecting Objects in Images

More information

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test

Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test Announcements (Part 3) CSE 152 Lecture 16 Homework 3 is due today, 11:59 PM Homework 4 will be assigned today Due Sat, Jun 4, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying

More information

Algorithm research of 3D point cloud registration based on iterative closest point 1

Algorithm research of 3D point cloud registration based on iterative closest point 1 Acta Technica 62, No. 3B/2017, 189 196 c 2017 Institute of Thermomechanics CAS, v.v.i. Algorithm research of 3D point cloud registration based on iterative closest point 1 Qian Gao 2, Yujian Wang 2,3,

More information

Face Detection and Recognition in an Image Sequence using Eigenedginess

Face Detection and Recognition in an Image Sequence using Eigenedginess Face Detection and Recognition in an Image Sequence using Eigenedginess B S Venkatesh, S Palanivel and B Yegnanarayana Department of Computer Science and Engineering. Indian Institute of Technology, Madras

More information

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology

More information

LOCALIZATION OF FACIAL REGIONS AND FEATURES IN COLOR IMAGES. Karin Sobottka Ioannis Pitas

LOCALIZATION OF FACIAL REGIONS AND FEATURES IN COLOR IMAGES. Karin Sobottka Ioannis Pitas LOCALIZATION OF FACIAL REGIONS AND FEATURES IN COLOR IMAGES Karin Sobottka Ioannis Pitas Department of Informatics, University of Thessaloniki 540 06, Greece e-mail:fsobottka, pitasg@zeus.csd.auth.gr Index

More information

Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions

Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions Tomokazu Wakasugi, Masahide Nishiura and Kazuhiro Fukui Corporate Research and Development Center, Toshiba Corporation

More information

Miniature faking. In close-up photo, the depth of field is limited.

Miniature faking. In close-up photo, the depth of field is limited. Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg

More information

calibrated coordinates Linear transformation pixel coordinates

calibrated coordinates Linear transformation pixel coordinates 1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial

More information

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and

Expression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and Wes Miller 5/11/2011 Comp Sci 534 Expression Detection in Video Abstract Expression detection is useful as a non-invasive method of lie detection and behavior prediction, as many facial expressions are

More information

Principal Component Analysis and Neural Network Based Face Recognition

Principal Component Analysis and Neural Network Based Face Recognition Principal Component Analysis and Neural Network Based Face Recognition Qing Jiang Mailbox Abstract People in computer vision and pattern recognition have been working on automatic recognition of human

More information

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION

COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA

More information

Stereo Image Rectification for Simple Panoramic Image Generation

Stereo Image Rectification for Simple Panoramic Image Generation Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,

More information

Image processing and features

Image processing and features Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry

More information

Towards the completion of assignment 1

Towards the completion of assignment 1 Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics

More information

What is Computer Vision?

What is Computer Vision? Perceptual Grouping in Computer Vision Gérard Medioni University of Southern California What is Computer Vision? Computer Vision Attempt to emulate Human Visual System Perceive visual stimuli with cameras

More information

Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors

Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors Keith Forbes 1 Anthon Voigt 2 Ndimi Bodika 2 1 Digital Image Processing Group 2 Automation and Informatics Group Department of Electrical

More information

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Costas Panagiotakis and Anastasios Doulamis Abstract In this paper, an unsupervised, automatic video human members(human

More information

Model-based segmentation and recognition from range data

Model-based segmentation and recognition from range data Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This

More information

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision

Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China

More information

Multi-view stereo. Many slides adapted from S. Seitz

Multi-view stereo. Many slides adapted from S. Seitz Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window

More information

Model Based Perspective Inversion

Model Based Perspective Inversion Model Based Perspective Inversion A. D. Worrall, K. D. Baker & G. D. Sullivan Intelligent Systems Group, Department of Computer Science, University of Reading, RG6 2AX, UK. Anthony.Worrall@reading.ac.uk

More information

A Statistical Consistency Check for the Space Carving Algorithm.

A Statistical Consistency Check for the Space Carving Algorithm. A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper

More information

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera Wolfgang Niem, Jochen Wingbermühle Universität Hannover Institut für Theoretische Nachrichtentechnik und Informationsverarbeitung

More information

Using temporal seeding to constrain the disparity search range in stereo matching

Using temporal seeding to constrain the disparity search range in stereo matching Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department

More information

VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural

VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD Ertem Tuncel and Levent Onural Electrical and Electronics Engineering Department, Bilkent University, TR-06533, Ankara, Turkey

More information

PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using

PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1 Tak-keung CHENG derek@cs.mu.oz.au Leslie KITCHEN ljk@cs.mu.oz.au Computer Vision and Pattern Recognition Laboratory, Department of Computer Science,

More information

Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform

Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform Torsten Palfner, Alexander Mali and Erika Müller Institute of Telecommunications and Information Technology, University of

More information

THE GENERATION of a stereoscopic image sequence

THE GENERATION of a stereoscopic image sequence IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 15, NO. 8, AUGUST 2005 1065 Stereoscopic Video Generation Based on Efficient Layered Structure and Motion Estimation From a Monoscopic

More information

Contents. 1 Introduction Background Organization Features... 7

Contents. 1 Introduction Background Organization Features... 7 Contents 1 Introduction... 1 1.1 Background.... 1 1.2 Organization... 2 1.3 Features... 7 Part I Fundamental Algorithms for Computer Vision 2 Ellipse Fitting... 11 2.1 Representation of Ellipses.... 11

More information

Local Image Registration: An Adaptive Filtering Framework

Local Image Registration: An Adaptive Filtering Framework Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,

More information

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS

METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires

More information

CS201 Computer Vision Camera Geometry

CS201 Computer Vision Camera Geometry CS201 Computer Vision Camera Geometry John Magee 25 November, 2014 Slides Courtesy of: Diane H. Theriault (deht@bu.edu) Question of the Day: How can we represent the relationships between cameras and the

More information

3D Model Acquisition by Tracking 2D Wireframes

3D Model Acquisition by Tracking 2D Wireframes 3D Model Acquisition by Tracking 2D Wireframes M. Brown, T. Drummond and R. Cipolla {96mab twd20 cipolla}@eng.cam.ac.uk Department of Engineering University of Cambridge Cambridge CB2 1PZ, UK Abstract

More information

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt

ECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

An Adaptive Eigenshape Model

An Adaptive Eigenshape Model An Adaptive Eigenshape Model Adam Baumberg and David Hogg School of Computer Studies University of Leeds, Leeds LS2 9JT, U.K. amb@scs.leeds.ac.uk Abstract There has been a great deal of recent interest

More information

IRIS SEGMENTATION OF NON-IDEAL IMAGES

IRIS SEGMENTATION OF NON-IDEAL IMAGES IRIS SEGMENTATION OF NON-IDEAL IMAGES William S. Weld St. Lawrence University Computer Science Department Canton, NY 13617 Xiaojun Qi, Ph.D Utah State University Computer Science Department Logan, UT 84322

More information

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze

More information

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation

Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation IJCSNS International Journal of Computer Science and Network Security, VOL.13 No.11, November 2013 1 Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial

More information

Multiview Reconstruction

Multiview Reconstruction Multiview Reconstruction Why More Than 2 Views? Baseline Too short low accuracy Too long matching becomes hard Why More Than 2 Views? Ambiguity with 2 views Camera 1 Camera 2 Camera 3 Trinocular Stereo

More information

Conversion of 2D Image into 3D and Face Recognition Based Attendance System

Conversion of 2D Image into 3D and Face Recognition Based Attendance System Conversion of 2D Image into 3D and Face Recognition Based Attendance System Warsha Kandlikar, Toradmal Savita Laxman, Deshmukh Sonali Jagannath Scientist C, Electronics Design and Technology, NIELIT Aurangabad,

More information

CS6670: Computer Vision

CS6670: Computer Vision CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection

More information

2D to pseudo-3d conversion of "head and shoulder" images using feature based parametric disparity maps

2D to pseudo-3d conversion of head and shoulder images using feature based parametric disparity maps University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2001 2D to pseudo-3d conversion of "head and shoulder" images using feature

More information

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H. Nonrigid Surface Modelling and Fast Recovery Zhu Jianke Supervisor: Prof. Michael R. Lyu Committee: Prof. Leo J. Jia and Prof. K. H. Wong Department of Computer Science and Engineering May 11, 2007 1 2

More information

Two-view geometry Computer Vision Spring 2018, Lecture 10

Two-view geometry Computer Vision Spring 2018, Lecture 10 Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of

More information

Abstract We present a system which automatically generates a 3D face model from a single frontal image of a face. Our system consists of two component

Abstract We present a system which automatically generates a 3D face model from a single frontal image of a face. Our system consists of two component A Fully Automatic System To Model Faces From a Single Image Zicheng Liu Microsoft Research August 2003 Technical Report MSR-TR-2003-55 Microsoft Research Microsoft Corporation One Microsoft Way Redmond,

More information

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation

Face Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation Face Tracking Amit K. Roy-Chowdhury and Yilei Xu Department of Electrical Engineering, University of California, Riverside, CA 92521, USA {amitrc,yxu}@ee.ucr.edu Synonyms Facial Motion Estimation Definition

More information

International Journal of Advanced Research in Computer Science and Software Engineering

International Journal of Advanced Research in Computer Science and Software Engineering Volume 2, Issue 8, August 2012 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Study on Block

More information

Motion Tracking and Event Understanding in Video Sequences

Motion Tracking and Event Understanding in Video Sequences Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem

More information

Perception and Action using Multilinear Forms

Perception and Action using Multilinear Forms Perception and Action using Multilinear Forms Anders Heyden, Gunnar Sparr, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: {heyden,gunnar,kalle}@maths.lth.se Abstract

More information

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION

CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester

More information

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition

Linear Discriminant Analysis in Ottoman Alphabet Character Recognition Linear Discriminant Analysis in Ottoman Alphabet Character Recognition ZEYNEB KURT, H. IREM TURKMEN, M. ELIF KARSLIGIL Department of Computer Engineering, Yildiz Technical University, 34349 Besiktas /

More information

Optimal Grouping of Line Segments into Convex Sets 1

Optimal Grouping of Line Segments into Convex Sets 1 Optimal Grouping of Line Segments into Convex Sets 1 B. Parvin and S. Viswanathan Imaging and Distributed Computing Group Information and Computing Sciences Division Lawrence Berkeley National Laboratory,

More information

AN ADAPTIVE MESH METHOD FOR OBJECT TRACKING

AN ADAPTIVE MESH METHOD FOR OBJECT TRACKING AN ADAPTIVE MESH METHOD FOR OBJECT TRACKING Mahdi Koohi 1 and Abbas Shakery 2 1 Department of Computer Engineering, Islamic Azad University, Shahr-e-Qods Branch,Tehran,Iran m.kohy@yahoo.com 2 Department

More information

Computer Vision Lecture 17

Computer Vision Lecture 17 Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week

More information

A Robust Two Feature Points Based Depth Estimation Method 1)

A Robust Two Feature Points Based Depth Estimation Method 1) Vol.31, No.5 ACTA AUTOMATICA SINICA September, 2005 A Robust Two Feature Points Based Depth Estimation Method 1) ZHONG Zhi-Guang YI Jian-Qiang ZHAO Dong-Bin (Laboratory of Complex Systems and Intelligence

More information

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary

More information

Facial Deformations for MPEG-4

Facial Deformations for MPEG-4 Facial Deformations for MPEG-4 Marc Escher, Igor Pandzic, Nadia Magnenat Thalmann MIRALab - CUI University of Geneva 24 rue du Général-Dufour CH1211 Geneva 4, Switzerland {Marc.Escher, Igor.Pandzic, Nadia.Thalmann}@cui.unige.ch

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

3D Geometry and Camera Calibration

3D Geometry and Camera Calibration 3D Geometry and Camera Calibration 3D Coordinate Systems Right-handed vs. left-handed x x y z z y 2D Coordinate Systems 3D Geometry Basics y axis up vs. y axis down Origin at center vs. corner Will often

More information

Two-View Geometry (Course 23, Lecture D)

Two-View Geometry (Course 23, Lecture D) Two-View Geometry (Course 23, Lecture D) Jana Kosecka Department of Computer Science George Mason University http://www.cs.gmu.edu/~kosecka General Formulation Given two views of the scene recover the

More information

Gender Classification Technique Based on Facial Features using Neural Network

Gender Classification Technique Based on Facial Features using Neural Network Gender Classification Technique Based on Facial Features using Neural Network Anushri Jaswante Dr. Asif Ullah Khan Dr. Bhupesh Gour Computer Science & Engineering, Rajiv Gandhi Proudyogiki Vishwavidyalaya,

More information

Recovering light directions and camera poses from a single sphere.

Recovering light directions and camera poses from a single sphere. Title Recovering light directions and camera poses from a single sphere Author(s) Wong, KYK; Schnieders, D; Li, S Citation The 10th European Conference on Computer Vision (ECCV 2008), Marseille, France,

More information

CS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow

CS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow CS 565 Computer Vision Nazar Khan PUCIT Lectures 15 and 16: Optic Flow Introduction Basic Problem given: image sequence f(x, y, z), where (x, y) specifies the location and z denotes time wanted: displacement

More information

Noise Reduction in Image Sequences using an Effective Fuzzy Algorithm

Noise Reduction in Image Sequences using an Effective Fuzzy Algorithm Noise Reduction in Image Sequences using an Effective Fuzzy Algorithm Mahmoud Saeid Khadijeh Saeid Mahmoud Khaleghi Abstract In this paper, we propose a novel spatiotemporal fuzzy based algorithm for noise

More information

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Video Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Literature Survey Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This literature survey compares various methods

More information

Dense 3D Reconstruction. Christiano Gava

Dense 3D Reconstruction. Christiano Gava Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction

More information

Agenda. Rotations. Camera calibration. Homography. Ransac

Agenda. Rotations. Camera calibration. Homography. Ransac Agenda Rotations Camera calibration Homography Ransac Geometric Transformations y x Transformation Matrix # DoF Preserves Icon translation rigid (Euclidean) similarity affine projective h I t h R t h sr

More information

Camera Calibration. COS 429 Princeton University

Camera Calibration. COS 429 Princeton University Camera Calibration COS 429 Princeton University Point Correspondences What can you figure out from point correspondences? Noah Snavely Point Correspondences X 1 X 4 X 3 X 2 X 5 X 6 X 7 p 1,1 p 1,2 p 1,3

More information

Detecting Ellipses via Bounding Boxes

Detecting Ellipses via Bounding Boxes Detecting Ellipses via Bounding Boxes CHUN-MING CHANG * Department of Information and Design, Asia University, Taiwan ABSTRACT A novel algorithm for ellipse detection based on bounding boxes is proposed

More information

Mouse Pointer Tracking with Eyes

Mouse Pointer Tracking with Eyes Mouse Pointer Tracking with Eyes H. Mhamdi, N. Hamrouni, A. Temimi, and M. Bouhlel Abstract In this article, we expose our research work in Human-machine Interaction. The research consists in manipulating

More information

Multiple View Geometry in Computer Vision

Multiple View Geometry in Computer Vision Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Structure Computation Lecture 18 March 22, 2005 2 3D Reconstruction The goal of 3D reconstruction

More information

Applications Video Surveillance (On-line or off-line)

Applications Video Surveillance (On-line or off-line) Face Face Recognition: Dimensionality Reduction Biometrics CSE 190-a Lecture 12 CSE190a Fall 06 CSE190a Fall 06 Face Recognition Face is the most common biometric used by humans Applications range from

More information

A Survey of Light Source Detection Methods

A Survey of Light Source Detection Methods A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light

More information

Depth Estimation for View Synthesis in Multiview Video Coding

Depth Estimation for View Synthesis in Multiview Video Coding MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Depth Estimation for View Synthesis in Multiview Video Coding Serdar Ince, Emin Martinian, Sehoon Yea, Anthony Vetro TR2007-025 June 2007 Abstract

More information

Efficient Block Matching Algorithm for Motion Estimation

Efficient Block Matching Algorithm for Motion Estimation Efficient Block Matching Algorithm for Motion Estimation Zong Chen International Science Inde Computer and Information Engineering waset.org/publication/1581 Abstract Motion estimation is a key problem

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

A Method of Automated Landmark Generation for Automated 3D PDM Construction

A Method of Automated Landmark Generation for Automated 3D PDM Construction A Method of Automated Landmark Generation for Automated 3D PDM Construction A. D. Brett and C. J. Taylor Department of Medical Biophysics University of Manchester Manchester M13 9PT, Uk adb@sv1.smb.man.ac.uk

More information

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe

Face detection and recognition. Many slides adapted from K. Grauman and D. Lowe Face detection and recognition Many slides adapted from K. Grauman and D. Lowe Face detection and recognition Detection Recognition Sally History Early face recognition systems: based on features and distances

More information

Finding a Best Fit Plane to Non-coplanar Point-cloud Data Using Non Linear and Linear Equations

Finding a Best Fit Plane to Non-coplanar Point-cloud Data Using Non Linear and Linear Equations AIJSTPME (013) 6(): 17-3 Finding a Best Fit Plane to Non-coplanar Point-cloud Data Using Non Linear and Linear Equations Mulay A. Production Engineering Department, College of Engineering, Pune, India

More information

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry

55:148 Digital Image Processing Chapter 11 3D Vision, Geometry 55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence

More information