Head Detection and Tracking by 2-D and 3-D Ellipsoid Fitting.
|
|
- Mervyn Albert Higgins
- 6 years ago
- Views:
Transcription
1 Head Detection and Tracking by 2-D and 3-D Ellipsoid Fitting. Nikos Grammalidis and Michael G.Strintzis Department of Electrical Engineering, University of Thessaloniki Thessaloniki , GREECE Abstract A novel procedure for segmenting a set of scattered 3- D data obtained from a head and shoulders multiview sequence is presented. The procedure consists of two steps. In the first step, two ellipses corresponding to the head and the body of the person are identified based on ellipse fitting of the outline of the person in each image. The fitting is based on a fast direct least squares method using the constraint that forces a general conic to be an ellipse. In order to achieve head/body segmentation, a K-means algorithm is used to minimise the fitting error between the points and the two ellipsoids. In the second step, a 3-D ellipsoid model corresponding to the head of the person is identified using an extension of the above method. Robustness and outlier removal can be achieved if 3-D ellipsoid model estimation technique is used in conjunction with the Median of Least Squares (MedLS) technique, which minimises the median of the errors corresponding to each 3-D point. An interesting application of the proposed method is the combination of the 3-D ellipsoid model with a generic face model which is adapted to the face images to provide information only for the high-detail front part of the head, while the 3-D ellipsoid is used for the back of the head, which is usually not visible. 1. Introduction Estimation of the position, shape and motion of the head from image sequences has been the focus of a significant amount of recent research [11, 8, 1]. In [11], a face segmentation and identification algorithm, utilising the elliptical structure of the human head is presented. An ellipse is fitted to a properly preprocessed edge map, in order to mark the boundary between the head and the background regions. A similar approach is used in [2], where a 2-D ellipse is fitted onto binary edge data and postprocessing techniques are used to eliminate detection errors. However, such approaches work well only with when the background is homogeneous and lightly to moderately cluttered. n [8], a technique is proposed for facial detection and coding. The detection of the face, as well as the facial features (eyes, nose, mouth, etc) is based on a parametric image model obtained using a Karhunen-Loève decomposition. The detected face image is then normalised by preprocessing, aiming to align the detected facial features with those of a standard model and also to compensate lighting and contrast variations. The normalised image is then coded by projecting it (by inner product) to a set of eigenfaces [13, 14], i.e. a subset of the eigenvectors obtained using a large database of normalised face images. If only a small number of primary eigenfaces are used, this technique results to successful face detection and coding using extremely low bitrates. Furthermore, the resulting face representations may additionally be used for face recognition applications. In [1], a 3-D ellipsoidal model is used for robust tracking of the rigid motion of the human head from a video sequence. This approach is based on the interpretation of the optical flow in terms of 3-D motion of the model. The method is seen to be robust to large head movements and to provide better results than those obtained using a simpler planar head model. Initialisation is based on predicting the shape of the ellipsoid from the positions of the face features (eyes, nose, mouth) which were previously estimated as in [8]. In [17], a hybrid system is presented using visual and sound cues for face localisation. The system exploits mainly luminance and motion information for locating and tracking the head edges from videoconferencing image sequences. In the present paper we shall assume a typical videoconferencing scene, with one person in front of the cameras. The initial set of 3-D points may be produced using depth estimation from stereoscopic or multiple views, provided that the camera geometry and calibration parameters are known. Furthermore, we assume that the outline of the person is available using some preprocessing technique. In the test
2 sequences, used for our experimental results, this outline is easily available since the background is homogeneous. A K-means algorithm, which is able to identify ellipses in the image, was then used for simultaneous ellipse fitting and segmentation of the outline. Specifically, a least squares fit of the data on a conic is performed, using an additional constraint, forcing the conic to be an ellipse. The algorithm is applied to estimate two ellipses, each corresponding to the head and body of the person. The basic ellipse fitting technique was also extended to the three dimensional case. This 3-D ellipsoid fitting technique was applied so as to estimate 3-D ellipsoids for the head of a person. Specifically, 3-D points that are projected inside each of the estimated ellipses in the previous step are used to estimate the 3-D ellipsoid model. Because of its high computational efficiency, this technique can be used for fast tracking of the head and body position and orientation. Since the segmentation of the nodes is only performed at the first time instant, only the estimation of the 3-D ellipsoid from the available data points has to be determined at subsequent instants. Finally, an interesting application of the proposed method is the combination of the 3-D ellipsoid model with a generic face model which is adapted to the face images to provide information only for the high-detail front part of the head, while the 3-D ellipsoid is used for the back of the head, which is usually not visible. 2. Estimating the initial set of 3-D points In order to estimate the initial 3-D point set, a multiocular system with N fully calibrated views is assumed. Each camera c k k =1 ::: N is modeled using a pinhole camera model based on perspective projection and we assume that accurate calibration information is available. A number of methods have been developed for depth estimation in stereoscopic [16, 15] and multiview [6, 4] image sequences. In this paper, we use the depth estimation and from trinocular image sequences developed in the ACTS 092 PANORAMA project. More specifically, depth is estimated from a fully calibrated trinocular image pair using the depth estimation algorithm in [4]. Furthermore, a reliability value is also estimated for depth estimate. A set of 3-D points where depth estimation with high reliability is then identified by selecting the 3-D points whose reliability exceeds a certain threshold. 3. Approximation of a 3-D data set using an 3- D Ellipsoid Model In [5], a new efficient method was presented for fitting ellipses to scattered 2-D data. The method is based on fitting the general equation of a conic to the given data, subject to a constraint which forces the conic to be an ellipse. In this paper, we develop an segmentation and estimation scheme using this technique in a K-means algorithm. Furthermore, we extend this technique for 3-D data, so that it can be used for approximating a general set of 3-D points using a 3-D ellipsoid. The general equation of a conic in the N-D space is: F (A b c x) =x T Ax + b T x + c =0 (1) where A is a symmetric matrix. In order for this general conic to be an ellipsoid in N-dimensions, the matrix A must be either positive or negative definite. As a result, the problem of fitting an ellipsoid into N data points x i can be solved by minimising the sum of squares of the algebraic distances NX d(a b c)= F (A b c x) 2 (2) i=1 subject to the constraint that A is either positive or negative definite. In general, this constrained problem is very difficult to solve since the Kuhn-Tucker conditions [9] do not guarantee a solution [5]. For the 2-D case, a direct solution to the problem was presented in [5]. The equation of the 2-D conic is F (a x) =ax = ax 2 + bxy + cy 2 + dx + ey + f =0 (3) where a = [a bcdef], x = [x 2 xy y 2 xy1] and the constraint to lead to an ellipse is 4ac ; b 2 > 0. Sincea is a free parameter, the equality constraint 4ac ; b 2 =1may be imposed instead. This in turn may be be written as ; a T Ca = a 26 T a =1 (4) 75 Using the above constraint, the following constrained fitting problem is formulated [5]: P N Problem 1 Minimise E = i=1 F 2 (a x i )=jjdajj 2 subject to the constraint a T Ca =1 where D = [x1 x2 x N ] T. This problem can be directly solved using a Lagrange multiplier, providing the following solution: a is the generalised eigenvector of D T Da = Ca (5) corresponding to the smallest positive eigenvalue.
3 In [5], it is proved that exactly one eigenvalue of eq. (5) is positive, thus a unique eigenvector a s is always found using the above method, which is non-iterative and thus extremely efficient. Even though this approach is not easily extended into the general N-D problem, a generalisation suitable for the 3-D case will now be formulated. The equation of a 3-D conic is: F (a x) =a x = = ax 2 +bxy +cy 2 +dxz +fz 2 +gz+hy +kz+l =0 (6) where a x are 10-dimensional vectors, can be written in the form of eq. (1), where A = U = 2 4 a b=2 d=2 b=2 c e=2 d=2 e=2 f a b=2 b=2 c 3 5 U v = v T f v =[d=2 e=2] T : The following Lemma provides the constraint forcing a general 3-D conic to be an ellipsoid. Lemma 1 A is positive definite or positive negative, provided that i. det(u) > 0 () 4ac ; b 2 > 0 ii. (a + c) det(a) > 0 Proof Let x =[xyz] T, u =[xy] T and u c = ;U ;1 vz. Then, x T Ax = u T Uu +2u T vz + fz 2 = (u ; u c ) T U(u ; u c )+z 2 (f ; v T U ;1 v)= (u ; u c ) T U(u ; u c )+ det(u) z2 (f det(u) ;v T c ;b=2 ;b=2 a v) = (u ; u c ) T U(u ; u c )+ z2 det(u) (acf ; fb2 =4 ;cd 2 =4 ; ae 2 =4+bde=4) = (u ; u c ) T U(u ; u c )+z 2 det(a) det(u) Assuming that U is positive definite, (u ; u c ) T U(u ; u c ) > 0 when x 6= 0. Furthermore, det(u) > 0 and tr(u) =a + c>0, thus if det(a) > 0, then eq. (7) yields x T Ax > 0 for x 6= 0, thus A is positive definite. Similarly, if U is negative definite, (u ; u c ) T U(u ; u c ) < 0, det(u) > 0 and tr(u) =a + c<0. Therefore, if det(a) < 0 then eq. (7) yields x T Ax < 0 for x 6= 0, thus A is negative definite, q.e.d. The first condition of Lemma 1 is very similar to the constraint 4ac ; b 2 > 0 used in the 2-D case. Therefore, we (7) can setup the problem of minimising the error E, asbefore using the constraint a T Ca =1,where: C = ; The solution is again obtained by solving Problem 1 using eq. (5) with the new C of (8). However, in this case, after obtaining the solution, we have to check whether the second condition of Lemma 1 is satisfied. If we select the sign of the solution eigenvector so that a + c>0, then the second condition is simplified to det(a) > Ellipse estimation and segmentation using K-means algorithm A significant advantage of the method described in the previous Section is its low computational requirements. This feature may be exploited by designing an algorithm for simultaneous ellipse estimation and segmentation of a set of 2-D points into ellipses. A simple algorithm that can be used for this purpose is a modified K-means algorithm, which is described below. In this version, we consider an algorithm using K =2,which is suitable for identifying the head and the body of a person from the person s outline in one of the available views (8) i. Initialisation: Divide the available points into K=2 sets, based on subsidiary information (e.g. relative position of head and body). ii. Fit an ellipse to each set, using the procedure described in the previous Section. iii. Reassign each point to the ellipse for which the distance F 2 (a x i ) in eq. (3) between the point x i and the ellipsoid is minimised. Points with distances larger than a given threshold are not assigned to any set. iv. End of the estimation procedure if the change in the estimated ellipsoid parameters is below a threshold else return to step ii.
4 5. Robustness and removal of outliers using Median of Least Squares When the initial 3-D point estimates are noisy, or contain outliers (e.g. points from the neck or the body of the person), the above method result to erroneous estimates of the 3-D ellipsoid parameters. Robustness can be achieved by using the MedLS technique which minimises the median of the errors corresponding to each 3-D point. Furthermore, this scheme allows the integration of the tests for the validity of the conditions forcing the estimated conic forms to be 3-D ellipsoids. More precisely, the procedure consists of the following steps: 1. Set E min to a large number. Randomly select a set of N 3-D points from the initial 3-D point set. For these points: 2. Estimate the ellipsoid parameters A b c. If the resulting conic is not an ellipsoid, discard the solution and return to Step Compute the Median of the Least Squares Fitting Error (E MedLS ) for each of point in the initial 3-D point set, using the estimated ellipsoid parameters. If E MedLS <E min,sete min = E MedLS and the corresponding parameters as the optimal parameters. 4. Iterate Steps 1-3 until E min is below a threshold or a maximum number of iterations is reached. 6 Combining 3-D Face Models with the 3-D Ellipsoid Another important application of the proposed methods is to combine a generic face model with a 3-D ellipsoid model estimated using the method described in Section 3. Various methods for adapting generic 3-D face models to face images have been proposed [12, 7, 3, 10]. In this paper, we have used the face adaptation method proposed in [10], where a generic 3D face model is transformed in a way that it adapts to the characteristics of an actual face observed in its front and profile image views. Assuming that a set of feature points have been detected on both views, the adaptation procedure initialises with a rigid transformation of the model aiming to minimise the distances of the 3D model feature nodes from the calculated 3D coordinates of the 2D feature points. Then, a non-rigid transformation ensures that the feature nodes are displaced optimally close to their exact calculated positions, dragging their neighbours in a way that does not deform the facial model in an unnatural way. In theory, any 3D model of the human face may be used. In practice however, the closer the characteristics of the model are to the characteristics of the image (e.g. model of female face for a female face) the fewer feature points will be needed for accurate adaptation. The estimated 3-D ellipsoid is then used to approximate the back of the head, while the adapted generic 3-D wireframe provides information for the high-detail front part of the head. In order to identify the part of the ellipsoid that corresponds to the back of the head, a check is performed as to whether the line that connects each node with the ellipsoid center intersects with any triangle of the wireframe corresponding to the front(face) wireframe. If so, the corresponding node is removed from the ellipsoid wireframe. Furthermore, each triangle in the ellipsoid wireframe containing this node is also removed. At this stage, two wireframe models corresponding to the front and the back part of the head have been generated. Since these parts are not connected, an algorithm was developed to automatically connect them. This is achieved by adding additional triangles between the list of nodes that define the largest wireframe boundary in each wireframe model. The largest wireframe boundary is simply the set of nodes corresponding to the largest path that can be defined using wireframe edges (edges belonging to only one triangle). 7. Experimental Results Results were obtained using the 3-view sequence Ludo The set of initial 3-D points is illustrated in Figure 1 (a triangulation was performed for display purposes). The approximation of the outline of the person with two ellipses corresponding to the head and the body area is shown in Figure 2. In this case a very simple initialisation was used: points lying higher (lower) than the mean height of the outline are attributed to the head (body) region. Convergence is achieved within a small number of iterations, typically 3, as shown in Figure 3, each requiring less than 0.5 sec at an SGI R4400 workstation. Then, the nodes of the 3-D wireframe of the head that are projected inside the head outline are used to estimate a 3-D ellipsoid model for the head. This model is illustrated in Figure 4. Table 1 shows the improvement at the performance of the proposed 3-D estimation technique when the MedLS technique is used to achieve robustness to noise. The comparison criterion used was the value of a fitting ratio r defined as follows: r = C intersection (9) C union where C intersection,c union are the number of pixels belonging to the intersection and the union respectively of the head area (defined manually after sketching the head contour) and the estimated ellipse obtained by projecting the 1 This sequence was provided by THOMPSON BROADCAST SYSTEMS for the the ACTS 092 project PANORAMA.
5 0.015 Error E Iterations Figure 3. Convergence of mean error E Figure 1. The initial 3-D wireframe Figure 2. The outline of the person (left view) and the two estimated ellipses using the K- means algorithm 3-D ellipsoid on the image plane. Larger values of the ratio r 2 (0 1] indicate better approximation of the head area with the ellipsoid. All methods were applied to 10 consecutive frames from the Ludo sequence (left view), and the average fitting ratio was calculated and is presented in Table 1. Two views of generated head models produced when combining a generic face model with the estimated 3-D ellipsoid model, as described in Section 6, are presented in Figures 5(a-b)(front and rear view respectively). 8. Conclusions An efficient technique to segment the head and body parts obtained from a head and shoulders multiview sequence and to estimate a 3-D ellipsoid model corresponding to the head was presented. A coarse approximation of an initial wireframe requiring minimal bitrate was thus obtained. Robustness and outlier removal was achieved by augmenting the above basic methods with either the Me- Figure 4. Estimated 3-D ellipsoid model superimposed on the head of Ludo (left view) dian of Least Squares (MedLS) technique which minimises the median of the errors corresponding to each 3-D point. Also, an application of the proposed method for combining of the 3-D ellipsoid model with a generic face model which is adapted to the face images was examined. 9. Acknowledgement This work was supported by the EU project ACTS VI- DAS, the GSRT PENED project and the GSRT AMEA Project: LIP-PHONE - A Videophone for the Deaf. Method fitting ratio r One-step 3-D parameter estimation MedLS Table 1. Improvement of fitting ratio r when using the MedLS robust estimation technique.
6 References [1] S. Basu, I. Essa, and A. Pentland. Motion Regularization for Model-based Head Tracking. In Proceedings, International Conference on Pattern Recognition, Vienna, Austria, August [2] A. Eleftheriadis and A. Jacquin. Automatic face location detection for model-assisted rate control in h.261 compatible coding of video. Signal Processing : Image Communication, 7: , November [3] M. Escher and N. Magnenat-Thalmann. Automatic 3d cloning and real-time animation of a human face. In Proceedings Computer Animation (CA97), Geneva, Switzerland, [4] L. Falkenhagen. Block-Based Depth Estimation from Image Triples with Unrestricted Camera Setup. In IEEE Workshop on Multimedia Image Processing, Princeton, NJ, June [5] A. W. Fitzgibbon, M. Pilu, and R. Fisher. Direct Least Squares Fitting of Ellipses. In Proc. International Conference on Pattern Recognition, Vienna, Austria, August [6] N. Grammalidis and M. G. Strintzis. Disparity and Occlusion Estimation in Multiocular Systems and their Coding for the Communication of Multiview Image Sequences. IEEE Trans. Circuits and Systems for Video Technology, 8(3): , June [7] N. Magnenat-Thalmann, P.Kalra, and M. Escher. Face to virtual face. Proc. IEEE, 86(5), [8] B. Moghaddam and A. Pentland. An Automatic System for Model-Based Coding of Faces. In IEEE Data Compression Conference, Snowbird, Utah, March [9] S. Rao. Optimization:Theory and Applications. Wiley Eastern, [10] N. Sarris and M. G. Strintzis. Three dimensional facial model adaptation. In submitted to ICIP 00, Vancouver, Canada, [11] S. Sirohey. Human Face Segmentation and Identification. Master s thesis, CV Laboratory, University of Maryland, College Park, MD, November [12] D. Terzopoulos and K. Waters. Physically-based facial modelling, analysis, and animation. The Journal of Visualization and Computer Animation, 1:73 80, [13] M. Turk and A. Pentland. Eigenfaces for recognition. Journal of Cognitive Neuro Science, 3(1):71 86, [14] M. A. Turk and A. P. Pentland. Face recognition using eigenfaces. In Proceedings IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages , Hawai, June [15] D. Tzovaras, N. Grammalidis, and M. G. Strintzis. Object- Based Coding of Stereo Image Sequences Using Joint 3D motion/disparity Compensation. IEEE Trans. on Ciscuits and Systems for Video Technology, 7(2): , April [16] D. Tzovaras, N. Grammalidis, and M. G. Strintzis. Disparity field and Depth Map Coding for Multiview 3D Image Generation. Signal Processing : Image Communication, 11(3): , January [17] C. Wang and M. Brandstein. A hybrid real-time face tracking system. In Proc. IEEE Int. Conf. Acoust., Speech and Signal Processing, Seattle, WA, May (a) (b) Figure 5. The combination of the generic face model and the 3-D ellipsoid model as viewed from two different viewpoints a) front view b) rear view.
Transactions on Information and Communications Technologies vol 19, 1997 WIT Press, ISSN
Hopeld Network for Stereo Correspondence Using Block-Matching Techniques Dimitrios Tzovaras and Michael G. Strintzis Information Processing Laboratory, Electrical and Computer Engineering Department, Aristotle
More informationMultiview Image Compression using Algebraic Constraints
Multiview Image Compression using Algebraic Constraints Chaitanya Kamisetty and C. V. Jawahar Centre for Visual Information Technology, International Institute of Information Technology, Hyderabad, INDIA-500019
More informationExploring Curve Fitting for Fingers in Egocentric Images
Exploring Curve Fitting for Fingers in Egocentric Images Akanksha Saran Robotics Institute, Carnegie Mellon University 16-811: Math Fundamentals for Robotics Final Project Report Email: asaran@andrew.cmu.edu
More informationSprite Generation and Coding in Multiview Image Sequences
302 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 10, NO. 2, MARCH 2000 Sprite Generation and Coding in Multiview Image Sequences Nikos Grammalidis, Student Member, IEEE, Dimitris
More informationStructure from Motion. Prof. Marco Marcon
Structure from Motion Prof. Marco Marcon Summing-up 2 Stereo is the most powerful clue for determining the structure of a scene Another important clue is the relative motion between the scene and (mono)
More informationLocal qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet:
Local qualitative shape from stereo without detailed correspondence Extended Abstract Shimon Edelman Center for Biological Information Processing MIT E25-201, Cambridge MA 02139 Internet: edelman@ai.mit.edu
More informationNEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING
NEW CONCEPT FOR JOINT DISPARITY ESTIMATION AND SEGMENTATION FOR REAL-TIME VIDEO PROCESSING Nicole Atzpadin 1, Serap Askar, Peter Kauff, Oliver Schreer Fraunhofer Institut für Nachrichtentechnik, Heinrich-Hertz-Institut,
More informationVideo Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin
Final Report Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This report describes a method to align two videos.
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 14 130307 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Review Stereo Dense Motion Estimation Translational
More informationSpeech Driven Synthesis of Talking Head Sequences
3D Image Analysis and Synthesis, pp. 5-56, Erlangen, November 997. Speech Driven Synthesis of Talking Head Sequences Peter Eisert, Subhasis Chaudhuri,andBerndGirod Telecommunications Laboratory, University
More informationStereo and Epipolar geometry
Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka
More informationImage Processing and Image Representations for Face Recognition
Image Processing and Image Representations for Face Recognition 1 Introduction Face recognition is an active area of research in image processing and pattern recognition. Since the general topic of face
More informationOn Modeling Variations for Face Authentication
On Modeling Variations for Face Authentication Xiaoming Liu Tsuhan Chen B.V.K. Vijaya Kumar Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213 xiaoming@andrew.cmu.edu
More informationRecognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17
Recognition (Part 4) CSE 152 Lecture 17 Announcements Homework 5 is due June 9, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying Images Chapter 17: Detecting Objects in Images
More informationProf. Fanny Ficuciello Robotics for Bioengineering Visual Servoing
Visual servoing vision allows a robotic system to obtain geometrical and qualitative information on the surrounding environment high level control motion planning (look-and-move visual grasping) low level
More informationFace Recognition At-a-Distance Based on Sparse-Stereo Reconstruction
Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,
More informationAnnouncements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test
Announcements (Part 3) CSE 152 Lecture 16 Homework 3 is due today, 11:59 PM Homework 4 will be assigned today Due Sat, Jun 4, 11:59 PM Reading: Chapter 15: Learning to Classify Chapter 16: Classifying
More informationAlgorithm research of 3D point cloud registration based on iterative closest point 1
Acta Technica 62, No. 3B/2017, 189 196 c 2017 Institute of Thermomechanics CAS, v.v.i. Algorithm research of 3D point cloud registration based on iterative closest point 1 Qian Gao 2, Yujian Wang 2,3,
More informationFace Detection and Recognition in an Image Sequence using Eigenedginess
Face Detection and Recognition in an Image Sequence using Eigenedginess B S Venkatesh, S Palanivel and B Yegnanarayana Department of Computer Science and Engineering. Indian Institute of Technology, Madras
More informationColour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation
ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology
More informationLOCALIZATION OF FACIAL REGIONS AND FEATURES IN COLOR IMAGES. Karin Sobottka Ioannis Pitas
LOCALIZATION OF FACIAL REGIONS AND FEATURES IN COLOR IMAGES Karin Sobottka Ioannis Pitas Department of Informatics, University of Thessaloniki 540 06, Greece e-mail:fsobottka, pitasg@zeus.csd.auth.gr Index
More informationRobust Lip Contour Extraction using Separability of Multi-Dimensional Distributions
Robust Lip Contour Extraction using Separability of Multi-Dimensional Distributions Tomokazu Wakasugi, Masahide Nishiura and Kazuhiro Fukui Corporate Research and Development Center, Toshiba Corporation
More informationMiniature faking. In close-up photo, the depth of field is limited.
Miniature faking In close-up photo, the depth of field is limited. http://en.wikipedia.org/wiki/file:jodhpur_tilt_shift.jpg Miniature faking Miniature faking http://en.wikipedia.org/wiki/file:oregon_state_beavers_tilt-shift_miniature_greg_keene.jpg
More informationcalibrated coordinates Linear transformation pixel coordinates
1 calibrated coordinates Linear transformation pixel coordinates 2 Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration with partial
More informationExpression Detection in Video. Abstract Expression detection is useful as a non-invasive method of lie detection and
Wes Miller 5/11/2011 Comp Sci 534 Expression Detection in Video Abstract Expression detection is useful as a non-invasive method of lie detection and behavior prediction, as many facial expressions are
More informationPrincipal Component Analysis and Neural Network Based Face Recognition
Principal Component Analysis and Neural Network Based Face Recognition Qing Jiang Mailbox Abstract People in computer vision and pattern recognition have been working on automatic recognition of human
More informationCOMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION
COMPARATIVE STUDY OF DIFFERENT APPROACHES FOR EFFICIENT RECTIFICATION UNDER GENERAL MOTION Mr.V.SRINIVASA RAO 1 Prof.A.SATYA KALYAN 2 DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PRASAD V POTLURI SIDDHARTHA
More informationStereo Image Rectification for Simple Panoramic Image Generation
Stereo Image Rectification for Simple Panoramic Image Generation Yun-Suk Kang and Yo-Sung Ho Gwangju Institute of Science and Technology (GIST) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Korea Email:{yunsuk,
More informationImage processing and features
Image processing and features Gabriele Bleser gabriele.bleser@dfki.de Thanks to Harald Wuest, Folker Wientapper and Marc Pollefeys Introduction Previous lectures: geometry Pose estimation Epipolar geometry
More informationTowards the completion of assignment 1
Towards the completion of assignment 1 What to do for calibration What to do for point matching What to do for tracking What to do for GUI COMPSCI 773 Feature Point Detection Why study feature point detection?
More informationIntroduction to Computer Vision
Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics
More informationWhat is Computer Vision?
Perceptual Grouping in Computer Vision Gérard Medioni University of Southern California What is Computer Vision? Computer Vision Attempt to emulate Human Visual System Perceive visual stimuli with cameras
More informationVisual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors
Visual Hulls from Single Uncalibrated Snapshots Using Two Planar Mirrors Keith Forbes 1 Anthon Voigt 2 Ndimi Bodika 2 1 Digital Image Processing Group 2 Automation and Informatics Group Department of Electrical
More informationUnsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme
Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Costas Panagiotakis and Anastasios Doulamis Abstract In this paper, an unsupervised, automatic video human members(human
More informationModel-based segmentation and recognition from range data
Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This
More informationAdaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision
Adaptive Zoom Distance Measuring System of Camera Based on the Ranging of Binocular Vision Zhiyan Zhang 1, Wei Qian 1, Lei Pan 1 & Yanjun Li 1 1 University of Shanghai for Science and Technology, China
More informationMulti-view stereo. Many slides adapted from S. Seitz
Multi-view stereo Many slides adapted from S. Seitz Beyond two-view stereo The third eye can be used for verification Multiple-baseline stereo Pick a reference image, and slide the corresponding window
More informationModel Based Perspective Inversion
Model Based Perspective Inversion A. D. Worrall, K. D. Baker & G. D. Sullivan Intelligent Systems Group, Department of Computer Science, University of Reading, RG6 2AX, UK. Anthony.Worrall@reading.ac.uk
More informationA Statistical Consistency Check for the Space Carving Algorithm.
A Statistical Consistency Check for the Space Carving Algorithm. A. Broadhurst and R. Cipolla Dept. of Engineering, Univ. of Cambridge, Cambridge, CB2 1PZ aeb29 cipolla @eng.cam.ac.uk Abstract This paper
More informationAutomatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera
Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera Wolfgang Niem, Jochen Wingbermühle Universität Hannover Institut für Theoretische Nachrichtentechnik und Informationsverarbeitung
More informationUsing temporal seeding to constrain the disparity search range in stereo matching
Using temporal seeding to constrain the disparity search range in stereo matching Thulani Ndhlovu Mobile Intelligent Autonomous Systems CSIR South Africa Email: tndhlovu@csir.co.za Fred Nicolls Department
More informationVIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural
VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD Ertem Tuncel and Levent Onural Electrical and Electronics Engineering Department, Bilkent University, TR-06533, Ankara, Turkey
More informationPRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1. We present some preliminary results on a system for tracking 3D motion using
PRELIMINARY RESULTS ON REAL-TIME 3D FEATURE-BASED TRACKER 1 Tak-keung CHENG derek@cs.mu.oz.au Leslie KITCHEN ljk@cs.mu.oz.au Computer Vision and Pattern Recognition Laboratory, Department of Computer Science,
More informationOptimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform
Optimized Progressive Coding of Stereo Images Using Discrete Wavelet Transform Torsten Palfner, Alexander Mali and Erika Müller Institute of Telecommunications and Information Technology, University of
More informationTHE GENERATION of a stereoscopic image sequence
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 15, NO. 8, AUGUST 2005 1065 Stereoscopic Video Generation Based on Efficient Layered Structure and Motion Estimation From a Monoscopic
More informationContents. 1 Introduction Background Organization Features... 7
Contents 1 Introduction... 1 1.1 Background.... 1 1.2 Organization... 2 1.3 Features... 7 Part I Fundamental Algorithms for Computer Vision 2 Ellipse Fitting... 11 2.1 Representation of Ellipses.... 11
More informationLocal Image Registration: An Adaptive Filtering Framework
Local Image Registration: An Adaptive Filtering Framework Gulcin Caner a,a.murattekalp a,b, Gaurav Sharma a and Wendi Heinzelman a a Electrical and Computer Engineering Dept.,University of Rochester, Rochester,
More informationMETRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS
METRIC PLANE RECTIFICATION USING SYMMETRIC VANISHING POINTS M. Lefler, H. Hel-Or Dept. of CS, University of Haifa, Israel Y. Hel-Or School of CS, IDC, Herzliya, Israel ABSTRACT Video analysis often requires
More informationCS201 Computer Vision Camera Geometry
CS201 Computer Vision Camera Geometry John Magee 25 November, 2014 Slides Courtesy of: Diane H. Theriault (deht@bu.edu) Question of the Day: How can we represent the relationships between cameras and the
More information3D Model Acquisition by Tracking 2D Wireframes
3D Model Acquisition by Tracking 2D Wireframes M. Brown, T. Drummond and R. Cipolla {96mab twd20 cipolla}@eng.cam.ac.uk Department of Engineering University of Cambridge Cambridge CB2 1PZ, UK Abstract
More informationECE 470: Homework 5. Due Tuesday, October 27 in Seth Hutchinson. Luke A. Wendt
ECE 47: Homework 5 Due Tuesday, October 7 in class @:3pm Seth Hutchinson Luke A Wendt ECE 47 : Homework 5 Consider a camera with focal length λ = Suppose the optical axis of the camera is aligned with
More informationStereo Vision. MAN-522 Computer Vision
Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in
More informationAn Adaptive Eigenshape Model
An Adaptive Eigenshape Model Adam Baumberg and David Hogg School of Computer Studies University of Leeds, Leeds LS2 9JT, U.K. amb@scs.leeds.ac.uk Abstract There has been a great deal of recent interest
More informationIRIS SEGMENTATION OF NON-IDEAL IMAGES
IRIS SEGMENTATION OF NON-IDEAL IMAGES William S. Weld St. Lawrence University Computer Science Department Canton, NY 13617 Xiaojun Qi, Ph.D Utah State University Computer Science Department Logan, UT 84322
More informationDepth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth
Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze
More informationMoving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial Region Segmentation
IJCSNS International Journal of Computer Science and Network Security, VOL.13 No.11, November 2013 1 Moving Object Segmentation Method Based on Motion Information Classification by X-means and Spatial
More informationMultiview Reconstruction
Multiview Reconstruction Why More Than 2 Views? Baseline Too short low accuracy Too long matching becomes hard Why More Than 2 Views? Ambiguity with 2 views Camera 1 Camera 2 Camera 3 Trinocular Stereo
More informationConversion of 2D Image into 3D and Face Recognition Based Attendance System
Conversion of 2D Image into 3D and Face Recognition Based Attendance System Warsha Kandlikar, Toradmal Savita Laxman, Deshmukh Sonali Jagannath Scientist C, Electronics Design and Technology, NIELIT Aurangabad,
More informationCS6670: Computer Vision
CS6670: Computer Vision Noah Snavely Lecture 7: Image Alignment and Panoramas What s inside your fridge? http://www.cs.washington.edu/education/courses/cse590ss/01wi/ Projection matrix intrinsics projection
More information2D to pseudo-3d conversion of "head and shoulder" images using feature based parametric disparity maps
University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2001 2D to pseudo-3d conversion of "head and shoulder" images using feature
More informationNonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.
Nonrigid Surface Modelling and Fast Recovery Zhu Jianke Supervisor: Prof. Michael R. Lyu Committee: Prof. Leo J. Jia and Prof. K. H. Wong Department of Computer Science and Engineering May 11, 2007 1 2
More informationTwo-view geometry Computer Vision Spring 2018, Lecture 10
Two-view geometry http://www.cs.cmu.edu/~16385/ 16-385 Computer Vision Spring 2018, Lecture 10 Course announcements Homework 2 is due on February 23 rd. - Any questions about the homework? - How many of
More informationAbstract We present a system which automatically generates a 3D face model from a single frontal image of a face. Our system consists of two component
A Fully Automatic System To Model Faces From a Single Image Zicheng Liu Microsoft Research August 2003 Technical Report MSR-TR-2003-55 Microsoft Research Microsoft Corporation One Microsoft Way Redmond,
More informationFace Tracking. Synonyms. Definition. Main Body Text. Amit K. Roy-Chowdhury and Yilei Xu. Facial Motion Estimation
Face Tracking Amit K. Roy-Chowdhury and Yilei Xu Department of Electrical Engineering, University of California, Riverside, CA 92521, USA {amitrc,yxu}@ee.ucr.edu Synonyms Facial Motion Estimation Definition
More informationInternational Journal of Advanced Research in Computer Science and Software Engineering
Volume 2, Issue 8, August 2012 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Study on Block
More informationMotion Tracking and Event Understanding in Video Sequences
Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Today: dense 3D reconstruction The matching problem
More informationPerception and Action using Multilinear Forms
Perception and Action using Multilinear Forms Anders Heyden, Gunnar Sparr, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: {heyden,gunnar,kalle}@maths.lth.se Abstract
More informationCHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION
CHAPTER 3 DISPARITY AND DEPTH MAP COMPUTATION In this chapter we will discuss the process of disparity computation. It plays an important role in our caricature system because all 3D coordinates of nodes
More informationComputer Vision Lecture 17
Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics 13.01.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Seminar in the summer semester
More informationLinear Discriminant Analysis in Ottoman Alphabet Character Recognition
Linear Discriminant Analysis in Ottoman Alphabet Character Recognition ZEYNEB KURT, H. IREM TURKMEN, M. ELIF KARSLIGIL Department of Computer Engineering, Yildiz Technical University, 34349 Besiktas /
More informationOptimal Grouping of Line Segments into Convex Sets 1
Optimal Grouping of Line Segments into Convex Sets 1 B. Parvin and S. Viswanathan Imaging and Distributed Computing Group Information and Computing Sciences Division Lawrence Berkeley National Laboratory,
More informationAN ADAPTIVE MESH METHOD FOR OBJECT TRACKING
AN ADAPTIVE MESH METHOD FOR OBJECT TRACKING Mahdi Koohi 1 and Abbas Shakery 2 1 Department of Computer Engineering, Islamic Azad University, Shahr-e-Qods Branch,Tehran,Iran m.kohy@yahoo.com 2 Department
More informationComputer Vision Lecture 17
Announcements Computer Vision Lecture 17 Epipolar Geometry & Stereo Basics Seminar in the summer semester Current Topics in Computer Vision and Machine Learning Block seminar, presentations in 1 st week
More informationA Robust Two Feature Points Based Depth Estimation Method 1)
Vol.31, No.5 ACTA AUTOMATICA SINICA September, 2005 A Robust Two Feature Points Based Depth Estimation Method 1) ZHONG Zhi-Guang YI Jian-Qiang ZHAO Dong-Bin (Laboratory of Complex Systems and Intelligence
More informationSupplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision
Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary
More informationFacial Deformations for MPEG-4
Facial Deformations for MPEG-4 Marc Escher, Igor Pandzic, Nadia Magnenat Thalmann MIRALab - CUI University of Geneva 24 rue du Général-Dufour CH1211 Geneva 4, Switzerland {Marc.Escher, Igor.Pandzic, Nadia.Thalmann}@cui.unige.ch
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More information3D Geometry and Camera Calibration
3D Geometry and Camera Calibration 3D Coordinate Systems Right-handed vs. left-handed x x y z z y 2D Coordinate Systems 3D Geometry Basics y axis up vs. y axis down Origin at center vs. corner Will often
More informationTwo-View Geometry (Course 23, Lecture D)
Two-View Geometry (Course 23, Lecture D) Jana Kosecka Department of Computer Science George Mason University http://www.cs.gmu.edu/~kosecka General Formulation Given two views of the scene recover the
More informationGender Classification Technique Based on Facial Features using Neural Network
Gender Classification Technique Based on Facial Features using Neural Network Anushri Jaswante Dr. Asif Ullah Khan Dr. Bhupesh Gour Computer Science & Engineering, Rajiv Gandhi Proudyogiki Vishwavidyalaya,
More informationRecovering light directions and camera poses from a single sphere.
Title Recovering light directions and camera poses from a single sphere Author(s) Wong, KYK; Schnieders, D; Li, S Citation The 10th European Conference on Computer Vision (ECCV 2008), Marseille, France,
More informationCS 565 Computer Vision. Nazar Khan PUCIT Lectures 15 and 16: Optic Flow
CS 565 Computer Vision Nazar Khan PUCIT Lectures 15 and 16: Optic Flow Introduction Basic Problem given: image sequence f(x, y, z), where (x, y) specifies the location and z denotes time wanted: displacement
More informationNoise Reduction in Image Sequences using an Effective Fuzzy Algorithm
Noise Reduction in Image Sequences using an Effective Fuzzy Algorithm Mahmoud Saeid Khadijeh Saeid Mahmoud Khaleghi Abstract In this paper, we propose a novel spatiotemporal fuzzy based algorithm for noise
More informationVideo Alignment. Literature Survey. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin
Literature Survey Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin Omer Shakil Abstract This literature survey compares various methods
More informationDense 3D Reconstruction. Christiano Gava
Dense 3D Reconstruction Christiano Gava christiano.gava@dfki.de Outline Previous lecture: structure and motion II Structure and motion loop Triangulation Wide baseline matching (SIFT) Today: dense 3D reconstruction
More informationAgenda. Rotations. Camera calibration. Homography. Ransac
Agenda Rotations Camera calibration Homography Ransac Geometric Transformations y x Transformation Matrix # DoF Preserves Icon translation rigid (Euclidean) similarity affine projective h I t h R t h sr
More informationCamera Calibration. COS 429 Princeton University
Camera Calibration COS 429 Princeton University Point Correspondences What can you figure out from point correspondences? Noah Snavely Point Correspondences X 1 X 4 X 3 X 2 X 5 X 6 X 7 p 1,1 p 1,2 p 1,3
More informationDetecting Ellipses via Bounding Boxes
Detecting Ellipses via Bounding Boxes CHUN-MING CHANG * Department of Information and Design, Asia University, Taiwan ABSTRACT A novel algorithm for ellipse detection based on bounding boxes is proposed
More informationMouse Pointer Tracking with Eyes
Mouse Pointer Tracking with Eyes H. Mhamdi, N. Hamrouni, A. Temimi, and M. Bouhlel Abstract In this article, we expose our research work in Human-machine Interaction. The research consists in manipulating
More informationMultiple View Geometry in Computer Vision
Multiple View Geometry in Computer Vision Prasanna Sahoo Department of Mathematics University of Louisville 1 Structure Computation Lecture 18 March 22, 2005 2 3D Reconstruction The goal of 3D reconstruction
More informationApplications Video Surveillance (On-line or off-line)
Face Face Recognition: Dimensionality Reduction Biometrics CSE 190-a Lecture 12 CSE190a Fall 06 CSE190a Fall 06 Face Recognition Face is the most common biometric used by humans Applications range from
More informationA Survey of Light Source Detection Methods
A Survey of Light Source Detection Methods Nathan Funk University of Alberta Mini-Project for CMPUT 603 November 30, 2003 Abstract This paper provides an overview of the most prominent techniques for light
More informationDepth Estimation for View Synthesis in Multiview Video Coding
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Depth Estimation for View Synthesis in Multiview Video Coding Serdar Ince, Emin Martinian, Sehoon Yea, Anthony Vetro TR2007-025 June 2007 Abstract
More informationEfficient Block Matching Algorithm for Motion Estimation
Efficient Block Matching Algorithm for Motion Estimation Zong Chen International Science Inde Computer and Information Engineering waset.org/publication/1581 Abstract Motion estimation is a key problem
More informationSegmentation and Tracking of Partial Planar Templates
Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract
More informationA Method of Automated Landmark Generation for Automated 3D PDM Construction
A Method of Automated Landmark Generation for Automated 3D PDM Construction A. D. Brett and C. J. Taylor Department of Medical Biophysics University of Manchester Manchester M13 9PT, Uk adb@sv1.smb.man.ac.uk
More informationFace detection and recognition. Many slides adapted from K. Grauman and D. Lowe
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe Face detection and recognition Detection Recognition Sally History Early face recognition systems: based on features and distances
More informationFinding a Best Fit Plane to Non-coplanar Point-cloud Data Using Non Linear and Linear Equations
AIJSTPME (013) 6(): 17-3 Finding a Best Fit Plane to Non-coplanar Point-cloud Data Using Non Linear and Linear Equations Mulay A. Production Engineering Department, College of Engineering, Pune, India
More information55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography Estimating homography from point correspondence
More information