Articulated Structure from Motion through Ellipsoid Fitting

Size: px
Start display at page:

Download "Articulated Structure from Motion through Ellipsoid Fitting"

Transcription

1 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV' Articulated Structure from Motion through Ellipsoid Fitting Peter Boyi Zhang, and Yeung Sam Hung Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, China {byzhang, Abstract - We present a new method to reconstruct non-rigid objects from orthographic projections based on the assumption of articulated model. We introduce an ellipsoid property to identify points belonging to a rigid subset. This enables us to formulate the problem of motion segmentation as an ellipsoid fitting problem. The obtained rigid subsets are then linked as kinematic chains to constitute 3D articulated structure. This method is practical of computational complexity O(N) mainly based on linear least squares. We demonstrate the effectiveness of this method through experiments on real tracking data, motion capture data, and challenging human dataset with missing data, in comparison with existing methods. Keywords: non-rigid structure from motion, articulated structure, ellipsoid fitting, motion segmentation, orthographic camera 1 Introduction The Non-Rigid Structure From Motion (NRSFM) problem aiming at recovering dynamic 3D structure from a sequence of 2D image measurements has attracted much attention over the years. As demonstrated in Figure 1, under affine projection, in each frame of measurement the depth information for all feature points needs to be estimated. This is an ill-posed problem unless additional constraints (e.g. based on the property of the object) are introduced to regulate the estimation to achieve meaningful results. Bregler et al. [1] introduce low rank condition on the object, constraining the object s movement to be a mean shape plus some degrees of freedom. Akhter et al. [2] also propose trajectory basis method based on this low rank assumption. Generally, low rank methods are only effective when the object is relatively rigid, and the movement is confined to be linear with few degrees of freedom. Rabaud and Belongie [3] tackle the problem assuming the shape of the object repeats itself at different times, and is thus captured in a few different frames. They identify these frames and recover the object in the same way as a rigid object. In a real scenario, many objects can fit to an articulated model. They can be modeled as a few rigid subsets connected to each other. With this assumption, as shown in Figure 1, feature points can be grouped into rigid subsets. Points in a same subset are fixed relative to each other. Furthermore, the relative depth between subsets can be recovered based on the knowledge of linkage between subsets. A major challenge of this problem is the grouping of feature points into rigid subsets, which amounts to motion segmentation. Tresadern and Reid [4] perform motion segmentation based on rank condition. Yan and Pollefeys [5] project points to low-rank manifolds and group points sharing the same subspace. Ross et al. [6] employ a probabilistic model to divide and merge the subsets iteratively. Russell et al. [7] propose an energy-based approach to label points as belonging to different subsets. The rank condition is not strong enough to accurately classify points into different subsets, and the number of subsets has to be provided a priori as an input. In this paper we show that beyond the low rank property, a group of points on a rigid structure also has the property of constituting an ellipsoid. The motion segmentation problem can thus be transformed to an ellipsoid model fitting problem. This helps achieve a more precise segmentation of feature points with low computational cost, and leads to better performance in 3D reconstruction. We describe the ellipsoid fitting method in section 2, and how to perform full 3D reconstruction in section 3. We demonstrate the effectiveness of the method through experiments on challenging datasets in section 4. Some conclusions are drawn in section 5. Figure 1. Without constraint, the feature points can have arbitrary depth. With the articulated assumption, the relative depths between feature points of the same rigid subset are fixed, and the relative depth between rigid subsets can be resolved by means of joint linkages. [8] 2 Motion segmentation Motion segmentation is a critical step for recovering articulated objects. In this section, we describe the problem, reveal the ellipsoid property of rigid subsets, and propose an efficient method to perform motion segmentation through ellipsoid fitting.

2 180 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV' Articulated object For an articulated object consisting of feature points measured over frames by an orthographic camera, let denote the number of rigid subsets ( is not known); let denote the measurements, and let denote the measurement of the th frame; let denote 3D structure of the th subset, where is the number of points in the th subset; let and denote the rotation and translation of the th subset in the th frame relative to a world coordinate. Let we have We denote, (1). (2),. (3) Through motion segmentation, we aim to group the feature points into subsets, so that each subset is rigid throughout all frames. To achieve that, firstly a property needs to be utilized to effectively classify the feature points into rigid subset; secondly an efficient method needs to be proposed to apply the property to different groups of points. We describe the ellipsoid property in section 2.2, and describe the grouping method in section Ellipsoid fitting The problem of grouping points into rigid subsets can be transformed to a problem of ellipsoid model fitting. In this section we reveal the ellipsoid property as a fundamental property of rigid structure under orthographic projection, and explain how ellipsoid fitting can help identify rigid subsets. An ellipsoid in centered at the origin is defined as a set of points satisfying subject to, (4), (5) where,, is a set of orthonormal vectors specifying the axis-direction of the ellipsoid;,, are constants representing the lengths of the semi-axes; and is a unit vector. Let denote the measurements of the th group of points. Subtract the mean of each row of from entries of the same row, we obtain a registered measurement matrix. Suppose this group is a rigid subset, we have. (6) We show rows of lie on an ellipsoid through the following analysis, and after that we will describe how to perform ellipsoid fitting. Express by SVD as then, (7). (8) We absorb the orthogonal matrix into. Let,, be the first three rows of ;,, be the diagonal elements of. Denote the entries of as then, (9). (10) According to (4) and (5), rows of lie on a same ellipsoid in. Moreover, the two rows of satisfy. (11) To fit ellipsoid to, we first project its row vectors to a 3D subspace. We register rows of to their centroid to get. As illustrated in Figure 2, each row of can be represented as a point in -dimensional space. We then perform a rank-3 approximation on, so that points in - dimensional space are projected to points in a 3D subspace, given by, (12) where,, is an orthonormal basis of the 3D subspace. (12) can be expressed in the form of (10) subject to (11) if and only if there exists, such that (13) This is equivalent to fitting an ellipsoid characterized by to the projected points. Since, we may write, (14)

3 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV' where is orthogonal. Then, (15). (16) In our method, we fit the ellipsoid by least squares as in (13), and then estimate using (15). Although this method of first fitting a subspace then an ellipsoid gives a sub-optimal solution compared to directly fitting an ellipsoid in, it greatly reduces the computational cost. The 3D shape of this subset can be recovered as We define 2D reprojection error. (17) (18) as a measurement of how much this group of points deviates from rigidity. Subset 1 Subset 2 Subset 3 Figure 2. Ellipsoid fitting. Each row in subset 3 is represented as a point in 5D space. The 5D space is projected to its most significant 3D subspace, and an ellipsoid is fitted to the points. 2.3 Grouping of points as rigid subsets Through the previous analysis, the problem of finding rigid subsets among feature points can be transformed to a problem of finding appropriate ellipsoids parameterized by and, so that each group of columns in can be fitted to an ellipsoid as. (19) To fit an ellipsoid to any possible group of points, there would be as many as groups to test. We propose a method to perform ellipsoid model fitting with only complexity. We first estimate the model by fitting ellipsoid to a minimal seed, and then add inliers to the estimated model. A minimal seed contains 4 feature points, because registering each row by its mean reduces the rank of the measurement matrix by 1, and we need at least 3 points to fit an ellipsoid centered at the origin in a 3D subspace. In a real scenario points of a rigid subset are usually close to each other, so in our case, instead of randomly selecting groups of 4 points as seeds, we only consider points close to each other. This would only require ellipsoid fitting computations. The groups with ellipsoid fitting error below a threshold are chosen as seeds for rigid subsets. For a seed and the fitted model,, we evaluate of each column in through reprojection as in (18), and include the points with to expand the rigid subset. As the rigid subset expands, and may be updated to minimize. In total, this process requires ellipsoid fitting and projections. After we expand each rigid seed, we might obtain a number of rigid subsets overlapping with each other. This is because in a real scenario the object might not be a perfectly articulated object, and there may be ambiguity as to which subset the points at the boundary should belong to. We resolve this issue by first picking subsets with minimal overlap, and then assign the points belonging to more than one rigid subsets to the subset of minimal reprojection error. If some points are not grouped to any rigid subset, they tend to introduce large error. Either we can choose to discard them; or we can include each point to the rigid subset for which it has minimal at the final stage after the recovery of and, to reduce their influence. The ability to distinguish outliers is an advantage of this algorithm. As we shall show in our experiments, the range of choice of to provide optimal result is wide. As increases from 0, the number of ungrouped points decreases, and the number of duplicate points (points occurring in more than one subsets) increases. As reaches the value that well distinguishes the rigid subsets, the number of ungrouped and duplicate points are kept small and stable. The threshold should be small if the object can be well modeled as an articulated object. should be large if the object is not strictly articulated and/or there is much noise. We shall also show that the 2D reprojection error can be taken as an indicator of the 3D error, for there is a correlation between them. 3 Reconstruction of 3D structure After segmenting the feature points into rigid subsets, we describe how to reconstruct the 3D structure in this section. We connect the subsets through kinematic chain, eliminate mirror ambiguity based on the linkage between subsets, resolve the relative depth between subsets, and handle missing data. 3.1 Kinematic chain Similar to Kirk et al. [9] and Yan and Pollefeys [5], we construct a graph recording the cost of connecting every two rigid subsets through joints (called joint cost), and build kinematic chains through performing minimum spanning tree

4 182 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV'15 search in the graph; however, we calculate the joint cost through a different method that only requires least squares computation. If two rigid subsets and can be connected through a link, we should be able to find virtual points and attached to the coordinate frame of and, respectively, such that the distance between trajectory of and is small (ideally zero). Place the coordinate frame of at the centroid of, can be recovered as the row mean of. Let denote the coefficients combining trajectories of to, the trajectory of as observed in the measurement is Substitute (6) into (20),. (20). (21) In our model, the joint points and are close to each other, but may not strictly coincide with each other. This relaxed condition allows us to model objects in the physical world that are not perfectly articulated, e.g. the human body. By (21), can be approximated directly from, without recovery of and. The joint cost can be found through least squares minimization as:. (22) We construct an undirected graph with the rigid subsets as nodes, and joint costs as weights of edges. In this graph we perform minimum spanning tree search to recover the connectivity among rigid subsets and then kinematic chains from connected rigid subsets. 3.2 Mirror ambiguity and can be recovered through (15) and (17). Yet a mirror ambiguity remains unresolved, since and is an equally acceptable solution. We resolve this ambiguity using the physical condition that two subsets linked by a joint generally have similar rotational transformation. For two subsets labeled and linked by joints, we align and with, and pick the camera motion with smaller error. We arbitrarily pick one subset as reference and propagate alignment to its neighbors and so on. As a result, we will only have one overall mirror ambiguity left to be handled manually D structure As we have recovered,, the exact joint location in the coordinate frame of can be reevaluated by applying (20) and (22) to subsets linked by joint. Now the only parameter remains to be estimated is the depth, the third element of. For a set of linked subsets and, we set the relative depth between and to be in each frame to minimize the joint distance, assigning one subset as the reference, the depth of its neighboring subsets can be solved through vector addition as shown in Figure 3. Figure 3. Two subsets linked by joints. The circles represent recovered joint locations. (Best viewed in color.) 3.4 Missing data Missing data does not present any problem and can be handled in a natural way in our method. For points that are visible in incomplete set of frames, we only use information in the observable frames to perform motion segmentation for these points. Once the points are grouped into rigid subsets, we can estimate the rigid substructure and its transformation in all frames. 4 Experimental Results We evaluate our method on both motion capture data and real tracking data. The motion capture datasets include 3D ground truth, providing a foundation for quantitative measurement of the accuracy of the algorithm. We evaluate the 2D reprojection error against the measurement matrix, and the 3D error against the 3D ground truth. The 2D reprojection error is defined as the Frobenius norm of the difference between 2D measurement and 2D reprojection, divided by. The 3D error is defined as the Frobenius norm of the difference between 3D ground truth and reconstructed 3D points, divided by the Frobenius norm of 3D ground truth. We evaluate our method qualitatively on real scenes two cranes and toy truck from [5]. We also perform 3D reconstruction on toy truck. The robustness of our method in the presence of missing data is demonstrated through motion capture data skin. Depending on its size, each of these datasets takes two seconds to a few minutes for an intel-i5 3.10GHz PC to process.

5 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV' Motion capture data The stepstool dataset taken from subject 40 trial 6 of Carnegie Mellon University Motion Capture Database [10] records a person that climbs, steps over, sits on, and jumps over a stepstool. The camera is fixed while the person performs substantial movements, making it a very challenging NRSFM problem. Before testing our algorithm, we down sampled the number of frames by a factor of 6, and removed the duplicate point tracks, resulting in 1097 frames and 149 points. From the 3D trajectory of all the tracks, we retain the information in and axis as 2D measurements. Dataset dance from [2] is a standard NRSFM dataset evaluated by many methods, it has 264 frames and 75 points. The camera rotates around the person horizontally for about 720 degrees throughout the frames. is set to 0.3% and 2.2% times for stepstool and dance respectively. Table 1. 3D error. Trajectory basis Kernel Ellipsoid Fitting Stepstool 22.70% 20.73% 1.60% Dance 18.99% 16.90% 8.85% We list the results in comparison with the trajectory basis method [2] and the kernel method by Gotardo and Martines [11] in Table 1. We calculate the 3D error after shifting each frame of the 3D ground truth to the centroid. Our method reduces the 3D error by more than an order of magnitude in stepstool, and by a factor of two in dance. This shows the limitation of methods based on low-rank assumptions for reconstructing highly flexible objects. Figure 4 (a) shows the effect of varying. The leading to optimal result is chosen where the number of ungrouped and duplicate points are small, the error is robust to change of, and the 3D error correlates with 2D error. Figure 4 (b) and (c) show the recovered kinematic chains. We observe that as few as 4 points is enough to determine a rigid subset. For comparison, we include result reported by Ross et al. [6] on stepstool to the right. Our method results in 20 subsets, distinguishing the chest, the waist and the hip, and also the shoulder from the upper arm. Although [6] use 3D data instead of 2D measurements as input, they can only identify 15 subsets. Moreover, we recover the joints at more reasonable positons. Figure 7 shows 3D stepstool results in different frames. 4.2 Real tracking data Data two cranes has 30 frames and 94 points. One crane rotates relative to the camera, while the other crane has two moving parts, but little camera motion. As shown in Figure 5, we segment the data into 3 parts, while Fayad et al. [12] further segment one more part on each crane, at least one of which appears to be redundant. Compared with Yan and Pollefeys result [5] where two points are misclassified as belonging to the other crane, our result is also better. Because there is almost no camera motion to provide depth information, it is not possible to perform 3D reconstruction. Yet this experiment shows that our method is capable of separating the rigid subsets as long as there is relative motion Frame 1 Frame 30 (a) (b) (c) Figure 4. (a) Effect of varying. Recovered kinematic structure of (b) dance, and (c) stepstool in comparison with result by Ross et al. [6] (Best viewed in color. See Figure 7 for meaning of symbols and coloring.) Figure 5. Experiments on two crane. From top to bottom are results by our method, Fayad et al. [12], and Yan and Pollefeys [5].

6 184 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV'15 Frame 1 Frame 50 Figure 6. Experiments on toy truck. The first row shows the image and segmentation result, the second shows the 3D reconstruction result. between them. Data toy truck has 60 frames and 83 points. It rotates slightly with respect to the camera, and its shovel moves up and down. Figure 6 shows that our method accurately differentiates the shovel from the rest of the truck. We also perform 3D reconstruction. As there is no ground truth for this type of real data, the reconstruction can only be assessed qualitatively. Figure 6 shows textured 3D views of frame 1 with shovel down and frame 50 with shovel up from a different angle, which look reasonable. 4.3 Data with missing observations The skin dataset from [13] depicts a person flexing his muscles. Approximately 350 markers are placed on the subject to capture the subtle movement of human skin. But since the motion capture system sometimes lose track of the markers because of occlusion, as a result 467 tracks are Stepstool Frame 81, climb Frame 185, sit Frame 387, stride Frame 456, pose Frame 643, jump Scene Setup Camera view Side view Skin Frame 16 Frame 32 Frame 52 Frame 68 Frame 92 Scene Setup [13] Camera view Side view Figure 7. Results on datasets stepstool and skin. The reconstructed points are colored according to different subsets, the gray circles represent ground truth, the colored circles are joint locations, and the black lines are recovered kinematic chains. (Best viewed in color.)

7 Int'l Conf. IP, Comp. Vision, and Pattern Recognition IPCV' captured, some full and some partial. Thus we have 97 frames and 467 points. In total 31.75% of data are missing in the measurements. The camera motion is also relatively small (about 45 degrees), making this very challenging data for NRSFM problem. We process this data with set from 0.5% to 0.9% times, resulting in 2D reprojection error ranging from 0.60% to 1.02%, and 3D error ranging from 2.85% to 3.94%. Figure 7 shows the segmentation and reconstruction with set to 0.8% times. The left foot exhibits larger 3D error than other rigid subsets because its motion relative to camera is very small. Comparatively, Fayad et al. [13] s reconstruction from the same dataset based on the same input and same measure of error results in a 3D error of 7.13%. Our 3D error is about half of theirs. Our segmentation further divides the upper body and the arm, capturing the subtle movement of the skin. 5 Conclusion We have presented a method to systematically reconstruct dynamic 3D structure from orthographic 2D measurements based on the articulated model assumption. We introduce the ellipsoid property as a property to effectively distinguish feature points belonging to different rigid subsets. This enables us to convert the motion segmentation problem to ellipsoid model fitting, and propose a practical method with computational complexity. Then, joint constraints are used to build kinematic chains and 3D articulated structures are recovered. The method is proven effective on both standard real tracking data and challenging human motion capture datasets with missing data, where the proposed method yields better motion segmentation and significant improvements in both 2D and 3D errors compared to existing methods. 6 Acknowledgements The work described in this paper was supported by the Research Grants Council of the Hong Kong Special Administrative Region, China (Project No. HKU E) and CRCG of the University of Hong Kong. We thank João Fayad for his kindness in sending us the datasets necessary to carry out the experiments. 7 References [1] C. Bregler, A. Hertzmann, and H. Biermann. Recovering non-rigid 3D shape from image streams. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, [3] V. Rabaud, and S. Belongie. Re-thinking non-rigid structure from motion. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, [4] P. Tresadern, and I. Reid. Articulated structure from motion by factorization. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, [5] J. Yan, and M. Pollefeys. A factorization-based approach for articulated non-rigid shape, motion and kinematic chain recovery from video. In IEEE Transactions on Pattern Analysis and Machine Intelligence, [6] D. Ross, D. Tarlow, and R. Zemel. Learning articulated structure and motion. In International Journal of Computer Vision, 88(2): , [7] C. Russell, J. Fayad, and L. Agapito. Energy based multiple model fitting for non-rigid structure from motion. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, [8] Robot arm model, [9] A. Kirk, J. O'Brien, and D. Forsyth. Skeletal parameter estimation from optical motion capture data. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, [10] Carnegie Mellon University motion capture database, [11] P. F. U. Gotardo, and A. M. Martinez. Kernal non-rigid structure from motion. In Proceedings of IEEE International Conference on Computer Vision, [12] J. Fayad, C. Russell, and L. Agapito. Automated articulated structure and 3D shape recovery from point correspondences. In Proceedings of IEEE International Conference on Computer Vision, [13] S. I. Park, and J. K. Hodgins. Capturing and animating skin deformation in human motion. In Proceedings of the ACM SIGGRAPH Conference on Computer Graphics, [2] I. Akhter, Y. Sheikh, S. Khan, and T. Kanade. Nonrigid structure from motion in trajectory space. In Neural Information Processing Systems, 2008.

Learning Articulated Skeletons From Motion

Learning Articulated Skeletons From Motion Learning Articulated Skeletons From Motion Danny Tarlow University of Toronto, Machine Learning with David Ross and Richard Zemel (and Brendan Frey) August 6, 2007 Point Light Displays It's easy for humans

More information

Unsupervised learning in Vision

Unsupervised learning in Vision Chapter 7 Unsupervised learning in Vision The fields of Computer Vision and Machine Learning complement each other in a very natural way: the aim of the former is to extract useful information from visual

More information

C280, Computer Vision

C280, Computer Vision C280, Computer Vision Prof. Trevor Darrell trevor@eecs.berkeley.edu Lecture 11: Structure from Motion Roadmap Previous: Image formation, filtering, local features, (Texture) Tues: Feature-based Alignment

More information

Graph-based High Level Motion Segmentation using Normalized Cuts

Graph-based High Level Motion Segmentation using Normalized Cuts Graph-based High Level Motion Segmentation using Normalized Cuts Sungju Yun, Anjin Park and Keechul Jung Abstract Motion capture devices have been utilized in producing several contents, such as movies

More information

Automatic Kinematic Chain Building from Feature Trajectories of Articulated Objects

Automatic Kinematic Chain Building from Feature Trajectories of Articulated Objects Automatic Kinematic Chain Building from Feature Trajectories of Articulated Objects Jingyu Yan and Marc Pollefeys Department of Computer Science The University of North Carolina at Chapel Hill Chapel Hill,

More information

Data-driven Approaches to Simulation (Motion Capture)

Data-driven Approaches to Simulation (Motion Capture) 1 Data-driven Approaches to Simulation (Motion Capture) Ting-Chun Sun tingchun.sun@usc.edu Preface The lecture slides [1] are made by Jessica Hodgins [2], who is a professor in Computer Science Department

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Modeling 3D Human Poses from Uncalibrated Monocular Images

Modeling 3D Human Poses from Uncalibrated Monocular Images Modeling 3D Human Poses from Uncalibrated Monocular Images Xiaolin K. Wei Texas A&M University xwei@cse.tamu.edu Jinxiang Chai Texas A&M University jchai@cse.tamu.edu Abstract This paper introduces an

More information

In Defense of Orthonormality Constraints for Nonrigid Structure from Motion

In Defense of Orthonormality Constraints for Nonrigid Structure from Motion In Defense of Orthonormality Constraints for Nonrigid Structure from Motion Ijaz Akhter 1 Yaser Sheikh 2 Sohaib Khan 1 akhter@lumsedupk yaser@cscmuedu sohaib@lumsedupk 1 LUMS School of Science and Engineering,

More information

Factorization with Missing and Noisy Data

Factorization with Missing and Noisy Data Factorization with Missing and Noisy Data Carme Julià, Angel Sappa, Felipe Lumbreras, Joan Serrat, and Antonio López Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona,

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00. Topic (Research Paper):

Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00. Topic (Research Paper): Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00 Topic (Research Paper): Jinxian Chai and Jessica K. Hodgins, Performance Animation

More information

NON-RIGID 3D SHAPE RECOVERY USING STEREO FACTORIZATION

NON-RIGID 3D SHAPE RECOVERY USING STEREO FACTORIZATION NON-IGID D SHAPE ECOVEY USING STEEO FACTOIZATION Alessio Del Bue ourdes Agapito Department of Computer Science Queen Mary, University of ondon ondon, E1 NS, UK falessio,lourdesg@dcsqmulacuk ABSTACT In

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribes: Jeremy Pollock and Neil Alldrin LECTURE 14 Robust Feature Matching 14.1. Introduction Last lecture we learned how to find interest points

More information

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction

Dynamic Time Warping for Binocular Hand Tracking and Reconstruction Dynamic Time Warping for Binocular Hand Tracking and Reconstruction Javier Romero, Danica Kragic Ville Kyrki Antonis Argyros CAS-CVAP-CSC Dept. of Information Technology Institute of Computer Science KTH,

More information

Stereo and Epipolar geometry

Stereo and Epipolar geometry Previously Image Primitives (feature points, lines, contours) Today: Stereo and Epipolar geometry How to match primitives between two (multiple) views) Goals: 3D reconstruction, recognition Jana Kosecka

More information

Occluded Facial Expression Tracking

Occluded Facial Expression Tracking Occluded Facial Expression Tracking Hugo Mercier 1, Julien Peyras 2, and Patrice Dalle 1 1 Institut de Recherche en Informatique de Toulouse 118, route de Narbonne, F-31062 Toulouse Cedex 9 2 Dipartimento

More information

Robust Model-Free Tracking of Non-Rigid Shape. Abstract

Robust Model-Free Tracking of Non-Rigid Shape. Abstract Robust Model-Free Tracking of Non-Rigid Shape Lorenzo Torresani Stanford University ltorresa@cs.stanford.edu Christoph Bregler New York University chris.bregler@nyu.edu New York University CS TR2003-840

More information

Piecewise Quadratic Reconstruction of Non-Rigid Surfaces from Monocular Sequences

Piecewise Quadratic Reconstruction of Non-Rigid Surfaces from Monocular Sequences Piecewise Quadratic Reconstruction of Non-Rigid Surfaces from Monocular Sequences João Fayad 1, Lourdes Agapito 1, and Alessio Del Bue 2 1 Queen Mary University of London. Mile End Road, London E1 4NS,

More information

Markerless human motion capture through visual hull and articulated ICP

Markerless human motion capture through visual hull and articulated ICP Markerless human motion capture through visual hull and articulated ICP Lars Mündermann lmuender@stanford.edu Stefano Corazza Stanford, CA 93405 stefanoc@stanford.edu Thomas. P. Andriacchi Bone and Joint

More information

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H. Nonrigid Surface Modelling and Fast Recovery Zhu Jianke Supervisor: Prof. Michael R. Lyu Committee: Prof. Leo J. Jia and Prof. K. H. Wong Department of Computer Science and Engineering May 11, 2007 1 2

More information

Multiple Motion Scene Reconstruction from Uncalibrated Views

Multiple Motion Scene Reconstruction from Uncalibrated Views Multiple Motion Scene Reconstruction from Uncalibrated Views Mei Han C & C Research Laboratories NEC USA, Inc. meihan@ccrl.sj.nec.com Takeo Kanade Robotics Institute Carnegie Mellon University tk@cs.cmu.edu

More information

Pose estimation using a variety of techniques

Pose estimation using a variety of techniques Pose estimation using a variety of techniques Keegan Go Stanford University keegango@stanford.edu Abstract Vision is an integral part robotic systems a component that is needed for robots to interact robustly

More information

SCAPE: Shape Completion and Animation of People

SCAPE: Shape Completion and Animation of People SCAPE: Shape Completion and Animation of People By Dragomir Anguelov, Praveen Srinivasan, Daphne Koller, Sebastian Thrun, Jim Rodgers, James Davis From SIGGRAPH 2005 Presentation for CS468 by Emilio Antúnez

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

Non-rigid structure from motion using quadratic deformation models

Non-rigid structure from motion using quadratic deformation models FAYAD et al.: NON-RIGID SFM USING QUADRATIC DEFORMATION MODELS Non-rigid structure from motion using quadratic deformation models João Fayad jrkf@dcs.qmul.ac.uk Alessio Del Bue adb@isr.ist.utl.pt Lourdes

More information

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models

A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models A novel approach to motion tracking with wearable sensors based on Probabilistic Graphical Models Emanuele Ruffaldi Lorenzo Peppoloni Alessandro Filippeschi Carlo Alberto Avizzano 2014 IEEE International

More information

Occlusion-Aware Reconstruction and Manipulation of 3D Articulated Objects

Occlusion-Aware Reconstruction and Manipulation of 3D Articulated Objects Occlusion-Aware Reconstruction and Manipulation of 3D Articulated Objects Xiaoxia Huang, Ian Walker, and Stan Birchfield Department of Electrical and Computer Engineering Clemson University, Clemson, SC

More information

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Accurate 3D Face and Body Modeling from a Single Fixed Kinect Accurate 3D Face and Body Modeling from a Single Fixed Kinect Ruizhe Wang*, Matthias Hernandez*, Jongmoo Choi, Gérard Medioni Computer Vision Lab, IRIS University of Southern California Abstract In this

More information

CS 664 Structure and Motion. Daniel Huttenlocher

CS 664 Structure and Motion. Daniel Huttenlocher CS 664 Structure and Motion Daniel Huttenlocher Determining 3D Structure Consider set of 3D points X j seen by set of cameras with projection matrices P i Given only image coordinates x ij of each point

More information

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems

Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Partial Calibration and Mirror Shape Recovery for Non-Central Catadioptric Systems Abstract In this paper we present a method for mirror shape recovery and partial calibration for non-central catadioptric

More information

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Structured Light II Johannes Köhler Johannes.koehler@dfki.de Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov Introduction Previous lecture: Structured Light I Active Scanning Camera/emitter

More information

Stereo Vision. MAN-522 Computer Vision

Stereo Vision. MAN-522 Computer Vision Stereo Vision MAN-522 Computer Vision What is the goal of stereo vision? The recovery of the 3D structure of a scene using two or more images of the 3D scene, each acquired from a different viewpoint in

More information

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation

Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation ÖGAI Journal 24/1 11 Colour Segmentation-based Computation of Dense Optical Flow with Application to Video Object Segmentation Michael Bleyer, Margrit Gelautz, Christoph Rhemann Vienna University of Technology

More information

Nonrigid Structure from Motion in Trajectory Space

Nonrigid Structure from Motion in Trajectory Space Nonrigid Structure from Motion in Trajectory Space Ijaz Akhter LUMS School of Science and Engineering Lahore, Pakistan akhter@lumsedupk Sohaib Khan LUMS School of Science and Engineering Lahore, Pakistan

More information

Perception and Action using Multilinear Forms

Perception and Action using Multilinear Forms Perception and Action using Multilinear Forms Anders Heyden, Gunnar Sparr, Kalle Åström Dept of Mathematics, Lund University Box 118, S-221 00 Lund, Sweden email: {heyden,gunnar,kalle}@maths.lth.se Abstract

More information

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO Stefan Krauß, Juliane Hüttl SE, SoSe 2011, HU-Berlin PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO 1 Uses of Motion/Performance Capture movies games, virtual environments biomechanics, sports science,

More information

The Template Update Problem

The Template Update Problem The Template Update Problem Iain Matthews, Takahiro Ishikawa, and Simon Baker The Robotics Institute Carnegie Mellon University Abstract Template tracking dates back to the 1981 Lucas-Kanade algorithm.

More information

Structure from Motion. Lecture-15

Structure from Motion. Lecture-15 Structure from Motion Lecture-15 Shape From X Recovery of 3D (shape) from one or two (2D images). Shape From X Stereo Motion Shading Photometric Stereo Texture Contours Silhouettes Defocus Applications

More information

Sequential Non-Rigid Structure-from-Motion with the 3D-Implicit Low-Rank Shape Model

Sequential Non-Rigid Structure-from-Motion with the 3D-Implicit Low-Rank Shape Model Sequential Non-Rigid Structure-from-Motion with the 3D-Implicit Low-Rank Shape Model Marco Paladini 1, Adrien Bartoli 2, and Lourdes Agapito 1 1 Queen Mary University of London, Mile End Road, E1 4NS London,

More information

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University.

3D Computer Vision. Structured Light II. Prof. Didier Stricker. Kaiserlautern University. 3D Computer Vision Structured Light II Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de 1 Introduction

More information

Non-Rigid Structure from Motion with Complementary Rank-3 Spaces

Non-Rigid Structure from Motion with Complementary Rank-3 Spaces Non-Rigid Structure from Motion with Complementary Rank-3 Spaces Paulo FU Gotardo and Aleix M Martinez Department of Electrical and Computer Engineering The Ohio State University, Columbus, OH 43210, USA

More information

Affine-invariant shape matching and recognition under partial occlusion

Affine-invariant shape matching and recognition under partial occlusion Title Affine-invariant shape matching and recognition under partial occlusion Author(s) Mai, F; Chang, CQ; Hung, YS Citation The 17th IEEE International Conference on Image Processing (ICIP 2010), Hong

More information

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Visual Odometry. Features, Tracking, Essential Matrix, and RANSAC. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Visual Odometry Features, Tracking, Essential Matrix, and RANSAC Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline The

More information

Non-Rigid Shape and Motion Recovery: Degenerate Deformations

Non-Rigid Shape and Motion Recovery: Degenerate Deformations Non-Rigid Shape and Motion Recovery: Degenerate Deformations Jing Xiao Takeo Kanade The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 {jxiao, tk}@cscmuedu Abstract This paper studies

More information

Registration of Dynamic Range Images

Registration of Dynamic Range Images Registration of Dynamic Range Images Tan-Chi Ho 1,2 Jung-Hong Chuang 1 Wen-Wei Lin 2 Song-Sun Lin 2 1 Department of Computer Science National Chiao-Tung University 2 Department of Applied Mathematics National

More information

CS 231A Computer Vision (Winter 2014) Problem Set 3

CS 231A Computer Vision (Winter 2014) Problem Set 3 CS 231A Computer Vision (Winter 2014) Problem Set 3 Due: Feb. 18 th, 2015 (11:59pm) 1 Single Object Recognition Via SIFT (45 points) In his 2004 SIFT paper, David Lowe demonstrates impressive object recognition

More information

Two-View Geometry (Course 23, Lecture D)

Two-View Geometry (Course 23, Lecture D) Two-View Geometry (Course 23, Lecture D) Jana Kosecka Department of Computer Science George Mason University http://www.cs.gmu.edu/~kosecka General Formulation Given two views of the scene recover the

More information

EECS 442: Final Project

EECS 442: Final Project EECS 442: Final Project Structure From Motion Kevin Choi Robotics Ismail El Houcheimi Robotics Yih-Jye Jeffrey Hsu Robotics Abstract In this paper, we summarize the method, and results of our projective

More information

Motion Tracking and Event Understanding in Video Sequences

Motion Tracking and Event Understanding in Video Sequences Motion Tracking and Event Understanding in Video Sequences Isaac Cohen Elaine Kang, Jinman Kang Institute for Robotics and Intelligent Systems University of Southern California Los Angeles, CA Objectives!

More information

CSE 252B: Computer Vision II

CSE 252B: Computer Vision II CSE 252B: Computer Vision II Lecturer: Serge Belongie Scribe: Haowei Liu LECTURE 16 Structure from Motion from Tracked Points 16.1. Introduction In the last lecture we learned how to track point features

More information

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet:

Local qualitative shape from stereo. without detailed correspondence. Extended Abstract. Shimon Edelman. Internet: Local qualitative shape from stereo without detailed correspondence Extended Abstract Shimon Edelman Center for Biological Information Processing MIT E25-201, Cambridge MA 02139 Internet: edelman@ai.mit.edu

More information

A Factorization Method for Structure from Planar Motion

A Factorization Method for Structure from Planar Motion A Factorization Method for Structure from Planar Motion Jian Li and Rama Chellappa Center for Automation Research (CfAR) and Department of Electrical and Computer Engineering University of Maryland, College

More information

Model-Based Human Motion Capture from Monocular Video Sequences

Model-Based Human Motion Capture from Monocular Video Sequences Model-Based Human Motion Capture from Monocular Video Sequences Jihun Park 1, Sangho Park 2, and J.K. Aggarwal 2 1 Department of Computer Engineering Hongik University Seoul, Korea jhpark@hongik.ac.kr

More information

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T

10/11/07 1. Motion Control (wheeled robots) Representing Robot Position ( ) ( ) [ ] T 3 3 Motion Control (wheeled robots) Introduction: Mobile Robot Kinematics Requirements for Motion Control Kinematic / dynamic model of the robot Model of the interaction between the wheel and the ground

More information

Learning a Manifold as an Atlas Supplementary Material

Learning a Manifold as an Atlas Supplementary Material Learning a Manifold as an Atlas Supplementary Material Nikolaos Pitelis Chris Russell School of EECS, Queen Mary, University of London [nikolaos.pitelis,chrisr,lourdes]@eecs.qmul.ac.uk Lourdes Agapito

More information

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical

More information

CS 231A: Computer Vision (Winter 2018) Problem Set 2

CS 231A: Computer Vision (Winter 2018) Problem Set 2 CS 231A: Computer Vision (Winter 2018) Problem Set 2 Due Date: Feb 09 2018, 11:59pm Note: In this PS, using python2 is recommended, as the data files are dumped with python2. Using python3 might cause

More information

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation , pp.162-167 http://dx.doi.org/10.14257/astl.2016.138.33 A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation Liqiang Hu, Chaofeng He Shijiazhuang Tiedao University,

More information

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics

Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Fast, Unconstrained Camera Motion Estimation from Stereo without Tracking and Robust Statistics Heiko Hirschmüller, Peter R. Innocent and Jon M. Garibaldi Centre for Computational Intelligence, De Montfort

More information

On-line, Incremental Learning of a Robust Active Shape Model

On-line, Incremental Learning of a Robust Active Shape Model On-line, Incremental Learning of a Robust Active Shape Model Michael Fussenegger 1, Peter M. Roth 2, Horst Bischof 2, Axel Pinz 1 1 Institute of Electrical Measurement and Measurement Signal Processing

More information

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Mobile Human Detection Systems based on Sliding Windows Approach-A Review Mobile Human Detection Systems based on Sliding Windows Approach-A Review Seminar: Mobile Human detection systems Njieutcheu Tassi cedrique Rovile Department of Computer Engineering University of Heidelberg

More information

Interpolation and extrapolation of motion capture data

Interpolation and extrapolation of motion capture data Interpolation and extrapolation of motion capture data Kiyoshi Hoshino Biological Cybernetics Lab, University of the Ryukyus and PRESTO-SORST, Japan Science and Technology Corporation Nishihara, Okinawa

More information

Occlusion Robust Multi-Camera Face Tracking

Occlusion Robust Multi-Camera Face Tracking Occlusion Robust Multi-Camera Face Tracking Josh Harguess, Changbo Hu, J. K. Aggarwal Computer & Vision Research Center / Department of ECE The University of Texas at Austin harguess@utexas.edu, changbo.hu@gmail.com,

More information

Sequential Non-Rigid Structure-from-Motion with the 3D-Implicit Low-Rank Shape Model

Sequential Non-Rigid Structure-from-Motion with the 3D-Implicit Low-Rank Shape Model Sequential Non-Rigid Structure-from-Motion with the 3D-Implicit Low-Rank Shape Model Marco Paladini 1, Adrien Bartoli 2, and Lourdes Agapito 1 1 Queen Mary University of London, Mile End Road, E1 4NS London,

More information

Measurement of 3D Foot Shape Deformation in Motion

Measurement of 3D Foot Shape Deformation in Motion Measurement of 3D Foot Shape Deformation in Motion Makoto Kimura Masaaki Mochimaru Takeo Kanade Digital Human Research Center National Institute of Advanced Industrial Science and Technology, Japan The

More information

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching Stereo Matching Fundamental matrix Let p be a point in left image, p in right image l l Epipolar relation p maps to epipolar line l p maps to epipolar line l p p Epipolar mapping described by a 3x3 matrix

More information

Epipolar Geometry in Stereo, Motion and Object Recognition

Epipolar Geometry in Stereo, Motion and Object Recognition Epipolar Geometry in Stereo, Motion and Object Recognition A Unified Approach by GangXu Department of Computer Science, Ritsumeikan University, Kusatsu, Japan and Zhengyou Zhang INRIA Sophia-Antipolis,

More information

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group)

Research Subject. Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group) Research Subject Dynamics Computation and Behavior Capture of Human Figures (Nakamura Group) (1) Goal and summary Introduction Humanoid has less actuators than its movable degrees of freedom (DOF) which

More information

Hand-Eye Calibration from Image Derivatives

Hand-Eye Calibration from Image Derivatives Hand-Eye Calibration from Image Derivatives Abstract In this paper it is shown how to perform hand-eye calibration using only the normal flow field and knowledge about the motion of the hand. The proposed

More information

Image Coding with Active Appearance Models

Image Coding with Active Appearance Models Image Coding with Active Appearance Models Simon Baker, Iain Matthews, and Jeff Schneider CMU-RI-TR-03-13 The Robotics Institute Carnegie Mellon University Abstract Image coding is the task of representing

More information

Recognition Rate. 90 S 90 W 90 R Segment Length T

Recognition Rate. 90 S 90 W 90 R Segment Length T Human Action Recognition By Sequence of Movelet Codewords Xiaolin Feng y Pietro Perona yz y California Institute of Technology, 36-93, Pasadena, CA 925, USA z Universit a dipadova, Italy fxlfeng,peronag@vision.caltech.edu

More information

BIL Computer Vision Apr 16, 2014

BIL Computer Vision Apr 16, 2014 BIL 719 - Computer Vision Apr 16, 2014 Binocular Stereo (cont d.), Structure from Motion Aykut Erdem Dept. of Computer Engineering Hacettepe University Slide credit: S. Lazebnik Basic stereo matching algorithm

More information

CS 223B Computer Vision Problem Set 3

CS 223B Computer Vision Problem Set 3 CS 223B Computer Vision Problem Set 3 Due: Feb. 22 nd, 2011 1 Probabilistic Recursion for Tracking In this problem you will derive a method for tracking a point of interest through a sequence of images.

More information

Abstract We present a system which automatically generates a 3D face model from a single frontal image of a face. Our system consists of two component

Abstract We present a system which automatically generates a 3D face model from a single frontal image of a face. Our system consists of two component A Fully Automatic System To Model Faces From a Single Image Zicheng Liu Microsoft Research August 2003 Technical Report MSR-TR-2003-55 Microsoft Research Microsoft Corporation One Microsoft Way Redmond,

More information

Computer Vision in a Non-Rigid World

Computer Vision in a Non-Rigid World 2 nd Tutorial on Computer Vision in a Non-Rigid World Dr. Lourdes Agapito Prof. Adrien Bartoli Dr. Alessio Del Bue Non-rigid Structure from Motion Non-rigid structure from motion To extract non-rigid 3D

More information

Structure from Motion

Structure from Motion Structure from Motion Lecture-13 Moving Light Display 1 Shape from Motion Problem Given optical flow or point correspondences, compute 3-D motion (translation and rotation) and shape (depth). 2 S. Ullman

More information

Non-rigid stereo factorization

Non-rigid stereo factorization [Journal Name], [Volumn Number], 1 15 ([Volumn Year]) c [Volumn Year] Kluwer Academic Publishers, Boston Manufactured in The Netherlands Non-rigid stereo factorization AESSIO DE BUE AND OURDES AGAPITO

More information

Multi-stable Perception. Necker Cube

Multi-stable Perception. Necker Cube Multi-stable Perception Necker Cube Spinning dancer illusion, Nobuyuki Kayahara Multiple view geometry Stereo vision Epipolar geometry Lowe Hartley and Zisserman Depth map extraction Essential matrix

More information

Using Subspace Constraints to Improve Feature Tracking Presented by Bryan Poling. Based on work by Bryan Poling, Gilad Lerman, and Arthur Szlam

Using Subspace Constraints to Improve Feature Tracking Presented by Bryan Poling. Based on work by Bryan Poling, Gilad Lerman, and Arthur Szlam Presented by Based on work by, Gilad Lerman, and Arthur Szlam What is Tracking? Broad Definition Tracking, or Object tracking, is a general term for following some thing through multiple frames of a video

More information

Chapter 2 Basic Structure of High-Dimensional Spaces

Chapter 2 Basic Structure of High-Dimensional Spaces Chapter 2 Basic Structure of High-Dimensional Spaces Data is naturally represented geometrically by associating each record with a point in the space spanned by the attributes. This idea, although simple,

More information

Animation. CS 465 Lecture 22

Animation. CS 465 Lecture 22 Animation CS 465 Lecture 22 Animation Industry production process leading up to animation What animation is How animation works (very generally) Artistic process of animation Further topics in how it works

More information

International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol. XXXIV-5/W10

International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol. XXXIV-5/W10 BUNDLE ADJUSTMENT FOR MARKERLESS BODY TRACKING IN MONOCULAR VIDEO SEQUENCES Ali Shahrokni, Vincent Lepetit, Pascal Fua Computer Vision Lab, Swiss Federal Institute of Technology (EPFL) ali.shahrokni,vincent.lepetit,pascal.fua@epfl.ch

More information

Human Upper Body Pose Estimation in Static Images

Human Upper Body Pose Estimation in Static Images 1. Research Team Human Upper Body Pose Estimation in Static Images Project Leader: Graduate Students: Prof. Isaac Cohen, Computer Science Mun Wai Lee 2. Statement of Project Goals This goal of this project

More information

CS 231A Computer Vision (Fall 2012) Problem Set 3

CS 231A Computer Vision (Fall 2012) Problem Set 3 CS 231A Computer Vision (Fall 2012) Problem Set 3 Due: Nov. 13 th, 2012 (2:15pm) 1 Probabilistic Recursion for Tracking (20 points) In this problem you will derive a method for tracking a point of interest

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Nov 2009 Perspective projection and affine motion Goals Today Perspective projection 3D motion Wed Projects Friday Regularization and robust statistics

More information

Detection and recognition of moving objects using statistical motion detection and Fourier descriptors

Detection and recognition of moving objects using statistical motion detection and Fourier descriptors Detection and recognition of moving objects using statistical motion detection and Fourier descriptors Daniel Toth and Til Aach Institute for Signal Processing, University of Luebeck, Germany toth@isip.uni-luebeck.de

More information

Chapter 9 Object Tracking an Overview

Chapter 9 Object Tracking an Overview Chapter 9 Object Tracking an Overview The output of the background subtraction algorithm, described in the previous chapter, is a classification (segmentation) of pixels into foreground pixels (those belonging

More information

On the Dimensionality of Deformable Face Models

On the Dimensionality of Deformable Face Models On the Dimensionality of Deformable Face Models CMU-RI-TR-06-12 Iain Matthews, Jing Xiao, and Simon Baker The Robotics Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 Abstract

More information

Rigging / Skinning. based on Taku Komura, Jehee Lee and Charles B.Own's slides

Rigging / Skinning. based on Taku Komura, Jehee Lee and Charles B.Own's slides Rigging / Skinning based on Taku Komura, Jehee Lee and Charles B.Own's slides Skeletal Animation Victoria 2 CSE 872 Dr. Charles B. Owen Advanced Computer Graphics Skinning http://www.youtube.com/watch?

More information

2D vs. 3D Deformable Face Models: Representational Power, Construction, and Real-Time Fitting

2D vs. 3D Deformable Face Models: Representational Power, Construction, and Real-Time Fitting 2D vs. 3D Deformable Face Models: Representational Power, Construction, and Real-Time Fitting Iain Matthews, Jing Xiao, and Simon Baker The Robotics Institute, Carnegie Mellon University Epsom PAL, Epsom

More information

Factorization Method Using Interpolated Feature Tracking via Projective Geometry

Factorization Method Using Interpolated Feature Tracking via Projective Geometry Factorization Method Using Interpolated Feature Tracking via Projective Geometry Hideo Saito, Shigeharu Kamijima Department of Information and Computer Science, Keio University Yokohama-City, 223-8522,

More information

Grasp Recognition using a 3D Articulated Model and Infrared Images

Grasp Recognition using a 3D Articulated Model and Infrared Images Grasp Recognition using a 3D Articulated Model and Infrared Images Koichi Ogawara Institute of Industrial Science, Univ. of Tokyo, Tokyo, Japan Jun Takamatsu Institute of Industrial Science, Univ. of Tokyo,

More information

Interactive Computer Graphics

Interactive Computer Graphics Interactive Computer Graphics Lecture 18 Kinematics and Animation Interactive Graphics Lecture 18: Slide 1 Animation of 3D models In the early days physical models were altered frame by frame to create

More information

Chapter 3 Image Registration. Chapter 3 Image Registration

Chapter 3 Image Registration. Chapter 3 Image Registration Chapter 3 Image Registration Distributed Algorithms for Introduction (1) Definition: Image Registration Input: 2 images of the same scene but taken from different perspectives Goal: Identify transformation

More information

Generating Different Realistic Humanoid Motion

Generating Different Realistic Humanoid Motion Generating Different Realistic Humanoid Motion Zhenbo Li,2,3, Yu Deng,2,3, and Hua Li,2,3 Key Lab. of Computer System and Architecture, Institute of Computing Technology, Chinese Academy of Sciences, Beijing

More information

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE

COMPUTER VISION > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE COMPUTER VISION 2017-2018 > OPTICAL FLOW UTRECHT UNIVERSITY RONALD POPPE OUTLINE Optical flow Lucas-Kanade Horn-Schunck Applications of optical flow Optical flow tracking Histograms of oriented flow Assignment

More information

ELEC Dr Reji Mathew Electrical Engineering UNSW

ELEC Dr Reji Mathew Electrical Engineering UNSW ELEC 4622 Dr Reji Mathew Electrical Engineering UNSW Review of Motion Modelling and Estimation Introduction to Motion Modelling & Estimation Forward Motion Backward Motion Block Motion Estimation Motion

More information

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation

Motion. 1 Introduction. 2 Optical Flow. Sohaib A Khan. 2.1 Brightness Constancy Equation Motion Sohaib A Khan 1 Introduction So far, we have dealing with single images of a static scene taken by a fixed camera. Here we will deal with sequence of images taken at different time intervals. Motion

More information

MOTION SEGMENTATION BASED ON INDEPENDENT SUBSPACE ANALYSIS

MOTION SEGMENTATION BASED ON INDEPENDENT SUBSPACE ANALYSIS MOTION SEGMENTATION BASED ON INDEPENDENT SUBSPACE ANALYSIS Zhimin Fan, Jie Zhou and Ying Wu Department of Automation, Tsinghua University, Beijing 100084, China Department of Electrical and Computer Engineering,

More information