PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO

Similar documents
Performance Capture from Sparse Multi-view Video

Multi-view stereo. Many slides adapted from S. Seitz

Multi-View Stereo for Static and Dynamic Scenes

HISTOGRAMS OF ORIENTATIO N GRADIENTS

Structured light 3D reconstruction

CS201: Computer Vision Introduction to Tracking

Processing 3D Surface Data

Dense 3D Reconstruction. Christiano Gava

Stereo Vision. MAN-522 Computer Vision

Step-by-Step Model Buidling

Dense 3D Reconstruction. Christiano Gava

Midterm Examination CS 534: Computational Photography

3D Modeling of Objects Using Laser Scanning

Processing 3D Surface Data

Measurement of 3D Foot Shape Deformation in Motion

Scanning Real World Objects without Worries 3D Reconstruction

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Dynamic Geometry Processing

Input Nodes. Surface Input. Surface Input Nodal Motion Nodal Displacement Instance Generator Light Flocking

Dynamic 3D Shape From Multi-viewpoint Images Using Deformable Mesh Model

Planetary Rover Absolute Localization by Combining Visual Odometry with Orbital Image Measurements

Geometric Modeling and Processing

Model-based Motion Capture for Crash Test Video Analysis

Heterogeneous Deformation Model for 3D Shape and Motion Recovery from Multi-Viewpoint Images

Recap: Features and filters. Recap: Grouping & fitting. Now: Multiple views 10/29/2008. Epipolar geometry & stereo vision. Why multiple views?

Multiple View Geometry

3D Models from Range Sensors. Gianpaolo Palma

Multiview Stereo COSC450. Lecture 8

3D Computer Vision. Depth Cameras. Prof. Didier Stricker. Oliver Wasenmüller

arxiv: v1 [cs.cv] 28 Sep 2018

Computational Design. Stelian Coros

Correspondence and Stereopsis. Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]

Multi-view Stereo. Ivo Boyadzhiev CS7670: September 13, 2011

Geometric Representations. Stelian Coros

Volume Illumination. Visualisation Lecture 11. Taku Komura. Institute for Perception, Action & Behaviour School of Informatics

3D Imaging from Video and Planar Radiography

Processing 3D Surface Data

3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: ,

Corner Detection. GV12/3072 Image Processing.

Computer Vision Lecture 17

Computer Vision. 3D acquisition

Computer Vision Lecture 17

SIFT: SCALE INVARIANT FEATURE TRANSFORM SURF: SPEEDED UP ROBUST FEATURES BASHAR ALSADIK EOS DEPT. TOPMAP M13 3D GEOINFORMATION FROM IMAGES 2014

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

Overview. Augmented reality and applications Marker-based augmented reality. Camera model. Binary markers Textured planar markers

Occlusion Detection of Real Objects using Contour Based Stereo Matching

Augmented Reality VU. Computer Vision 3D Registration (2) Prof. Vincent Lepetit

Shape Matching. Michael Kazhdan ( /657)

TEXTURE OVERLAY ONTO NON-RIGID SURFACE USING COMMODITY DEPTH CAMERA

Prof. Fanny Ficuciello Robotics for Bioengineering Visual Servoing

Textureless Layers CMU-RI-TR Qifa Ke, Simon Baker, and Takeo Kanade

Surface Reconstruction. Gianpaolo Palma

CS231A Midterm Review. Friday 5/6/2016

Surface Registration. Gianpaolo Palma

Video Google: A Text Retrieval Approach to Object Matching in Videos

Image Based Reconstruction II

Vision-based Mobile Robot Localization and Mapping using Scale-Invariant Features

Last update: May 4, Vision. CMSC 421: Chapter 24. CMSC 421: Chapter 24 1

BSB663 Image Processing Pinar Duygulu. Slides are adapted from Selim Aksoy

Stereo and Epipolar geometry

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Index C, D, E, F I, J

Reconstruction of complete 3D object model from multi-view range images.

Local Feature Detectors

Talk plan. 3d model. Applications: cultural heritage 5/9/ d shape reconstruction from photographs: a Multi-View Stereo approach

Geometric Reconstruction Dense reconstruction of scene geometry

CSE452 Computer Graphics

L2 Data Acquisition. Mechanical measurement (CMM) Structured light Range images Shape from shading Other methods

ELEC Dr Reji Mathew Electrical Engineering UNSW

Multi-View 3D-Reconstruction

Large Mesh Deformation Using the Volumetric Graph Laplacian

Automatic Reconstruction of 3D Objects Using a Mobile Monoscopic Camera

Camera Drones Lecture 3 3D data generation

An Evaluation of Volumetric Interest Points

CS 231A Computer Vision (Winter 2018) Problem Set 3

Automatic Generation of Animatable 3D Personalized Model Based on Multi-view Images

Motion Analysis. Motion analysis. Now we will talk about. Differential Motion Analysis. Motion analysis. Difference Pictures

EE795: Computer Vision and Intelligent Systems

Volumetric stereo with silhouette and feature constraints

Shape from Silhouettes I

Stereo. 11/02/2012 CS129, Brown James Hays. Slides by Kristen Grauman

Depth. Common Classification Tasks. Example: AlexNet. Another Example: Inception. Another Example: Inception. Depth

HUMAN motion capture has been under investigation for many

Computer Graphics. Shadows

A Statistical Consistency Check for the Space Carving Algorithm.

Structured Light II. Thanks to Ronen Gvili, Szymon Rusinkiewicz and Maks Ovsjanikov

Real-Time Shape Editing using Radial Basis Functions

Speech Driven Synthesis of Talking Head Sequences

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

CS4495/6495 Introduction to Computer Vision

A Simple 3D Scanning System of the Human Foot Using a Smartphone with Depth Camera

TEXTURE OVERLAY ONTO NON-RIGID SURFACE USING COMMODITY DEPTH CAMERA

Combining Appearance and Topology for Wide

Project Updates Short lecture Volumetric Modeling +2 papers

Easy modeling generate new shapes by deforming existing ones

SLAM with SIFT (aka Mobile Robot Localization and Mapping with Uncertainty using Scale-Invariant Visual Landmarks ) Se, Lowe, and Little

Miniature faking. In close-up photo, the depth of field is limited.

The SIFT (Scale Invariant Feature

Deformable Mesh Model for Complex Multi-Object 3D Motion Estimation from Multi-Viewpoint Video

3D model-based human modeling and tracking

Transcription:

Stefan Krauß, Juliane Hüttl SE, SoSe 2011, HU-Berlin PERFORMANCE CAPTURE FROM SPARSE MULTI-VIEW VIDEO 1

Uses of Motion/Performance Capture movies games, virtual environments biomechanics, sports science, medicine robot control human-computer interaction 2

Types of Motion/Performance Capture inertial optical marker-based electromagnetic electromechanical optical markerless 3

Processing Steps 4

Data Retrieval full body laser scan 8 cameras, 24fps, 1004x1004 pixels, placed in circular configuration 5

Processing Steps 6

Pre-Processing color-based background subtraction for silhouette of model generate triangle mesh via Poisson surface reconstruction create tetrahedral version of surface scan Delaunay tetrahedralization 7

Deformation Problem on Triangle Meshes 8

Processing Steps 9

Capturing the Global Model Pose recover global pose of tetrahedral mesh for each time step of video subdivide surface of model into 100-200 regions compute deformation constraints from each pair of subsequent time steps 1. Pose Initialization from Image Features 2. Refining the Pose using Silhouette Rims 3. Optimizing Key Handle positions 10

Pose Initialization from Image Features SIFT scale-invariant feature transform features largely invariant under illumination out-of-plane rotation image noise small changes in viewpoint scaling reliable correspondence finding for fast scene motion 11

Pose Initialization from Image Features associate vertices of tetrahedral mesh with descriptors at time t compare intersection point of reprojected rays with vertex find corresponding association at time t + 1 12

Pose Initialization from Image Features each vertex found is candidate for deformation handle find one best handle for each region: normal n of vertex v most collinear with difference vector of intersection point p and vertex v define intermediate target position step-wise deformation until silhouette overlap error is minimal 13

Processing Steps 14

Volumetric Deformation 1. Linear Laplacian deformation 2. Rotation extraction 3. Update Iterate until non-rigid deformation is minimized 15

Linear Laplacian Deformation calculate gradient operator matrix G for each vertex of every tetrahedron construct Laplacian operator from G and volumina D (and constraints) solve Laplacian System: Lv = δ with L = G T DG and δ = G T Dg calculate new vertex-positions 16

Rotation Extraction and Update extract transformation matrix for each tetrahedron split transformation matrix in rigid and non-rigid part iterate until non-rigid-part is minimum update differential 17

Capturing the Global Model Pose recover global pose of tetrahedral mesh for each time step of video subdivide surface of model into 100-200 regions compute deformation constraints from each pair of subsequent time steps 1. Pose Initialization from Image Features 2. Refining the Pose using Silhouette Rims 3. Optimizing Key Handle positions 18

Refining the Pose using Silhouette Rims find rim vertices compute distance field value from distance field at projected location defines displacement length 19

Optimizing Key Handle positions find optimal multi-view silhouette overlap choose 15-25 key vertices manually optimize vertex positions until tetrahedral deformation produces optimal silhouette overlap (based on overlap error) model and silhouette overlap before optimization after key vertex optimization 20

Processing Steps 21

Deformation Transfer match vertices of input triangle mesh to input tetrahedral mesh vertices of triangle mesh are linear combination of tetrahedral mesh combine linear coefficients in matrix B T tri = T tet B 22

Processing Steps 23

Capturing Surface Detail map tetrahedral mesh to triangle mesh capture shape detail at each time step 1. Adaptation along Silhouette Contours 2. Model-Guided Multi-View Stereo 24

Adaptation along Silhouette Contours calculate rim vertices find closest 2D point on silhouette boundary for each rim vertex compute image gradients at input silhouette points and reprojected model contour image if distance between gradients smaller threshold, add as constraint 25

Model-Guided Multi-View Stereo multi-view stereo method by [Goesele et al. 2006] project pixel p of view R into bounding box reproject point to neighbouring views N 0, N 1 compute cross-correlation for window around p and corresponding window around q 0, q 1 high score for valid depth d 26

Model-Guided Multi-View Stereo constrain correlation search to 3D points at most ±2cm away from triangle mesh merge depth maps into single point cloud project points from triangle mesh onto point cloud projected points provide additional position constraints 27

Processing Steps 28

Surface-based Deformation calculate differential coordinates δ L is cotangent Laplacian matrix C is diagonal matrix, only entries for constrained vertices q minimize: Lv δ 2 + Cv q 2 new triangle mesh in v 29

Processing Steps 30

Summary 31

Summary overlay between input image and model silhouette overlap [%] blue: global pose estimation green: surface detail reconstruction 32

Summary non-intrusive method all test input scenes were reconstructed fully automatically minimal user input: mark head and foot regions to exclude from surface detail capture mark 25 deformation handles for key handle optimization only few isolated pose errors due to motion blur or strong shadows 33

Summary limitations: topological structure of input silhouette must be close to reprojected model silhouette (no outdoor scenes) topology of input scanned model preserved (movement of hair cannot be captured) deformation technique mimics elastic deformation rubbery look resolution limit to deformation capture (high-frequency details baked in ) reconstruction using a detailed (m) and a coarse (r) model 34

Summary large application range of algorithm tight and loose apparel fabrics with prominent texture as well as plain colors simple walks and different dance styles complicated self-occlusions 35

References De Aguiar et al, 2008 Performance Capture from Sparse Multi-view Video Stoll et al, 2007, A Volumetric Approach to Interactive Shape Editing Goesele et al, 2006, Multi-view stereo revisited Taku Komura, http://homepages.inf.ed.ac.uk/tkomura/cav/index.html, access: 27.05.2011 36