Fast Guided Global Interpolation for Depth and. Yu Li, Dongbo Min, Minh N. Do, Jiangbo Lu

Similar documents
SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs. Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu

Optical flow. Cordelia Schmid

Flow Estimation. Min Bai. February 8, University of Toronto. Min Bai (UofT) Flow Estimation February 8, / 47

Efficient Coarse-to-Fine PatchMatch for Large Displacement Optical Flow

Supplementary Material for "FlowNet 2.0: Evolution of Optical Flow Estimation with Deep Networks"

Introduction to Computer Graphics. Image Processing (1) June 8, 2017 Kenshi Takayama

A Hardware-Friendly Bilateral Solver for Real-Time Virtual-Reality Video

Recovering Realistic Texture in Image Super-resolution by Deep Spatial Feature Transform. Xintao Wang Ke Yu Chao Dong Chen Change Loy

Deep learning for dense per-pixel prediction. Chunhua Shen The University of Adelaide, Australia

CNN for Low Level Image Processing. Huanjing Yue

Peripheral drift illusion

Ninio, J. and Stevens, K. A. (2000) Variations on the Hermann grid: an extinction illusion. Perception, 29,

5638 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 12, DECEMBER 2014

Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting

Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging

JOINT DETECTION AND SEGMENTATION WITH DEEP HIERARCHICAL NETWORKS. Zhao Chen Machine Learning Intern, NVIDIA

MOTION ESTIMATION USING CONVOLUTIONAL NEURAL NETWORKS. Mustafa Ozan Tezcan

LEARNING RIGIDITY IN DYNAMIC SCENES FOR SCENE FLOW ESTIMATION

Joint Inference in Image Databases via Dense Correspondence. Michael Rubinstein MIT CSAIL (while interning at Microsoft Research)

Light Field Occlusion Removal

EE795: Computer Vision and Intelligent Systems

Online Object Tracking with Proposal Selection. Yang Hua Karteek Alahari Cordelia Schmid Inria Grenoble Rhône-Alpes, France

Models Matter, So Does Training: An Empirical Study of CNNs for Optical Flow Estimation

arxiv: v2 [cs.cv] 19 May 2015

Martian lava field, NASA, Wikipedia

OPTICAL flow, a formulation of the dense correspondence. for Dense Correspondence. Yunsong Li, Yinlin Hu, Rui Song, Peng Rao and Yangli Wang

Fast Edge-Preserving PatchMatch for Large Displacement Optical Flow

Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation

ACCURATE estimation of optical flow is a useful but challenging. Pyramidal Gradient Matching for Optical Flow Estimation

Multi-stable Perception. Necker Cube

Interactive Non-Linear Image Operations on Gigapixel Images

Multi-View 3D Object Detection Network for Autonomous Driving

Robust Local Optical Flow: Dense Motion Vector Field Interpolation

Supplemental Material Deep Fluids: A Generative Network for Parameterized Fluid Simulations

Large Displacement Optical Flow & Applications

Motion Estimation (II) Ce Liu Microsoft Research New England

Learning 6D Object Pose Estimation and Tracking

Flow-Based Video Recognition

BOP: Benchmark for 6D Object Pose Estimation

Deep Back-Projection Networks For Super-Resolution Supplementary Material

Fully Convolutional Networks for Semantic Segmentation

Lecture 7: Semantic Segmentation

Feature Tracking and Optical Flow

Seeing the unseen. Data-driven 3D Understanding from Single Images. Hao Su

Single Image Super-resolution. Slides from Libin Geoffrey Sun and James Hays

Structured Face Hallucination

Image Composition. COS 526 Princeton University

arxiv: v1 [cs.cv] 11 Apr 2016

LEARNING TO SEGMENT MOVING OBJECTS IN VIDEOS FRAGKIADAKI ET AL. 2015

A MULTI-RESOLUTION APPROACH TO DEPTH FIELD ESTIMATION IN DENSE IMAGE ARRAYS F. Battisti, M. Brizzi, M. Carli, A. Neri

Dense 3D Modelling and Monocular Reconstruction of Deformable Objects

Shape Preserving RGB-D Depth Map Restoration

Multiframe Scene Flow with Piecewise Rigid Motion. Vladislav Golyanik,, Kihwan Kim, Robert Maier, Mathias Nießner, Didier Stricker and Jan Kautz

ActiveStereoNet: End-to-End Self-Supervised Learning for Active Stereo Systems (Supplementary Materials)

Combined Bilateral Filter for Enhanced Real-Time Upsampling of Depth Images

Direct Methods in Visual Odometry

SSD: Single Shot MultiBox Detector. Author: Wei Liu et al. Presenter: Siyu Jiang

High Performance GPU-Based Preprocessing for Time-of-Flight Imaging in Medical Applications

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Deep Incremental Scene Understanding. Federico Tombari & Christian Rupprecht Technical University of Munich, Germany

Fast and Accurate Optical Flow based Depth Map Estimation from Light Fields

Optical flow. Cordelia Schmid

Feature Tracking and Optical Flow

Object Localization, Segmentation, Classification, and Pose Estimation in 3D Images using Deep Learning

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

FAST: A Framework to Accelerate Super- Resolution Processing on Compressed Videos

Switchable Temporal Propagation Network

ECCV Presented by: Boris Ivanovic and Yolanda Wang CS 331B - November 16, 2016

Learning to Detect Motion Boundaries

Structured Prediction using Convolutional Neural Networks

Beyond bags of features: Adding spatial information. Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba

CS231N Section. Video Understanding 6/1/2018

Temporally Consistence Depth Estimation from Stereo Video Sequences

arxiv: v1 [cs.cv] 22 Jan 2016

PatchMatch: A Randomized Correspondence Algorithm for Structural Image Editing

Real-Time Virtual Viewpoint Generation on the GPU for Scene Navigation

P-CNN: Pose-based CNN Features for Action Recognition. Iman Rezazadeh

Part II: Modeling Aspects

Convolutional Neural Network Implementation of Superresolution Video

Mask R-CNN. presented by Jiageng Zhang, Jingyao Zhan, Yunhan Ma

SparseFlow: Sparse Matching for Small to Large Displacement Optical Flow

Overview. Spectral Processing of Point- Sampled Geometry. Introduction. Introduction. Fourier Transform. Fourier Transform

Towards Robust Visual Object Tracking: Proposal Selection & Occlusion Reasoning Yang HUA

FLaME: Fast Lightweight Mesh Estimation using Variational Smoothing on Delaunay Graphs

Iterative methods for use with the Fast Multipole Method

Semi-Supervised Learning for Optical Flow with Generative Adversarial Networks

A Gentle Introduction to Bilateral Filtering and its Applications 10/10: Conclusions

Fundamentals of Stereo Vision Michael Bleyer LVA Stereo Vision

UnFlow: Unsupervised Learning of Optical Flow with a Bidirectional Census Loss

Urban Scene Segmentation, Recognition and Remodeling. Part III. Jinglu Wang 11/24/2016 ACCV 2016 TUTORIAL

Light Field Super Resolution with Convolutional Neural Networks

DeepIM: Deep Iterative Matching for 6D Pose Estimation - Supplementary Material

One Network to Solve Them All Solving Linear Inverse Problems using Deep Projection Models

Depth Learning: When Depth Estimation Meets Deep Learning

Face Recognition A Deep Learning Approach

Problem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation

Ping Tan. Simon Fraser University

Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture David Eigen, Rob Fergus

Project Updates Short lecture Volumetric Modeling +2 papers

Bidirectional Recurrent Convolutional Networks for Video Super-Resolution

Transcription:

Fast Guided Global Interpolation for Depth and Yu Li, Dongbo Min, Minh N. Do, Jiangbo Lu

Introduction Depth upsampling and motion interpolation are often required to generate a dense, high-quality, and high resolution depth map or optical flow field. Low-res & noisy depth (ToF) 68 86 High-res. color guidance 1088 1376 High-res. depth 1088 1376 Depth upsampling (color guided) Input TOF depth: noisy, low resolution, regularly distributed

Introduction Depth upsampling and motion interpolation are often required to generate a dense, high-quality, and high resolution depth map or optical flow field. Color frame and sparse matches from DM Dense optical flow field Data density < 1% Motion interpolation Input matches: typically reliable, but highly scattered, varying density [DM: Weinzaepfel et al., "DeepFlow: Large displacement optical flow with deep matching, ICCV 2013.]

Motivation Existing methods are often tailored to one specific task: Depth upsampling Motion interpolation JBF [Kopf et al. 2007], MRF+nlm [Park et al. 2011], TGV [Ferstl et al. 2013], JGU [Liu et al. 2013], AR [Yang et al. 2014], Data-driven [Kwon et al. 2015], etc EpicFlow [Revaud et al. 2015], [Drayer and Brox 2015], [Leordeanu et al. 2013], etc The common objective for both tasks is to densify a set of sparse data points, either regularly distributed or scattered, to a full image grid through a 2D guided interpolation process. Our approach: Fast Guided Global Interpolation (FGI) A unified approach that casts the guided interpolation problem into a hierarchical, global optimization framework.

Several Challenges e.g. Texture-copy artifacts due to inconsistent structures Color guidance Simple joint filtering Ground truth Large occlusions, long-range propagation and extrapolation EpicFlow Loss of thin structures, missing motion boundaries Complex algorithms and computationally inefficient

Single-scale WLS Method Vs. Our Method [WLS: Farbman et al., Edge-preserving decompositions for multi-scale tone and detail manipulation, SIGGRAPH 2008] [FGS: Min et al., "Fast global image smoothing based on weighted least squares," TIP 2014.]

Our Pipeline: Overview A hierarchical (coarse-to-fine), multi-pass guided interpolation framework Divide the problems into a sequence of interpolation tasks each with smaller scale factors Gradually fill the large gap between the sparse measurement and the dense data

Our Pipeline: Filtering with Alternating Guidances From the coarse level l= L-1, we upsample the signal by a factor of 2 at each level by solving the following weighted least square (WLS) using the recent FGS solver. Guided interp. : The color image c l as the guidance A cl is the spatially varying Laplacian matrix defined by c l [FGS: Min et al., "Fast global image smoothing based on weighted least squares," TIP 2014.] Why FGS? 100 ms for filteirng 1MPixels RGB images on 1 CPU core. More details in our paper.

Our Pipeline: Filtering with Alternating Guidances Next, another WLS is solved with the output d * as guidance and bicubic interpoplated signal as input. Joint filtering: The intermediate interpolated map d as the guidance A d is the spatially varying Laplacian matrix defined by d

Our Pipeline: Consensus-Based Data Augmentation Then, check the consistency between the output and the bicubic upsampled data, and pick the most consistent points to add to the data mask map m l The bibubic upsampled data is free from texture-copying Proceed in a non-overlapping patch fashion (2x2 patches) The entire process is repeated until the finest level (l = 0) is reached.

1D Scanline Illustration Guidance Ground truth One-pass WLS Ground truth

1D Scanline Illustration l=2 l=2 l=1 l=1 Guidance ~ d * d d 2 * d1 ~ l=0 l=0 Ground truth ~ d d = our result One-pass WLS Ground truth * 0

Depth Upsamling Error (MAD) Pipeline Validation on Depth Upsampling 4 Single scale WLS +Cascaded filtering with alternating guidances single scale (Sec. 3.1) +Hierarchical process +Consensus based data point augmentation (Sec. 3.2) 3.5 3 2.5 2 1.5 1 0.5 0 2x 4x 8x 16x

Depth Upsampling Results Average runtime to upsample a 272 344 depth to 1088 1376 (in seconds) MRF+nlm TGV AR GF CLMF FGI(ours) 170 420 900 1.3 2.4 0.6 1000x faster than AR Average Depth Upsampling Error on ToF Synthetic Dataset (6 cases) 5.00 4.50 4.00 3.50 3.00 2.50 2.00 1.50 1.00 0.50 0.00 2X 4X 8X 16X Close to AR MRF+nlm TGV WLS CLMF JGF AR FGI (ours)

Our framework also improves other edge-aware smoothing filters, e.g. the guided filter Depth avg. error Single-pass GF GF in our framework 2x 4x 8x 16x 1.31 1.54 2.04 3.12 1.06 1.21 1.63 2.59 [GF: He et al., Guided image filtering," ECCV 2010.]

Error map Depth Error map Depth Depth Upsampling Results ToF Synthetic Dataset 1000x faster than AR ToFMark Dataset 650x faster than TGV

Motion Interpolation Results Performance (EPE) on MPI Sintel training set WLS EpicFlow-NW EpicFlow-LA FGI(ours) Clean 3.23 3.17 2.65 2.75 Final 4.68 4.55 4.10 4.14 Runtime (sec) 0.21 0.80 0.94 0.39 Close to EpicFlow, but over 2x faster Performance (EPE) on the MPI Sintel testing benchmark

Conclusion General & versatile technique: Tackle both depth and motion interpolation tasks, and potentially more Generally applicable to other edge-aware smoothing filters, e.g. GF Competitive results while running much faster than taskspecific state-of-the-art methods Simple & effective: No color edge detection & variational minimization in [Revaud et al., CVPR 15] No domain transform filtering for post-smoothing in [Barron & Poole, ECCV 16] Further acceleration on GPUs and FPGA, offering a common engine for guided interpolation Project page (code is available): http://publish.illinois.edu/visual-modeling-and-analytics/fastguided-global-interpolation/