Tutorial: Compressive Sensing of Video

Similar documents
Compressive Sensing of High-Dimensional Visual Signals. Aswin C Sankaranarayanan Rice University

Short-course Compressive Sensing of Videos

Computational Cameras: Exploiting Spatial- Angular Temporal Tradeoffs in Photography

Flexible Voxels for Motion-Aware Videography

Coding and Modulation in Cameras

Compressive Sensing: Opportunities and Perils for Computer Vision

Compressive. Graphical Models. Volkan Cevher. Rice University ELEC 633 / STAT 631 Class

CS-MUVI: Video Compressive Sensing for Spatial-Multiplexing Cameras

Compressive Sensing. A New Framework for Sparse Signal Acquisition and Processing. Richard Baraniuk. Rice University

Compressive Sensing: Opportunities and pitfalls for Computer Vision

P2C2: Programmable Pixel Compressive Camera for High Speed Imaging

Mohit Gupta. Assistant Professor, Computer Science Department University of Wisconsin-Madison. Ph.D. in Robotics, Carnegie Mellon University

Compressive Sensing Applications and Demonstrations: Synthetic Aperture Radar

Introduction to Topics in Machine Learning

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

The Fundamentals of Compressive Sensing

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Compressive Parameter Estimation with Earth Mover s Distance via K-means Clustering. Dian Mo and Marco F. Duarte

Sparse Reconstruction / Compressive Sensing

Non-Differentiable Image Manifolds

Computational Photography Denoising

Lecture 17 Sparse Convex Optimization

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Video Compressive Sensing for Spatial Multiplexing Cameras Using Motion-Flow Models

Go With The Flow. Optical Flow-based Transport for Image Manifolds. Chinmay Hegde Rice University

Compressed Sensing for Rapid MR Imaging

Image reconstruction based on back propagation learning in Compressed Sensing theory

Multi-stable Perception. Necker Cube

ELEG Compressive Sensing and Sparse Signal Representations

Modified Iterative Method for Recovery of Sparse Multiple Measurement Problems

Super-Resolution. Many slides from Miki Elad Technion Yosi Rubner RTC and more

Structurally Random Matrices

Learning Splines for Sparse Tomographic Reconstruction. Elham Sakhaee and Alireza Entezari University of Florida

Ping Tan. Simon Fraser University

Optical flow and tracking

Introduction. Wavelets, Curvelets [4], Surfacelets [5].

Parameter Estimation in Compressive Sensing. Marco F. Duarte

CS-Video: Algorithms, Architectures, and Applications for Compressive Video Sensing

Compressive Sensing: Theory and Practice

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

The Benefit of Tree Sparsity in Accelerated MRI

Computer Vision 2. SS 18 Dr. Benjamin Guthier Professur für Bildverarbeitung. Computer Vision 2 Dr. Benjamin Guthier

COMPRESSIVE VIDEO SAMPLING

Motion Tracking and Event Understanding in Video Sequences

Lecture 19: Depth Cameras. Visual Computing Systems CMU , Fall 2013

Feature Tracking and Optical Flow

Computer Vision Lecture 20

Computer Vision Lecture 20

MAS.963 Special Topics: Computational Camera and Photography

Motion and Tracking. Andrea Torsello DAIS Università Ca Foscari via Torino 155, Mestre (VE)

Comparison between Motion Analysis and Stereo

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

Capturing, Modeling, Rendering 3D Structures

Feature Tracking and Optical Flow

Super-Resolution from Image Sequences A Review

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Iterative CT Reconstruction Using Curvelet-Based Regularization

EECS 556 Image Processing W 09

Sparse wavelet expansions for seismic tomography: Methods and algorithms

Compressive Acquisition of Dynamic Scenes Æ

EECS150 - Digital Design Lecture 14 FIFO 2 and SIFT. Recap and Outline

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

The SIFT (Scale Invariant Feature

G Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing

Motion and Optical Flow. Slides from Ce Liu, Steve Seitz, Larry Zitnick, Ali Farhadi

Introduction to Image Super-resolution. Presenter: Kevin Su

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

Peripheral drift illusion

EE 264: Image Processing and Reconstruction. Image Motion Estimation I. EE 264: Image Processing and Reconstruction. Outline

Structure from Motion

BIL Computer Vision Apr 16, 2014

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Stereo Vision. MAN-522 Computer Vision

Finally: Motion and tracking. Motion 4/20/2011. CS 376 Lecture 24 Motion 1. Video. Uses of motion. Motion parallax. Motion field

Bilevel Sparse Coding

CS 4495 Computer Vision A. Bobick. Motion and Optic Flow. Stereo Matching

Compressive Sensing for High-Dimensional Data

Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging

Randomized sampling strategies

Stereo and structured light

Compensating for Motion During Direct-Global Separation

Compressive Sensing Based Image Reconstruction using Wavelet Transform

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation. Range Imaging Through Triangulation

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

Compressive Sensing. Part IV: Beyond Sparsity. Mark A. Davenport. Stanford University Department of Statistics

Applications: Sampling/Reconstruction. Sampling and Reconstruction of Visual Appearance. Motivation. Monte Carlo Path Tracing.

Spatially-Localized Compressed Sensing and Routing in Multi-Hop Sensor Networks 1

An Intuitive Explanation of Fourier Theory

Computer Vision Lecture 20

Sparse Models in Image Understanding And Computer Vision

Weighted-CS for reconstruction of highly under-sampled dynamic MRI sequences

Inverse Problems and Machine Learning

Fundamental matrix. Let p be a point in left image, p in right image. Epipolar relation. Epipolar mapping described by a 3x3 matrix F

Histograms of Sparse Codes for Object Detection

HISTOGRAMS OF ORIENTATIO N GRADIENTS

ECE Digital Image Processing and Introduction to Computer Vision

Transcription:

Tutorial: Compressive Sensing of Video Venue: CVPR 2012, Providence, Rhode Island, USA June 16, 2012 Richard Baraniuk Mohit Gupta Aswin Sankaranarayanan Ashok Veeraraghavan

Tutorial: Compressive Sensing of Video Unit 3 Compressive Video Sensing Systems Ashok Veeraraghavan

RECAP: Compressive Sensing random measurements sparse signal nonzero entries Signal recovery via L 1 optimization [Candes, Romberg, Tao; Donoho]

RECAP: Compressive Sensing random measurements sparse signal Signal recovery via iterative greedy algorithm (orthogonal) matching pursuit [Gilbert, Tropp] iterated thresholding [Nowak, Figueiredo; Kingsbury, Reeves; Daubechies, Defrise, De Mol; Blumensath, Davies; ] CoSaMP [Needell and Tropp] nonzero entries

The Plan for the next hour Model Based Compressive Video Sensing Coded Strobing for Periodic signals CS-Linear Dynamical Systems Linear Motion Model Compressive Sensing of Generic Videos Single Pixel Video Camera Voxel Subsampling Voxel Multiplexing

Model Based CS Video Periodic or Repetitive Videos CS Strobing Linear Dynamical Systems CS LDS Linear Motion Model Multi-Exposure Video

Coded Strobing Camera Ashok Veeraraghavan Dikpal Reddy Ramesh Raskar PAMI 2010

8 Periodic Visual Signals

Strobing: Applications -Strobo-laryngoscopy -Observing machine parts, motors etc. 9

Periodic Signals Periodic Signal (x(t)) with period P P t Periodic signal with period P and band-limited to f Max. The Fourier domain contains only terms that are multiples of f P =1/P. - f Max -4f 0 f P =1/P 2f P -3f P -2f P -f P P 3f P 4f P f Max 11

Normal Camera and Aliasing Integrates out the high frequency information and this information is irrevocably lost. P = 10ms T Frame = Frame Duration = 40ms Causes aliasing of high frequency information. - f Max 0 f f S /2 = Sampling Frequency/2 Max 12

High Speed Camera Nyquist Sampling of x(t) When each period of x has very fine high frequency variations Nyquist sampling rate is very high. P = 10ms T s = 1/(2 f Max ) High speed video camera is inefficient since its bandwidth is wasted on the periodic signal since it is sparse in the Fourier domain - f Max -4f 0 f P =1/P 2f P -3f P -2f P -f P P 3f P 4f P f Max 14

Traditional Strobing When sampling rate of camera is low, generate beat frequencies of the high speed periodic signal which now can be easily captured P = 10ms t The light throughput is also very low in order to avoid blurring of the features during the strobe time. The period of the signal needs to be known a priori 15

16 Coded Strobing

Coded strobing: Time Domain Proposed Design: In every exposure duration observe different linear combinations of the underlying periodic signal. P = 10ms t Advantage of the design: Exposure coding is independent of the frequency of the underlying periodic signal. Further, light throughput is on an average 50% which is far greater than can be achieved via traditional strobing. 17

Coded strobing: Frequency domain Measure linear combinations of the Fourier components. Decode by enforcing sparsity of the reconstructed signal in Fourier domain. - f Max -4f P -3f P -f P 0 f P =1/P 2f P 3f P 4f P f Max Measure Linear Combinations Sparsity Enforcing Reconstruction Algorithm 18

Signal model t Non-zero elements b1 b2 bn Signal Fourier Basis Basis Coeff x = B s 19

Observation Model N unknowns Frame 1 Coded Strobing Frame M Observed Intensity y = Frame Integration Period T S C Signal x t 20

Signal and Observation model Signal Model: x = B s Observation Model: y = C x Observed low-frame rate video y Mixing matrix ( satisfies Restricted Isometry under certain conditions ) = C B s = A s Fourier coefficients of the signal ( sparse if periodic) A of size M x N, M<<N 21

Linear Inversion? y = A s highly underdetermined. M << N For example: Capture at 10 fps for 30s with 1 ms resolution has M=300, N=30000 Highly unstable in noise To reconstruct can we use extra information about the signal? 22

Extra info: Periodic signal y = A s Observed low rate frames Mixing matrix Very few (K) non-zero elements Sparse Basis Coeff 24 - f Max -4f 0 f P =1/P 2f P -3f P -2f P -f P P 3f P f Max

Compressed Sensing Estimate sparse s that satisfies y = A s Solve min s s. t. y As 1 2 (P1) P1 is a linear program. Fast and robust algorithms available for solving P1 26

Signal model and recovery Signal Model x = B s Observation Model y = C x Compressed sensing Solve min s s. t. y As 1 2 Sparsity and location enforced recovery 27

Periodic Signals Periodic Signal (x(t)) with period P P t Periodic signal with period P and band-limited to f Max. The Fourier domain contains only terms that are multiples of f P =1/P. - f Max -4f 0 f P =1/P 2f P -3f P -2f P -f P P 3f P 4f P f Max 28

Structure of Sparsity y = A s Observed low rate frames Mixing matrix Sparse Basis Coeff Fundamental frequency and its harmonics

Sparsity and location In addition to sparsity, use the location of the non-zero coefficients as constraint All the non-zero coefficients equally spaced at f P Solve min s s. t. y As 1 2 (P1t) with truncated mixing matrix and basis coefficients 30

Signal model and recovery Signal Model x = B s Observation Model y = C x Compressed sensing Sparsity and location enforced recovery Solve min s s. t. y As Solve 1 2 min s s. t. y As 1 2 31

Implementation Captured at 10fps using PGR Dragonfly2 Can be strobed at 1ms resolution Can flutter at 250us FLC Shutter 32

Example: Fractal simulation Original Video: Period 25ms Recovered Video 33 Captured by 10fps camera Captured by 10fps coded strobing camera

Rotating Mill Tool (Dragonfly2) Normal Video: 10fps Reconstructed Video 4500 RPM Reconstructed Video 3000 RPM Reconstructed Video 6000 RPM Reconstructed Video 12000 RPM 34

Compressive Sensing: Linear Dynamical Systems Aswin Sankaranarayanan Pavan Turaga Richard Baraniuk Rama Chellappa

CSLDS: Video CS under LDS model Motivation Extensive use in dynamic textures, sequence of LDS for activity analysis, etc LDS == Frames of the video lie on a low dimensional subspace. Linearity applicable over short durations Benefits LDS is a parametric model. Video CS is equivalent to parameter estimation Majority of the LDS parameters are static. This is extremely helpful in achieving high compression at sensing

Linear dynamical system example Video sequence Few frames Six key images All frames can be estimated using linear combinations of SIX images

LDS model Observation model Data lies on a low dimensional subspace C is a basis for the subspace. For uncompressed data, C is usually learnt using PCA on training examples. x t R N x t, v t Cv, d Terminology x t are the frames of the video (high dim) v t are the subspace coeffficients (low dim) R d t N

Compressive LDS Photo-detector x t R N x t, v t Cv R d t, d N Digital micro-mirror device y t y t A x R M t, M AC v N t Can we recover C and v t from y t? Very hard since the unknowns are bilinear

Challenges and solution Key challenge: Unknown variables appear as bilinear relationships CS theory handles unknowns in a linear relationship Solution: A novel measurement model that reduces the bilinear problem to a sequence of linear/convex problems Stable reconstruction at a measurement rate proportional to the state space dimension. Scene x :T CS camera Common measurements 1 ~ t Innovations measurements y 1:T ~ y 1:T Estimate state sequence using SVD v ˆ1 :T Model based CoSaMP for recovery of observation matrix xˆ Cˆ t v t Ĉ x ˆ1 :T

Results Ground truth Comp 20x SNR: 24.9 db Comp 50x SNR: 20.4 db Ground truth Comp20x Comp50x SNR: 25.3 db SNR:23.8 db

Linear Motion Model: Invertible Motion Blur in Video Amit Agrawal Yi Xu Ramesh Raskar

Traditional Exposure Video Motion PSF (Box Filter) Fourier Transform Information is lost Exposure Time Slide courtesy Agrawal

Varying Exposure Video Exposure Time Fourier Transform Slide courtesy Agrawal

Varying Exposure Video No common nulls Exposure Time Fourier Transform Exposure Time Slide courtesy Agrawal

Varying Exposure Video No common nulls Exposure Time Fourier Transform Exposure Time Exposure Time Slide courtesy Agrawal

Varying Exposure Video = PSF Null-Filling Fourier Transform Joint Frequency Spectrum Preserves All Frequencies Slide courtesy Agrawal

Varying Exposure Video Slide courtesy Agrawal

Blurred Photos Deblurred Result Slide courtesy Agrawal

Single Image Deblurring (SID) Multiple Image Deblurring (MID) Slide courtesy Agrawal

The Story so far Periodic or Repetitive Videos CS Strobing Linear Dynamical Systems CS LDS Linear Motion Model Multi-Exposure Video

How about general Videos? General Multiplexing Single Pixel Camera Voxel Subsampling Flexible Voxels Coded per-pixel Exposure Video Voxel Multiplexing Programmable Pixel Compressive Camera

Single pixel camera Each configuration of micro-mirrors yield ONE compressive measurement Photo-detector Non-visible wavelengths Sensor material costly in IR/UV bands Digital micro-mirror device Light throughput Half the light in the scene is directed to the photo-detector

Single-Pixel Camera M randomized measurements target N=65536 pixels M=11000 measurements (16%) M=1300 measurements (2%)

From Image to Video Sensing SPC works under the assumption of a static scene However, for dynamic scenes changes as we collect measurements Need to measure temporal events at their information rate Need to borrow richer models for videos and exploit them at sensing and reconstruction

Single Pixel Camera in SWIR Single pixel camera setup Object InGaAs Photo-detector (Short-wave IR) SPC sampling rate: 10,000 sample/s Number of compressive measurements: M = 16,384 Recovered video: N = 128 x 128 x 61 = 61*M

Joint work with Xu and Kelly CS-MUVI: IR spectrum initial estimate Upsampled Recovered Video

CS-MUVI on SPC Naïve frame-to-frame recovery CS-MUVI Joint work with Xu and Kelly

Voxel Subsampling Compressive Video Flexible Voxels for Content Aware Videography ECCV 2010 Mohit Gupta, Amit Agrawal, Ashok Veeraraghavan, Srinivas Narasimhan Video from a Single Exposure Coded Photograph ICCV 2011 Yasunobu Hitomi, Jinwei Gu, Mohit Gupta, Tomoo Mitsunaga, Shree K. Nayar

Temporal resolution (fps) Fundamental Resolution Trade-off 500 High-speed Cameras 30 Conventional video camera 64K 256x256 1M 1024x1024 Spatial resolution (pixels)

Imaging Dynamic Scenes Low-Frame Rate (30 fps) Camera Large Motion Blur High-Speed (480 fps) Camera Low Spatial Resolution 68

Motion Aware Camera Variable Spatio-Temporal Resolution 69

Spatio-Temporal Sampling High SR, Low TR Low SR, High TR Motion Aware Video Fast Moving y Static Slow Moving t x Thin and long voxels Fat and short voxels Flexible voxels

Sampling of the Space-Time Volume Conventional Sampling Scheme: Sensor Plane Frame 1 Frame 2 Frame N Camera Integration Time Time Our Sampling Scheme: Frame 1 Frame 2 Frame N Camera

Space Multi-resolution Sampling Captured Data Desired Spatio-temporal Interpretations Re-binning Time TR = 1X SR = 1X TR = 2X SR = ½ X TR = 4X SR = ¼X TR = Gain in Temporal Resolution SR = Gain in Spatial Resolution

TR = 1X SR = 1X TR = 2X SR = ½ X TR = 4X SR = ¼X Multi-resolution Sampling Captured Data Different Spatio-temporal Interpretations Re-binning What is the correct firing order? TR = 1X SR = 1X TR = 2X SR = ½ X Unsampled Bins TR = 4X SR = ¼X Re-binning

Computing the Firing Order Captured Data Desired Spatio-temporal Interpretations x p 0,1 B ij p = location index K = Number of pixels in a group B ij = i th block in j th interpretation Integer Linear Program: Each Block contains exactly one measurement p B ij x p 1 B ij Number of measurements = Number of pixels K 2 p 1 x p K

y Firing Order and Reconstructions for a 4x4 Block 1 8 4 5 16 2 15 9 7 10 13 3 14 12 6 x 11 16 15 1 16 2 15 1 2 9 10 8 9 7 10 8 7 4 13 3 14 13 14 4 3 5 12 6 11 12 11 5 6 TR = 1, SR = 1/1 TR = 2, SR = 1/2 16 15 t 9 10 13 14 8 7 12 11 1 2 5 6 4 3 16 15 t 13 14 12 11 9 10 8 7 5 6 4 3 1 2 t 15 16 12 13 14 t 1 2 3 4 5 6 7 8 9 10 11 TR = 4, SR = 1/4 TR = 8, SR = 1/8 TR = 16, SR = 1/16

Co-located Projector-Camera Setup Scene Camera Integration Time Projector Pattern Beam Splitter Image Plane Image Plane Projector Pixel 1 Pixel 2 Pixel K Camera Time

Co-located Projector-Camera Setup Scene Beam Splitter Image Plane Image Plane Projector Camera

LCOS Hardware Prototype 78

Multiple Balls Bouncing and Colliding (15 FPS) Close-up Large Motion Blur

Reconstruction at 4X Temporal Resolution (60FPS) Aliasing Artifacts on the Background

Reconstruction at 16X Temporal Resolution (240FPS) Aliasing Artifacts on the Background

Motion-aware Video Increasing Temporal Resolution + + Captured Frame Different Spatio-temporal Interpretations Motion Analysis Optical Motion-Aware Flow Magnitudes Video

Motion-aware Video Simple Rebinning Motion-Aware Video

Multiple Balls Bouncing Input Sequence Motion-Aware Video

Comparison Input Sequence Motion-Aware Video

Voxel Subsampling with Dictionary Learning? Slide Courtesy Hitomi et. al.

Slide Courtesy Hitomi et. al. Pixel-wise Coded Exposure Sampling. = Original scene Exposure pattern (white=on, black=off) Input Coded image

Dictionary learning Learning video set 20 scenes, 8 spatial rotations, Play forward and backward Overcomplete dictionary bases Slide Courtesy Hitomi et. al. 91

Simulations Slide Courtesy Hitomi et. al. 92

Road Traffic Original scene (450x300x36) Exposure Coded image Pattern (450x300x36) (450x300) Slide Courtesy Hitomi et. al. 93

Simulation Results Thin-out movie (13x9x36 450x300x36) Grid sampling + interpolation (450x300x36) 3D DCT representation + sparse reconstruction (450x300x36) Our method (450x300x36) Slide Courtesy Hitomi et. al. 94

Dog s Tongue Original scene (384x216x36) Coded image (384x216) Slide Courtesy Hitomi et. al. 95

Simulation Results Slide Courtesy Hitomi et. al. Thin-out movie (11x6x36 384x216x36) Grid sampling + interpolation (384x216x36) 3D DCT representation + sparse reconstruction (384x216x36) Our method (384x216x36) 96

Experimental Results

Slide Courtesy Hitomi et. al. Falling Ball Coded image (320x320, 18ms) Reconstructed Movie (320x320x18)

Slide Courtesy Hitomi et. al. Fluid (milk) pouring Coded image (320x320, 18ms) Reconstructed Movie (320x320x18)

Slide Courtesy Hitomi et. al. Blinking Eye Coded image (320x320, 27ms) Reconstructed Movie (320x320x9)

Voxel Multiplexing Compressive Video Programmable Pixel Compressive Camera CVPR 2011 Dikpal Reddy, Ashok Veeraraghavan, Rama Chellappa

P2C2: Programmable Pixel Compressive Camera Fast moving scene Capture Camera operating at 25 fps 0.25 megapixel Sensor array Captured low-res coded frames from a 25fps camera Spatio-temporal mask modulates every pixel independently. Mask has 1 megapixels and modulates at 200 fps. Recovery Optimization using Brightness constancy & Wavelet domain sparsity Captured low-res coded frames from a 25fps camera Video credit: TechImaging

P2C2: Prototype Objective Lens F=80 mm D=40 mm Polarizing beamsplitter LCOS Mirror SXGA-3DM Camera Lens Pentax 35mm

P2C2: Programmable pixel compressive camera Camera operating at 25 fps X (:, :, t) X Modulation at 200 fps (:, :, t) M (:, :, t) Integration Coded motion blur Subsampling spatially by 2 D( ) Frame captured by P2C2 Y (:, :, tl ) Reconstruction Algorithm Frame captured by P2C2 Temporal superres Spatial superres L t L s 8 2

P2C2: Programmable pixel compressive camera Observations and unknowns L s 1, Lt 4 = P2C2 captured frame y 2 mask m 6 high speed frame x 6

P2C2: Programmable pixel compressive camera Observations and unknowns L s 1, Lt 4 = y x Severely ill-conditioned system of equations (2 16 x 2 20 ) E data y x 2 2

P2C2: Brightness constancy constraints t=1 t=2 t=3 t=4 Forward flow -1 1 x 1 x 2 = 0 x 3 x 4 Brightness constancy x2( ) x1 () 0

P2C2: Brightness constancy constraints t=1 t=2 t=3 t=4 Forward flow -1 1 x 1 Backward flow 1-1 -1.1.3.2.4 x 2 = 0 Subpixel flow Bilinear interpolation Row removed x 3 Occluded pixel x 4 Pixel occluded in t=4 x 0 E x BC 2 2

P2C2: Transform sparsity Image DCT Wavelet: DB6 x t 1 x t 1 E x spatial t 1

P2C2: Dancers video Original high speed scene P2C2 capture Normal camera capture x y Temporal superres Spatial superres L t L s 8 1 Video credit: Photron

If brightness constancy constraints available Solve min P2C2: Cost function E data E BC E spatial min y x But optical flow not available apriori 2 2 x t 1 Optical flow can be extracted from smooth, low spatial resolution images 2 2 T 1 x t 1 Estimate sub-frames without brightness constancy constraints

Recover sub-frames by solving Solve min P2C2: Initial estimate E data E spatial x 0 arg min Significantly ill conditioned y x 2 2 T t 1 1 x t 1 Set high-frequency coefficients to zero x 0 smooth, high spatial frequency is lost preserves motion allowing OF to be estimated

P2C2: Estimate Optical Flow Recover flow from estimated sub-frames Perform consistency check using backward and forward flow Perform occlusion reasoning to prune the brightness constancy constraints OF from original high speed frames OF from initial recovered frames abs( u iv) abs( u 0 iv 0 )

If brightness constancy constraints available Solve min P2C2: Cost function E data E BC E spatial min y x But optical flow not available apriori 2 2 x t 1 Optical flow can be extracted from smooth, low spatial resolution images 2 2 T 1 x t 1 Estimate sub-frames without brightness constancy constraints

P2C2: Algorithm Estimate sub-frames x 0 without using BC constraints Begin iteration (iteration index k) Estimate OF ( u k, v k 1 1 ) from k 1 x Build BC constraint matrix k 1 Estimate sub-frames End iteration x k by minimizing E total

P2C2: Recovered Examples Reconstructed scene Normal camera capture Original high speed scene row Row-Time slice time occlusion

P2C2: Recovered Examples Reconstructed scene Normal camera capture Original high speed scene row Row-Time slice time occlusion

How about general Videos? General Multiplexing Single Pixel Camera Voxel Subsampling Flexible Voxels Coded per-pixel Exposure Video Voxel Multiplexing Programmable Pixel Compressive Camera

Motion Blur Whats next? Flutter Shutter Parabolic Exposure Rolling Shutter Cameras Coded Rolling Shutter High Dimensional Visual Compressive Sensing Compressive Light Transport Compressive volumetric Scattering Reinterpretable Imaging