Short-course Compressive Sensing of Videos

Similar documents
Compressive. Graphical Models. Volkan Cevher. Rice University ELEC 633 / STAT 631 Class

Compressive Sensing. A New Framework for Sparse Signal Acquisition and Processing. Richard Baraniuk. Rice University

Randomized Dimensionality Reduction

The Fundamentals of Compressive Sensing

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

Compressive Sensing. Part IV: Beyond Sparsity. Mark A. Davenport. Stanford University Department of Statistics

Compressive Sensing of High-Dimensional Visual Signals. Aswin C Sankaranarayanan Rice University

Compressive Sensing. and Applications. Volkan Cevher Justin Romberg

Compressive Sensing: Opportunities and Perils for Computer Vision

ELEG Compressive Sensing and Sparse Signal Representations

Tutorial: Compressive Sensing of Video

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Compressive Sensing: Theory and Practice

Sparse Reconstruction / Compressive Sensing

Introduction to Topics in Machine Learning

Compressive Sensing: Opportunities and pitfalls for Computer Vision

Agenda. Pressure is on Digital Sensors

Compressive Sensing for High-Dimensional Data

Compressive Parameter Estimation with Earth Mover s Distance via K-means Clustering. Dian Mo and Marco F. Duarte

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

Lecture 17 Sparse Convex Optimization

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

Collaborative Sparsity and Compressive MRI

Tutorial on Image Compression

Modified Iterative Method for Recovery of Sparse Multiple Measurement Problems

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

Deconvolution with curvelet-domain sparsity Vishal Kumar, EOS-UBC and Felix J. Herrmann, EOS-UBC

COMPRESSIVE VIDEO SAMPLING

Non-Differentiable Image Manifolds

Sparse wavelet expansions for seismic tomography: Methods and algorithms

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Parameter Estimation in Compressive Sensing. Marco F. Duarte

Learning Splines for Sparse Tomographic Reconstruction. Elham Sakhaee and Alireza Entezari University of Florida

Main Menu. Summary. sampled) f has a sparse representation transform domain S with. in certain. f S x, the relation becomes

Detection Performance of Radar Compressive Sensing in Noisy Environments

The Benefit of Tree Sparsity in Accelerated MRI

Randomized sampling strategies

Recovery of Piecewise Smooth Images from Few Fourier Samples

Structurally Random Matrices

Algebraic Iterative Methods for Computed Tomography

Compressed Sensing and Applications by using Dictionaries in Image Processing

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

The convex geometry of inverse problems

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Face Recognition via Sparse Representation

Inverse Problems and Machine Learning

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

Incoherent noise suppression with curvelet-domain sparsity Vishal Kumar, EOS-UBC and Felix J. Herrmann, EOS-UBC

A Study on Compressive Sensing and Reconstruction Approach

SPARSITY ADAPTIVE MATCHING PURSUIT ALGORITHM FOR PRACTICAL COMPRESSED SENSING

Robust image recovery via total-variation minimization

Compressive Sensing based image processing in TrapView pest monitoring system

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis

Generalized Principal Component Analysis CVPR 2007

Compressive Topology Identification of Interconnected Dynamic Systems via Clustered Orthogonal Matching Pursuit

Image reconstruction based on back propagation learning in Compressed Sensing theory

arxiv: v1 [stat.ml] 14 Jan 2017

Tomographic reconstruction: the challenge of dark information. S. Roux

Adaptive step forward-backward matching pursuit algorithm

Compressed Sensing for Rapid MR Imaging

Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm

RECONSTRUCTION ALGORITHMS FOR COMPRESSIVE VIDEO SENSING USING BASIS PURSUIT

Blind Compressed Sensing Using Sparsifying Transforms

Algebraic Iterative Methods for Computed Tomography

Learning based face hallucination techniques: A survey

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

The convex geometry of inverse problems

Sparse Models in Image Understanding And Computer Vision

IMAGE DENOISING TO ESTIMATE THE GRADIENT HISTOGRAM PRESERVATION USING VARIOUS ALGORITHMS

Total Variation Denoising with Overlapping Group Sparsity

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

SPARSE signal representations provide a general signal

CS 664 Segmentation. Daniel Huttenlocher

Image Reconstruction from Multiple Sparse Representations

Compressed Sensing and L 1 -Related Minimization

Structure from Motion

Outlier Pursuit: Robust PCA and Collaborative Filtering

Open Access Reconstruction Technique Based on the Theory of Compressed Sensing Satellite Images

Signal Processing with Side Information

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

Motivation. My General Philosophy. Assumptions. Advanced Computer Graphics (Spring 2013) Precomputation-Based Relighting

Array geometries, signal type, and sampling conditions for the application of compressed sensing in MIMO radar

Coded Acquisition of High Speed Videos with. Multiple Cameras

ELEC Dr Reji Mathew Electrical Engineering UNSW

Recognition, SVD, and PCA

Stereo Vision. MAN-522 Computer Vision

Department of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2

Iterative CT Reconstruction Using Curvelet-Based Regularization

Distributed Compressed Estimation Based on Compressive Sensing for Wireless Sensor Networks

EXAM SOLUTIONS. Image Processing and Computer Vision Course 2D1421 Monday, 13 th of March 2006,

An Approach for Reduction of Rain Streaks from a Single Image

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Applications: Sampling/Reconstruction. Sampling and Reconstruction of Visual Appearance. Motivation. Monte Carlo Path Tracing.

Spatially-Localized Compressed Sensing and Routing in Multi-Hop Sensor Networks 1

SPARSE SIGNAL RECONSTRUCTION FROM NOISY COMPRESSIVE MEASUREMENTS USING CROSS VALIDATION. Petros Boufounos, Marco F. Duarte, Richard G.

When Sparsity Meets Low-Rankness: Transform Learning With Non-Local Low-Rank Constraint For Image Restoration

Image Reconstruction based on Block-based Compressive Sensing

Hyperspectral Data Classification via Sparse Representation in Homotopy

Transcription:

Short-course Compressive Sensing of Videos Venue CVPR 2012, Providence, RI, USA June 16, 2012 Organizers: Richard G. Baraniuk Mohit Gupta Aswin C. Sankaranarayanan Ashok Veeraraghavan

Part 2: Compressive sensing Motivation, theory, recovery Linear inverse problems Sensing visual signals Compressive sensing Theory Hallmark Recovery algorithms Model-based compressive sensing Models specific to visual signals

Linear inverse problems

Linear inverse problems Many classic problems in computer can be posed as linear inverse problems Notation Signal of interest Observations x 2 R N y 2 R M measurement matrix Measurement model Problem definition: given y = x + e measurement noise y, recover x

Linear inverse problems y = x + e x 2 R N y 2 R M Problem definition: given y, recover x Scenario 1 We can invert the system of equations ˆx = M N 1 y Focus more on robustness to noise via signal priors

Linear inverse problems y = x + e x 2 R N y 2 R M Problem definition: given y, recover x Scenario 2 Measurement matrix has a (N-M) dimensional null-space Solution is no longer unique Many interesting vision problem fall under this scenario Key quantity of concern: Under-sampling ratio M/N

Image super-resolution Low resolution input/observation 128x128 pixels

Image super-resolution y = x + e 2x super-resolution y (1,1) =(x 1,1 + x 1,2 + x 2,1 + x 2,2 ) /4

Image super-resolution y = x + e 4x super-resolution y (1,1) =(x 1,1 + x 1,2 +...+ x 4,3 + x 4,4 ) /16

Image super-resolution y = x + e Super-resolution factor D Under-sampling factor M/N = 1/D 2 General rule: The smaller the under-sampling, the more the unknowns and hence, the harder the superresolution problem

Many other vision problems Affine rank minimizaiton, Matrix completion, deconvolution, Robust PCA Image synthesis Infilling, denoising, etc. Light transport Reflectance fields, BRDFs, Direct global separation, Light transport matrices Sensing

Sensing visual signals

High-dimensional visual signals Reflection Fog Volumetric scattering Human skin Sub-surface scattering Refraction Electron microscopy Tomography

Adelson and Bergen (91) The plenoptic function Collection of all variations of light in a scene time (1D) space (3D) spectrum (1D) angle (2D) Different slices reveal different scene properties

The plenoptic function time (1D) space (3D) angle (2D) spectrum (1D) High-speed cameras Hyper-spectral imaging Lytro light-field camera

Sensing the plenoptic function High-dimensional 1000 samples/dim == 10^21 dimensional signal Greater than all the storage in the world Traditional theories of sensing fail us!

Resolution trade-off Key enabling factor: Spatial resolution is cheap! Commercial cameras have 10s of megapixels One idea is the we trade-off spatial resolution for resolution in some other axis

Spatio-angular tradeoff [Ng, 2005]

Spatio-angular tradeoff [Levoy et al. 2006]

Spatio-temporal tradeoff Stagger pixel-shutter within each exposure [Bub et al., 2010]

Spatio-temporal tradeoff Rearrange to get high temporal resolution video at lower spatialresolution [Bub et al., 2010]

Resolution trade-off Very powerful and simple idea Drawbacks Does not extend to non-visible spectrum 1 Megapixel SWIR camera costs 50-100k Linear and global tradeoffs With today s technology, cannot obtain more than 10x for video without sacrificing spatial resolution completely

Compressive sensing

Sense by Sampling sample

Sense by Sampling sample too much data!

Sense then Compress sample compress JPEG JPEG2000 decompress

Sparsity pixels large wavelet coefficients (blue = 0)

Sparsity pixels large wavelet coefficients (blue = 0) wideband signal samples frequency large Gabor (TF) coefficients time

Concise Signal Structure Sparse signal: only K out of N coordinates nonzero sorted index

Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional subspaces aligned w/ coordinate axes sorted index

Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional subspaces Compressible signal: sorted coordinates decay rapidly with power-law power-law decay sorted index

Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional subspaces Compressible signal: sorted coordinates decay rapidly with power-law model: ball: power-law decay sorted index

What s Wrong with this Picture? Why go to all the work to acquire N samples only to discard all but K pieces of data? sample compress decompress

What s Wrong with this Picture? linear processing linear signal model (bandlimited subspace) nonlinear processing nonlinear signal model (union of subspaces) sample compress decompress

Compressive Sensing Directly acquire compressed data via dimensionality reduction Replace samples by more general measurements compressive sensing recover

Sampling Signal is -sparse in basis/dictionary WLOG assume sparse in space domain sparse signal nonzero entries

Sampling Signal is -sparse in basis/dictionary WLOG assume sparse in space domain Sampling measurements sparse signal nonzero entries

Compressive Sampling When data is sparse/compressible, can directly acquire a condensed representation with no/little information loss through linear dimensionality reduction measurements sparse signal nonzero entries

How Can It Work? Projection not full rank and so loses information in general Ex: Infinitely many (null space) s map to the same

How Can It Work? Projection not full rank and so loses information in general columns But we are only interested in sparse vectors

How Can It Work? Projection not full rank and so loses information in general columns But we are only interested in sparse vectors is effectively MxK

How Can It Work? Projection not full rank and so loses information in general columns But we are only interested in sparse vectors Design so that each of its MxK submatrices are full rank (ideally close to orthobasis) Restricted Isometry Property (RIP)

Restricted Isometry Property (RIP) Preserve the structure of sparse/compressible signals K-dim subspaces

Restricted Isometry Property (RIP) Stable embedding RIP of order 2K implies: for all K-sparse x 1 and x 2 K-dim subspaces

RIP = Stable Embedding An information preserving projection preserves the geometry of the set of sparse signals RIP ensures that

How Can It Work? Projection not full rank and so loses information in general columns Design so that each of its MxK submatrices are full rank (RIP) Unfortunately, a combinatorial, NP-complete design problem

Insight from the 70 s [Kashin, Gluskin] Draw at random iid Gaussian iid Bernoulli columns Then provided has the RIP with high probability

Randomized Sensing Measurements = random linear combinations of the entries of the signal y = x No information loss for sparse vectors whp measurements sparse signal nonzero entries

CS Signal Recovery Goal: Recover signal from measurements Problem: Random projection not full rank (ill-posed inverse problem) Solution: Exploit the sparse/compressible geometry of acquired signal

Random projection not full rank CS Signal Recovery Recovery problem: given find Null space Search in null space for the best according to some criterion ex: least squares (N-M)-dim hyperplane at random angle

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: Closed-form solution:

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: Closed-form solution: Wrong answer!

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: Closed-form solution: Wrong answer!

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: find sparsest vector in translated nullspace

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: Correct! find sparsest vector in translated nullspace But NP-Complete alg

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: Convexify the optimization Candes Romberg Tao Donoho

Signal Recovery Recovery: given (ill-posed inverse problem) find (sparse) Optimization: Convexify the optimization Correct! Polynomial time alg (linear programming)

Compressive Sensing random measurements sparse signal nonzero entries Signal recovery via [Candes, Romberg, Tao; Donoho] optimization

Compressive Sensing random measurements sparse signal nonzero entries Signal recovery via iterative greedy algorithm (orthogonal) matching pursuit [Gilbert, Tropp] iterated thresholding [Nowak, Figueiredo; Kingsbury, Reeves; Daubechies, Defrise, De Mol; Blumensath, Davies; ] CoSaMP [Needell and Tropp]

Greedy recovery algorithm #1 Consider the following problem min ky xk 2 s.t kxk 0 apple K Suppose we wanted to minimize just the cost, then steepest gradient descent works as bx k = bx k 1 + T (y bx k 1 ) But, the new estimate is no longer K-sparse bx k =thresh bx k 1 + T (y bx k 1 ),K

Iterated Hard Thresholding update signal estimate prune signal es4mate (best K- term approx) update residual

Greedy recovery algorithm #2 Consider the following problem sparse signal 1 sparse Can we recover the support?

Greedy recovery algorithm #2 Consider the following problem sparse signal 1 sparse If then =[ 1, 2,..., N ] arg max h i,yi gives the support of x How to extend to K-sparse signals?

Greedy recovery algorithm #2 sparse signal K sparse residue: find atom: r = y bx k 1 k = arg max h i,ri Add atom to support: S = S [ {k} Signal estimate x (Least squares over support) k =( S ) y

Orthogonal matching pursuit goal: given y = columns of x, recover a sparse x are unit-norm initialize: bx 0 =0,r = y, ={},i=0 iteration: i = i +1 T r b = k = arg max{ b(1), b(2),..., b(n) } = S k (bx i ) =( ) y, (bx i ) c =0 Find atom with largest support Update signal estimate r = y bx i update residual

Specialized solvers CoSAMP [Needell and Tropp, 2009] SPG_l1 [Friedlander, van der Berg, 2008] http://www.cs.ubc.ca/labs/scl/spgl1/ FPC [Hale, Yin, and Zhang, 2007] http://www.caam.rice.edu/~optimization/l1/fpc/ AMP [Donoho, Montanari and Maleki, 2010] many many others, see dsp.rice.edu/cs and https://sites.google.com/site/igorcarron2/cscodes

CS Hallmarks Stable acquisition/recovery process is numerically stable Asymmetrical (most processing at decoder) conventional: smart encoder, dumb decoder CS: dumb encoder, smart decoder Democratic each measurement carries the same amount of information robust to measurement loss and quantization digital fountain property Random measurements encrypted Universal same random projections / hardware can be used for any sparse signal class (generic)

Universality Random measurements can be used for signals sparse in any basis

Universality Random measurements can be used for signals sparse in any basis

Universality Random measurements can be used for signals sparse in any basis sparse coefficient vector nonzero entries

Summary: CS Compressive sensing randomized dimensionality reduction exploits signal sparsity information integrates sensing, compression, processing Why it works: with high probability, random projections preserve information in signals with concise geometric structures sparse signals compressible signals

Summary: CS Encoding: = random linear combinations of the entries of measurements sparse signal nonzero entries Decoding: Recover from via optimization

Image/Video specific signal models and recovery algorithms

Transform basis Recall Universality: Random measurements can be used for signals sparse in any basis DCT/FFT/Wavelets Fast transforms; very useful in large scale problems

Dictionary learning For many signal classes (ex: videos, light-fields), there are no obvious sparsifying transform basis Can we learn a sparsifying transform instead? GOAL: Given training data x 1,x 2,...,x T xi 2 R N learn a dictionary D, such that s i are sparse. x i = Ds i D 2 R N Q s i 2 R Q

Dictionary learning GOAL: Given training data x 1,x 2,...,x T xi 2 R N learn a dictionary D, such that s i are sparse. x i = Ds i D 2 R N Q s i 2 R Q min D,S kx DSk F s.t 8i, ks i k 0 apple K

Dictionary learning min D,S kx DSk F s.t 8i, ks i k 0 apple K Non-convex constraint Bilinear in D and S

Dictionary learning min D,S kx DSk F + X k ks k k 1 Biconvex in D and S Given D, the optimization problem is convex in s k Given S, the optimization problem is a least squares problem K-SVD: Solve using alternate minimization techniques Start with D = wavelet or DCT bases Additional pruning steps to control size of the dictionary Aharon et al., TSP 2006

Dictionary learning min D,S kx DSk F + X k ks k k 1 Pros Ability to handle arbitrary domains Cons Learning dictionaries can be computationally intensive for high-dimensional problems; need for very large amount of data Recovery algorithms may suffer due to lack of fast transforms

Models on image gradients Piecewise constant images Sparse image gradients Natural image statistics Heavy tailed distributions

Total variation prior TV(I) = X u,v kri(u, v)k 2 = X u,v q I 2 x(u, v)+i 2 y(u, v) TV norm Sparse-gradient promoting norm Formulation of recovery problem min TV(x) s.t y = x

Total variation prior Optimization problem Convex Often, works better than transform basis methods Variants 3D (video) Anisotropic TV min TV(x) s.t y = x Code TVAL3 Many many others (see dsp.rice/cs)

Beyond sparsity Model-based CS

Beyond Sparse Models Sparse signal model captures simplistic primary structure wavelets: natural images Gabor atoms: chirps/tones pixels: background subtracted images

Beyond Sparse Models Sparse signal model captures simplistic primary structure Modern compression/processing algorithms capture richer secondary coefficient structure wavelets: natural images Gabor atoms: chirps/tones pixels: background subtracted images

Sparse Signals K-sparse signals comprise a particular set of K-dim subspaces

Structured-Sparse Signals A K-sparse signal model comprises a particular (reduced) set of K-dim subspaces [Blumensath and Davies] Fewer subspaces <> relaxed RIP <> stable recovery using fewer measurements M

Wavelet Sparse Typical of wavelet transforms of natural signals and images (piecewise smooth)

Tree-Sparse Model: K-sparse coefficients + significant coefficients lie on a rooted subtree Typical of wavelet transforms of natural signals and images (piecewise smooth)

Wavelet Sparse Model: K-sparse coefficients + significant coefficients lie on a rooted subtree RIP: stable embedding K-planes

Tree-Sparse Model: K-sparse coefficients + significant coefficients lie on a rooted subtree Tree-RIP: stable embedding [Blumensath and Davies] K-planes

Tree-Sparse Model: K-sparse coefficients + significant coefficients lie on a rooted subtree Tree-RIP: [Blumensath and Davies] stable embedding Recovery: inject tree-sparse approx into IHT/CoSaMP

Recall: Iterated Thresholding update signal estimate prune signal es4mate (best K- term approx) update residual

Iterated Model Thresholding update signal estimate prune signal es4mate (best K- term model approx) update residual

Tree-Sparse Signal Recovery signal length N=1024 random measurements M=80 target signal CoSaMP, (RMSE=1.12) Tree-sparse CoSaMP (RMSE=0.037) L1-minimization (RMSE=0.751)

Clustered Signals Probabilistic approach via graphical model Model clustering of significant pixels in space domain using Ising Markov Random Field Ising model approximation performed efficiently using graph cuts [Cevher, Duarte, Hegde, Baraniuk 08] target Ising-model recovery CoSaMP recovery LP (FPC) recovery

Part 2: Compressive sensing Motivation, theory, recovery Linear inverse problems Sensing visual signals Compressive sensing Theory Hallmark Recovery algorithms Model-based compressive sensing Models specific to visual signals