Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Similar documents
Compressive. Graphical Models. Volkan Cevher. Rice University ELEC 633 / STAT 631 Class

Compressive Sensing. A New Framework for Sparse Signal Acquisition and Processing. Richard Baraniuk. Rice University

Compressive Sensing for High-Dimensional Data

Compressive Sensing: Theory and Practice

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Randomized Dimensionality Reduction

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

Sparse Reconstruction / Compressive Sensing

Ultra-low power wireless sensor networks: distributed signal processing and dynamic resources management

Introduction to Topics in Machine Learning

TERM PAPER ON The Compressive Sensing Based on Biorthogonal Wavelet Basis

Robust image recovery via total-variation minimization

G Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing

Compressive Sensing. and Applications. Volkan Cevher Justin Romberg

Compressive Imaging for Video Representation and Coding

Face Recognition via Sparse Representation

The Fundamentals of Compressive Sensing

Recovery of Piecewise Smooth Images from Few Fourier Samples

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Agenda. Pressure is on Digital Sensors

Structurally Random Matrices

Compressive Sensing. Part IV: Beyond Sparsity. Mark A. Davenport. Stanford University Department of Statistics

Weighted-CS for reconstruction of highly under-sampled dynamic MRI sequences

COMPRESSIVE VIDEO SAMPLING

Compressed Sensing and Applications by using Dictionaries in Image Processing

Non-Differentiable Image Manifolds

Robust Video Coding. Heechan Park. Signal and Image Processing Group Computer Science Department University of Warwick. for CS403

Randomized sampling strategies

Compressive Sensing of High-Dimensional Visual Signals. Aswin C Sankaranarayanan Rice University

Collaborative Sparsity and Compressive MRI

Multiple-View Object Recognition in Band-Limited Distributed Camera Networks

Adaptive compressed image sensing based on wavelet-trees

LOSSLESS COMPRESSION OF ENCRYPTED GREY-LEVEL AND COLOR IMAGES

Lecture 17 Sparse Convex Optimization

Compressive Sensing: Opportunities and Perils for Computer Vision

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

Off-the-Grid Compressive Imaging: Recovery of Piecewise Constant Images from Few Fourier Samples

Compressive Sensing: Opportunities and pitfalls for Computer Vision

Compressive Sensing Based Image Reconstruction using Wavelet Transform

Spatially-Localized Compressed Sensing and Routing in Multi-Hop Sensor Networks 1

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

Wireless Networks Research Seminar April 22nd 2013

ELEG Compressive Sensing and Sparse Signal Representations

Efficient Data Collection with Sampling in WSNs: Making Use of Matrix Completion Techniques

1-bit Compressive Data Gathering for Wireless Sensor Networks

Image Reconstruction from Multiple Sparse Representations

Distributed Compressed Estimation Based on Compressive Sensing for Wireless Sensor Networks

University of Alberta. Parallel Sampling and Reconstruction with Permutation in Multidimensional Compressed Sensing. Hao Fang

Image Denoising Based on Hybrid Fourier and Neighborhood Wavelet Coefficients Jun Cheng, Songli Lei

Tutorial on Image Compression

The Smashed Filter for Compressive Classification and Target Recognition

Compressed Sensing for Rapid MR Imaging

EE123 Digital Signal Processing

THE TRANSFORM AND DATA COMPRESSION HANDBOOK

Tomographic reconstruction: the challenge of dark information. S. Roux

Image reconstruction based on back propagation learning in Compressed Sensing theory

Randomized sampling without repetition in timelapse seismic surveys

The limits of adaptive sensing

CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT

A Comprehensive Review of Distributed Coding Algorithms for Visual Sensor Network (VSN)

Compressive Parameter Estimation with Earth Mover s Distance via K-means Clustering. Dian Mo and Marco F. Duarte

SPARSE SIGNAL RECONSTRUCTION FROM NOISY COMPRESSIVE MEASUREMENTS USING CROSS VALIDATION. Petros Boufounos, Marco F. Duarte, Richard G.

REMOTE-SENSED LIDAR USING RANDOM SAMPLING AND SPARSE RECONSTRUCTION

Compressive Sensing based image processing in TrapView pest monitoring system

Bipartite Graph based Construction of Compressed Sensing Matrices

VC 12/13 T16 Video Compression

Short-course Compressive Sensing of Videos

IMAGE DE-NOISING IN WAVELET DOMAIN

A Study on Compressive Sensing and Reconstruction Approach

Image/video compression: howto? Aline ROUMY INRIA Rennes

Rank Minimization over Finite Fields

Introduction to Video Compression

A Representative Sample Selection Approach for SRC

Frequency Band Coding Mode Selection for Key Frames of Wyner-Ziv Video Coding

Optimal Sampling Geometries for TV-Norm Reconstruction of fmri Data

Compressive Single Pixel Imaging Andrew Thompson University of Edinburgh. 2 nd IMA Conference on Mathematics in Defence

Introduction. Wavelets, Curvelets [4], Surfacelets [5].

ComputerLab: compressive sensing and application to MRI

MOTION ESTIMATION AT THE DECODER USING MAXIMUM LIKELIHOOD TECHNIQUES FOR DISTRIBUTED VIDEO CODING. Ivy H. Tseng and Antonio Ortega

Binary Graphs and Message Passing Strategies for Compressed Sensing in the Noiseless Setting

Lecture 5: Compression I. This Week s Schedule

Compression of RADARSAT Data with Block Adaptive Wavelets Abstract: 1. Introduction

Compressive Sensing Algorithms for Fast and Accurate Imaging

Computational Photography Denoising

GPR IMAGING USING COMPRESSED MEASUREMENTS

Topic 5 Image Compression

Sparse sampling in MRI: From basic theory to clinical application. R. Marc Lebel, PhD Department of Electrical Engineering Department of Radiology

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

MULTIMEDIA COMMUNICATION

Main Menu. Summary. sampled) f has a sparse representation transform domain S with. in certain. f S x, the relation becomes

Survey for Image Representation Using Block Compressive Sensing For Compression Applications

Introduction to Inverse Problems

Image Transformation Techniques Dr. Rajeev Srivastava Dept. of Computer Engineering, ITBHU, Varanasi

Geometry-Driven Distributed Compression of the Plenoptic Function: Performance Bounds and Constructive Algorithms

Compressive Sensing Applications and Demonstrations: Synthetic Aperture Radar

An Intuitive Explanation of Fourier Theory

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

Reconstruction Improvements on Compressive Sensing

DIGITAL IMAGE PROCESSING WRITTEN REPORT ADAPTIVE IMAGE COMPRESSION TECHNIQUES FOR WIRELESS MULTIMEDIA APPLICATIONS

Transcription:

Measurements and Bits: Compressed Sensing meets Information Theory Dror Baron ECE Department Rice University dsp.rice.edu/cs

Sensing by Sampling Sample data at Nyquist rate Compress data using model (e.g., sparsity) encode coefficient locations and values Lots of work to throw away >80% of the coefficients Most computation at sensor (asymmetrical) Brick wall to performance of modern acquisition systems sample compress transmit/store sparse wavelet transform receive decompress

Sparsity / Compressibility Many signals are sparse or compressible in some representation/basis (Fourier, wavelets, ) pixels large wavelet coefficients wideband signal samples large Gabor coefficients

Compressed Sensing Shannon/Nyquist sampling theorem worst case bound for any bandlimited signal too pessimistic for some classes of signals does not exploit signal sparsity/compressibility Seek direct sensing of compressible information Compressed Sensing (CS) sparse signals can be recovered from a small number of nonadaptive (fixed) linear measurements [Candes et al.; Donoho; Kashin; Gluskin; Rice ] based on new uncertainty principles beyond Heisenberg ( incoherency )

Incoherent Bases (matrices) Spikes and sines (Fourier)

Incoherent Bases Spikes and random noise

Compressed Sensing via Random Projections Measure linear projections onto incoherent basis where data is not sparse/compressible random projections are universally incoherent fewer measurements no location information Reconstruct via optimization Highly asymmetrical (most computation at receiver) project transmit/store receive reconstruct

CS Encoding Replace samples by more general encoder based on a few linear projections (inner products) Matrix vector multiplication potentially analog measurements sparse signal # non-zeros

Universality via Random Projections Random projections Universally incoherent with any compressible/sparse signal class measurements sparse signal # non-zeros

Reconstruction Before-CS Goal: Given measurements find signal Fewer rows than columns in measurement matrix Ill-posed: infinitely many solutions Classical solution: least squares

Reconstruction Before-CS Goal: Given measurements find signal Fewer rows than columns in measurement matrix Ill-posed: infinitely many solutions Classical solution: least squares Problem: small L 2 doesn t imply sparsity

Ideal Solution Ideal solution: exploit sparsity of Of the infinitely many solutions seek sparsest one number of nonzero entries

Ideal Solution Ideal solution: exploit sparsity of Of the infinitely many solutions seek sparsest one If M K then w/ high probability this can t be done If M K+1 then perfect reconstruction w/ high probability [Bresler et al.; Wakin et al.] But not robust and combinatorial complexity

The CS Revelation Of the infinitely many solutions seek the one with smallest L 1 norm

The CS Revelation Of the infinitely many solutions seek the one with smallest L 1 norm If then perfect reconstruction w/ high probability [Candes et al.; Donoho] Robust to measurement noise Linear programming

CS Hallmarks CS changes the rules of data acquisition game exploits a priori signal sparsity information (signal is compressible) Hardware: Universality same random projections / hardware for any compressible signal class simplifies hardware and algorithm design Processing: Information scalability random projections ~ sufficient statistics same random projections for range of tasks reconstruction > estimation > recognition > detection far fewer measurements required to detect/recognize Next generation data acquisition new imaging devices and A/D converters [DARPA] new reconstruction algorithms new distributed source coding algorithms [Baron et al.]

Random Projections in Analog

Optical Computation of Random Projections CS encoder integrates sensing, compression, processing Example: new cameras and imaging algorithms

First Image Acquisition (M=0.38N) ideal 64x64 image (4096 pixels) 400 wavelets image on DMD array 1600 random meas.

A/D Conversion Below Nyquist Rate Modulator Filter Downsample Challenge: wideband signals (radar, communications, ) currently impossible to sample at Nyquist rate Proposed CS-based solution: sample at information rate simple hardware components good reconstruction performance

Connections Between Compressed Sensing and Information Theory

Measurement Reduction via CS CS reconstruction via If then perfect reconstruction w/ high probability [Candes et al.; Donoho] Linear programming Compressible signals (signal components decay) also requires polynomial complexity (BPDN) [Candes et al.] cannot reduce order of [Kashin,Gluskin]

Fundamental Goal: Minimize Compressed sensing aims to minimize resource consumption due to measurements Donoho: Why go to so much effort to acquire all the data when most of what we get will be thrown away?

Fundamental Goal: Minimize Compressed sensing aims to minimize resource consumption due to measurements Donoho: Why go to so much effort to acquire all the data when most of what we get will be thrown away? Recall sparse signals only measurements for reconstruction not robust and combinatorial complexity

Rich Design Space What performance metric to use? Determine support set of nonzero entries [Wainwright] this is distortion metric but why let tiny nonzero entries spoil the fun? metric??

Rich Design Space What performance metric to use? Determine support set of nonzero entries [Wainwright] this is distortion metric but why let tiny nonzero entries spoil the fun? metric?? What complexity class of reconstruction algorithms? any algorithms? polynomial complexity? near-linear or better?

Rich Design Space What performance metric to use? Determine support set of nonzero entries [Wainwright] this is distortion metric but why let tiny nonzero entries wreck spoil the fun? metric?? What complexity class of reconstruction algorithms? any algorithms? polynomial complexity? near-linear or better? How to account for imprecisions? noise in measurements? compressible signal model?

Lower Bound on Number of Measurements

Measurement Noise Measurement process is analog Analog systems add noise, non-linearities, etc. Assume Gaussian noise for ease of analysis

Setup Signal is iid Additive white Gaussian noise Noisy measurement process

Setup Signal is iid Additive white Gaussian noise Noisy measurement process Random projection of tiny coefficients (compressible signals) similar to measurement noise

Measurement and Reconstruction Quality Measurement signal to noise ratio Reconstruct using decoder mapping Reconstruction distortion metric Goal: minimize CS measurement rate

Measurement Channel Model process as measurement channel Capacity of measurement channel Measurements are bits!

Lower Bound [Sarvotham et al.] Theorem: For a sparse signal with rate-distortion function, lower bound on measurement rate subject to measurement quality and reconstruction distortion satisfies Direct relationship to rate-distortion content Applies to any linear signal acquisition system

Lower Bound [Sarvotham et al.] Theorem: For a sparse signal with rate-distortion function, lower bound on measurement rate subject to measurement quality and reconstruction distortion satisfies Proof sketch: each measurement provides bits information content of source bits source-channel separation for continuous amplitude sources minimal number of measurements obtain measurement rate via normalization by

Example Spike process - spikes of uniform amplitude Rate-distortion function Lower bound Numbers: signal of length 10 7 10 3 spikes SNR=10 db SNR=-20 db If interesting portion of signal has relatively small energy then need significantly more measurements! Upper bound (achievable) in progress

CS Reconstruction Meets Channel Coding

Why is Reconstruction Expensive? Culprit: dense, unstructured measurements sparse signal nonzero entries

Fast CS Reconstruction LDPC measurement matrix (sparse) Only 0/1 in Each row of contains randomly placed 1 s Fast matrix multiplication fast encoding fast reconstruction measurements sparse signal nonzero entries

Ongoing Work: CS Using BP Considering noisy CS signals Application of Belief Propagation BP over real number field sparsity is modeled as prior in graph States Q Coefficients X Measurements Y

Promising Results 80 60 40 x(j) 20 l2 norm of (x x_hat) vs M 0 20 40 0 200 400 600 j l2 reconstruction error 500 400 300 200 100 l=5 l=10 l=15 l=20 l=25 norm(x) norm(x_n) 0 500 1000 1500 2000 2500 3000 Number of measurements (M)

Theoretical Advantages of CS-BP Low complexity Provable reconstruction with noisy measurements using Success of LDPC+BP in channel coding carried over to CS!

Distributed Compressed Sensing (DCS) CS for distributed signal ensembles

Why Distributed? Networks of many sensor nodes sensor, microprocessor for computation, wireless communication, networking, battery can be spread over large geographical area Must be energy efficient minimize communication at expense of computation motivates distributed compression

Distributed Sensing destination raw data Transmitting raw data typically inefficient

Correlation? destination Can we exploit intra-sensor and inter-sensor correlation to jointly compress? Ongoing challenge in information theory (distributed source coding)

Collaborative Sensing destination compressed data Collaboration introduces inter-sensor communication overhead complexity at sensors

Distributed Compressed Sensing destination compressed data Exploit intra- and inter-sensor correlations with zero inter-sensor communication overhead low complexity at sensors Distributed source coding via CS

Model 1: Common + Innovations

Common + Innovations Model Motivation: measuring signals in smooth field average temperature value common at multiple locations innovations driven by wind, rain, clouds, etc. Joint sparsity model: length-n sequences x 1 and x 2 z C is length-n common component z 1, z 2 length-n innovations components z C, z 1, z 2 have sparsity K C, K 1, K 2 Measurements

Measurement Rate Region with Separate Reconstruction Encoder f 1 Decoder g 1 Encoder f 2 Decoder g 2 separate encoding & recon

Slepian-Wolf Theorem (Distributed lossless coding) Theorem: [Slepian and Wolf 1973] R 1 >H(X 1 X 2 ) R 2 >H(X 2 X 1 ) R 1 +R 2 >H(X 1,X 2 ) (conditional entropy) (conditional entropy) (joint entropy) R 2 separate encoding & separate recon H(X 2 ) H(X 2 X 1 ) H(X 1 X 2 ) H(X 1 ) Slepian-Wolf joint recon R 1

Measurement Rate Region with Joint Reconstruction Encoder f 1 Decoder g Encoder f 2 separate encoding & joint recon Inspired by Slepian-Wolf coding

Measurement Rate Region [Baron et al.] separate reconstruction simulation achievable converse

Multiple Sensors

Model 2: Common Sparse Supports

Common Sparse Supports Model Ex: Many audio signals sparse in Fourier Domain same frequencies received by each node different attenuations and delays (magnitudes and phases)

Common Sparse Supports Model Signals share sparse components but different coefficients Intuition: Each measurement vector holds clues about coefficient support set

Required Number of Measurements [Baron et al. 2005] Theorem: M=K measurements per sensor do not suffice to reconstruct signal ensemble Theorem: As number of sensors J increases, M=K+1 measurements suffice to reconstruct Joint reconstruction with reasonable computational complexity

K=5 N=50 Results for Common Sparse Supports Reconstruction Separate Joint

Real Data Example Light levels in Intel Berkeley Lab 49 sensors, 1024 samples each Compare: wavelet approx 100 terms per sensor separate CS 400 measurements per sensor joint CS (SOMP) 400 measurements per sensor Correlated signal ensemble

Light Intensity at Node 19

Model 3: Non-Sparse Common Component

Non-Sparse Common Model Motivation: non-sparse video frame + sparse motion Length-N common component z C is non-sparse Each signal is incompressible Innovation sequences z j may share supports sparse not sparse Intuition: each measurement vector contains clues about common component z C

K=5 N=50 Results for Non-Sparse Common (same supports) Impact of z C vanishes as J! 1

Summary Compressed Sensing random projections process sparse signals using far fewer measurements universality and information scalability Determination of measurement rates in CS measurements are bits lower bound on measurement rate direct relationship to rate-distortion content Promising results with LDPC measurement matrices Distributed CS new models for joint sparsity analogy with Slepian-Wolf coding from information theory compression of sources w/ intra- and inter-sensor correlation Much potential and much more to be done Compressed sensing meets information theory dsp.rice.edu/cs

THE END

With High Probability