Compressive Parameter Estimation with Earth Mover s Distance via K-means Clustering. Dian Mo and Marco F. Duarte

Similar documents
Parameter Estimation in Compressive Sensing. Marco F. Duarte

Compressive Parameter Estimation with Earth Mover s Distance via K-Median Clustering

Compressive Sensing. Part IV: Beyond Sparsity. Mark A. Davenport. Stanford University Department of Statistics

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

The Fundamentals of Compressive Sensing

Compressive. Graphical Models. Volkan Cevher. Rice University ELEC 633 / STAT 631 Class

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

ELEG Compressive Sensing and Sparse Signal Representations

Modified Iterative Method for Recovery of Sparse Multiple Measurement Problems

Compressive Sensing: Theory and Practice

Detection Performance of Radar Compressive Sensing in Noisy Environments

Compressive Topology Identification of Interconnected Dynamic Systems via Clustered Orthogonal Matching Pursuit

Image Reconstruction from Multiple Sparse Representations

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

Compressive Sensing Applications and Demonstrations: Synthetic Aperture Radar

Coherence Pattern Guided Compressive Sensing with Unresolved Grids

Sparse Solutions to Linear Inverse Problems. Yuzhe Jin

Coherence-Pattern-Guided Compressive Sensing with Unresolved Grids

Randomized sampling strategies

Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm

Compressive Sensing for High-Dimensional Data

Progress In Electromagnetics Research, Vol. 125, , 2012

SPARSE signal representations provide a general signal

Compressive Sensing Based Image Reconstruction using Wavelet Transform

Compressed Sensing and Applications by using Dictionaries in Image Processing

Structurally Random Matrices

Main Menu. Summary. sampled) f has a sparse representation transform domain S with. in certain. f S x, the relation becomes

The Viterbi Algorithm for Subset Selection

Short-course Compressive Sensing of Videos

Non-Differentiable Image Manifolds

IMAGE MASKING SCHEMES FOR LOCAL MANIFOLD LEARNING METHODS. Hamid Dadkhahi and Marco F. Duarte

COMPRESSIVE VIDEO SAMPLING

Compressive Sensing based image processing in TrapView pest monitoring system

Measurements and Bits: Compressed Sensing meets Information Theory. Dror Baron ECE Department Rice University dsp.rice.edu/cs

Randomized Dimensionality Reduction

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Learning based face hallucination techniques: A survey

Sparse Reconstruction / Compressive Sensing

SUMMARY. In combination with compressive sensing, a successful reconstruction

SPARSITY ADAPTIVE MATCHING PURSUIT ALGORITHM FOR PRACTICAL COMPRESSED SENSING

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

Super-Resolution by Compressive Sensing Algorithms

Incoherent noise suppression with curvelet-domain sparsity Vishal Kumar, EOS-UBC and Felix J. Herrmann, EOS-UBC

Dimensionality reduction of MALDI Imaging datasets using non-linear redundant wavelet transform-based representations

Introduction to Topics in Machine Learning

Compressive Sensing. and Applications. Volkan Cevher Justin Romberg

A Relationship between the Robust Statistics Theory and Sparse Compressive Sensed Signals Reconstruction

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Signal and Image Recovery from Random Linear Measurements in Compressive Sampling

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

A Nearly-Linear Time Framework for Graph-Structured Sparsity

Lecture 17 Sparse Convex Optimization

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

Compressive Sensing. A New Framework for Sparse Signal Acquisition and Processing. Richard Baraniuk. Rice University

Compressive Sensing Reconstruction Algorithm using L1-norm Minimization via L2-norm Minimization

arxiv: v1 [stat.ml] 14 Jan 2017

The Benefit of Tree Sparsity in Accelerated MRI

RECONSTRUCTION ALGORITHMS FOR COMPRESSIVE VIDEO SENSING USING BASIS PURSUIT

Invariant shape similarity. Invariant shape similarity. Invariant similarity. Equivalence. Equivalence. Equivalence. Equal SIMILARITY TRANSFORMATION

Open Access Reconstruction Technique Based on the Theory of Compressed Sensing Satellite Images

Machine Learning Feature Creation and Selection

Locally Weighted Learning for Control. Alexander Skoglund Machine Learning Course AASS, June 2005

Predict Outcomes and Reveal Relationships in Categorical Data

The convex geometry of inverse problems

Compressive Sensing: Opportunities and pitfalls for Computer Vision

Sparsity Based Regularization

CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT

Multiple-View Object Recognition in Band-Limited Distributed Camera Networks

Compressive Sensing: Opportunities and Perils for Computer Vision

Sparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

Earth Mover s Distance and The Applications

Collaborative Sparsity and Compressive MRI

ECS 234: Data Analysis: Clustering ECS 234

SPARSE SIGNAL RECONSTRUCTION FROM NOISY COMPRESSIVE MEASUREMENTS USING CROSS VALIDATION. Petros Boufounos, Marco F. Duarte, Richard G.

Adaptive Sparse Recovery by Parametric Weighted L Minimization for ISAR Imaging of Uniformly Rotating Targets

A* Orthogonal Matching Pursuit: Best-First Search for Compressed Sensing Signal Recovery

Deconvolution with curvelet-domain sparsity Vishal Kumar, EOS-UBC and Felix J. Herrmann, EOS-UBC

Distributed Compressed Estimation Based on Compressive Sensing for Wireless Sensor Networks

Image reconstruction based on back propagation learning in Compressed Sensing theory

CoSaMP: Iterative Signal Recovery from Incomplete and Inaccurate Samples By Deanna Needell and Joel A. Tropp

Document Clustering: Comparison of Similarity Measures

Sparse Signal Reconstruction using Weight Point Algorithm

Clustered Compressive Sensing: Application on Medical Imaging

AEDL Algorithm for Change Detection in Medical Images - An Application of Adaptive Dictionary Learning Techniques

Histograms of Sparse Codes for Object Detection

Survey of the Mathematics of Big Data

The Curse of Dimensionality

Iterative CT Reconstruction Using Curvelet-Based Regularization

Distance-based Methods: Drawbacks

G Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing

4692 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 56, NO. 10, OCTOBER X/$ IEEE

1 Introduction Motivation and Aims Functional Imaging Computational Neuroanatomy... 12

KSVD - Gradient Descent Method For Compressive Sensing Optimization

Image Analysis & Retrieval. CS/EE 5590 Special Topics (Class Ids: 44873, 44874) Fall 2016, M/W Lec 18.

Survey for Image Representation Using Block Compressive Sensing For Compression Applications

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 59, NO. 6, JUNE

Reweighted thresholding and orthogonal projections for simultaneous source separation Satyakee Sen*, Zhaojun Liu, James Sheng, and Bin Wang, TGS

Image Reconstruction based on Block-based Compressive Sensing

Transcription:

Compressive Parameter Estimation with Earth Mover s Distance via K-means Clustering Dian Mo and Marco F. Duarte

Compressive Sensing (CS) Integrates linear acquisition with dimensionality reduction linear measurements discrete-time signal nonzero components in suitable transform [Candès, Romberg, Tao; Donoho]

Parametric Dictionaries for Sparsity Integrates sparsity/cs with parameter estimation Parametric dictionaries (PDs) collect observations for a set of values of parameter of interest (one per column) Simple signals (e.g., few localization targets) can be expressed via PDs using sparse coefficient vectors [Gorodntisky and Rao 1997] [Malioutov, Cetin, Willsky 2005] [Cevher, Duarte, Baraniuk 2008] [Cevher, Gurbuz, McClellan, Chellapa 2008][...]

Issues with Parametric Dictionaries As parameter resolution (e.g., number of grid points) increases, PD becomes increasingly coherent, hampering sparse approximation algorithms PD s high coherence is a manifestation of resolution issues in underlying estimation problem Structured sparsity models can mitigate this issues by preventing PD elements with coherence above target maximum from appearing simultaneously in recovered signal 1 0.8 0.6 0.4 0.2 1 0.8 0.6 0.4 0.2 0 100 200 300 400, meters Near-Field Localization 0 20 40 60 80, degrees Far-Field Localization [Duarte and Baraniuk 2010] [Duarte, 2012]

Standard Sparse Signal Recovery Iterative Hard Thresholding Inputs: Measurement vector y Measurement matrix Sparsity K Output: PD coefficient estimate Initialize: While halting criterion false, (estimate signal) (obtain best sparse approx.) (calculate residual) Return estimate [Blumensath and Davies 2009]

Structured Sparse Signal Recovery Structured Iterative Hard Thresholding (SIHT) Inputs: Measurement vector y Measurement matrix Structured sparse approx. algorithm Output: PD coefficient estimate Initialize: While halting criterion false, (estimate signal) (obtain best structured sparse approx.) (calculate residual) Return estimate Can be applied to a variety of greedy algorithms (CoSaMP, OMP, Subspace Pursuit, etc.) [Baraniuk, Cevher, Duarte, Hegde 2009]

Issues with Parametric Dictionaries Structured sparsity models need careful control of maximal coherence parameter Correlation function provides measure of coherence between PD elements Correlation function connects parameter resolution to maximal coherence value In most cases, coherence control equals band exclusion 1 0.8 0.6 0.4 0.2 0 100 200 300 400, meters Near-Field Localization 1 0.8 0.6 0.4 0.2 0 20 40 60 80, degrees Far-Field Localization [Duarte and Baraniuk, 2010] [Duarte, 2012] [Fannjiang and Liao, 2012]

Issues with Parametric Dictionaries Structured sparsity models need careful control of maximal coherence parameter Correlation function provides measure of coherence between PD elements Correlation function connects parameter resolution to maximal coherence value In most cases, coherence control equals band exclusion f0 Line Spectral Estimation f [Duarte and Baraniuk, 2010] [Duarte, 2012] [Fannjiang and Liao, 2012]

Average Parameter Estimation [ µs] Issues with PDs/Structured Sparsity: Sensitivity to Maximal Coherence Value 10 1 10 0 10 1 Parameter T =1 µs T =2 µs T =3 µs T =4 µs T =5 µs 0.2 0.4 0.6 0.8 1 Example: Compressive Time Delay Estimation (TDE) with PD and CS Performance depends on measurement ratio Structured sparsity used to enable highresolution TDE set to optimal value for chirp of length As length of chirp wave increases, performance of compressive TDE varies widely Shape of correlation function dependent on chirp length

Issues with PDs: Euclidean Norm Guarantees Most recovery methods provide guarantees to keep Euclidean norm error small This metric, however, is not connected to quality of parameter estimates Example: both estimates have same Euclidean error localization grid but provide very different location estimates. We search for a performance metric better suited to the use of PD coefficient vectors (i.e., )

Improved Performance Metric: Earth Mover s Distance Earth Mover s Distance (EMD) is based on concept of mass flowing between the entries of first vector in order to match the second EMD value is minimum work needed (measured as mass x transport distance) for first vector to match second: s.t.

Improved Performance Metric: Earth Mover s Distance When PDs are used, EMD captures parameter estimation error by measuring distance traveled by mass Parameter values must be proportional to indices in PD coefficient vector How to introduce EMD metric into CS recovery process? s.t.

Sparse Approximation with Earth Mover s Distance To integrate into greedy algorithms, we will need to solve the EMD-optimal K-sparse approximation problem It can be shown that approximation can be obtained by performing K-median clustering on set of points at locations with respective weights Cluster centroids provide support of, values can be easily computed to minimize EMD/estimation error [Indyk and Price 2009]

More Intuition Behind EMD Sparse Approximation Greedy algorithms generally compute a signal proxy /residual of the form For sufficiently large number of measurements, signal proxy will resemble correlation function convolved with original sparse coefficient vector EMD sparse approximation (K-median clustering) converts each correlation function into single cluster with centroid at correlation peak Small estimation biases may appear if translated correlation function is asymmetric or noise is present f0 f

Structured Sparse Signal Recovery Band-Excluding IHT Inputs: Measurement vector y Measurement matrix Structured sparse approx. algorithm Output: PD coefficient estimate Initialize: While halting criterion false, Can be applied to a variety of greedy algorithms (CoSaMP, OMP, Subspace Pursuit, etc.) (estimate signal) (obtain band-excluding sparse approx.) (calculate residual) Return estimate [Baraniuk, Cevher, Duarte, Hegde 2009] [Fannjiang and Liao, 2012]

EMD + Sparse Signal Recovery Clustered IHT Inputs: Measurement vector y Measurement matrix Sparsity K Output: PD coefficient estimate Initialize: While halting criterion false, (estimate signal) (best sparse approx. in EMD) (calculate residual) Return estimate Can be applied to a variety of greedy algorithms (CoSaMP, OMP, Subspace Pursuit, etc.)

Average Parameter Estimation [ µs] 10 1 10 0 10 1 Numerical Results T =1 µs T =2 µs T =3 µs T =4 µs T =5 µs 0.2 0.4 0.6 0.8 1 Band-Excluding Subspace Pursuit 0.2 0.4 0.6 0.8 1 Example: Compressive TDE with PD & random projections Performance depends on measurement ratio TDE performance varies widely as chirp length increases Consistent behavior for EMD-based signal recovery, but consistent bias observed Bias partially due to parameter space discretization Average Parameter Estimation [ µs] 10 1 10 0 10 1 T =1 µs T =2 µs T =3 µs T =4 µs T =5 µs Clustered Subspace Pursuit

Another PD Issue: Discretization Every PD can be conceived as a sampling from an infinite set of parametrizable signals for a discrete set of parameter values When signal vector varies smoothly as a function of, signal set can be represented by nonlinear manifold If manifold is well behaved, resolution can be improved by interpolating between PD samples... Parameter space [Fyhn, Dadkhahi, Duarte 2013] [Fyhn, Jensen, Duarte 2013]

Numerical Results Average Parameter Estimation [ µs] 10 0 10 5 10 10 10 15 10 20 T =1 µs T =2 µs T =3 µs T =4 µs T =5 µs 0.2 0.4 0.6 0.8 1 Band-Excluding Subspace Pursuit Average Parameter Estimation [ µs] 10 0 10 5 10 10 10 15 10 20 T =1 µs T =2 µs T =3 µs T =4 µs T =5 µs 0.2 0.4 0.6 0.8 1 Clustered Subspace Pursuit Example: Compressive TDE with PD and CS Performance depends on measurement ratio When integrated with polar interpolation, performance of compressive TDE improves significantly Sensitivity of Band-Excluding SP becomes more severe, while Clustered SP remains robust

Conclusions Retrofitting sparsity via parametric dictionaries is not enough! PDs enable use of CS, but often are coherent band exclusion can help, but must be highly precise issues remain with guarantees (Euclidean is not useful) PDs also discretize parameter space, limiting resolution Earth Mover s Distance is a suitable alternative easily implementable by leveraging K-median clustering EMD is suitable for dictionaries with well-behaved (compact) correlation functions from PDs to manifolds via interpolation techniques ongoing work: theoretical guarantees, bias issues, sensitivity to noise... localization, bearing estimation, radar imaging,... http://www.ecs.umass.edu/~mduarte mduarte@ecs.umass.edu