A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

Similar documents
A FAST ALGORITHM FOR SPARSE RECONSTRUCTION BASED ON SHRINKAGE, SUBSPACE OPTIMIZATION AND CONTINUATION. January, 2009

Convex or non-convex: which is better?

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Structurally Random Matrices

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

The Benefit of Tree Sparsity in Accelerated MRI

Introduction to Optimization Problems and Methods

Characterizing Improving Directions Unconstrained Optimization

Bilevel Sparse Coding

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

Introduction to Topics in Machine Learning

REDUCED ORDER MODELING IN MULTISPECTRAL PHOTOACOUSTIC TOMOGRAPHY

Convexization in Markov Chain Monte Carlo

Collaborative Sparsity and Compressive MRI

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007

Sparse Reconstruction / Compressive Sensing

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments

Sparsity Based Regularization

Programming, numerics and optimization

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

ELEG Compressive Sensing and Sparse Signal Representations

Recent Developments in Model-based Derivative-free Optimization

Introduction to Optimization

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Image Reconstruction from Multiple Sparse Representations

The convex geometry of inverse problems

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Chapter 3 Numerical Methods

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Lagrangian Relaxation: An overview

Compressive Sensing MRI with Wavelet Tree Sparsity

Iterative Shrinkage/Thresholding g Algorithms: Some History and Recent Development

Compressed Sensing and L 1 -Related Minimization

Compressive Sensing Algorithms for Fast and Accurate Imaging

Algorithms for convex optimization

Iteratively Re-weighted Least Squares for Sums of Convex Functions

Algorithms for sparse X-ray CT image reconstruction of objects with known contour

A New Non-Convex Regularized Sparse Reconstruction Algorithm for Compressed Sensing Magnetic Resonance Image Recovery

Inter and Intra-Modal Deformable Registration:

Detection Performance of Radar Compressive Sensing in Noisy Environments

Iterative CT Reconstruction Using Curvelet-Based Regularization

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Learning with infinitely many features

Compressed Sensing for Electron Tomography

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

A Course in Machine Learning

Generalized trace ratio optimization and applications

Sparse Solutions to Linear Inverse Problems. Yuzhe Jin

16.410/413 Principles of Autonomy and Decision Making

Image Restoration and Background Separation Using Sparse Representation Framework

Optimization for Machine Learning

25. NLP algorithms. ˆ Overview. ˆ Local methods. ˆ Constrained optimization. ˆ Global methods. ˆ Black-box methods.

Convex Optimization MLSS 2015

Introduction. Wavelets, Curvelets [4], Surfacelets [5].

Image Reconstruction based on Block-based Compressive Sensing

Efficient MR Image Reconstruction for Compressed MR Imaging

Parallel and Distributed Sparse Optimization Algorithms

Parameter Estimation and Model Order Identification of LTI Systems

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

Primal and Dual Methods for Optimisation over the Non-dominated Set of a Multi-objective Programme and Computing the Nadir Point

Efficient Iterative Semi-supervised Classification on Manifold

Divide and Conquer Kernel Ridge Regression

Outlier Pursuit: Robust PCA and Collaborative Filtering

A Geometric Analysis of Subspace Clustering with Outliers

A primal-dual framework for mixtures of regularizers

Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY

The Alternating Direction Method of Multipliers

Reconstruction of CT Images from Sparse-View Polyenergetic Data Using Total Variation Minimization

SPARSITY ADAPTIVE MATCHING PURSUIT ALGORITHM FOR PRACTICAL COMPRESSED SENSING

Modified Iterative Method for Recovery of Sparse Multiple Measurement Problems

Lecture 9. Introduction to Numerical Techniques

INTRODUCTION TO LINEAR AND NONLINEAR PROGRAMMING

Compressive Sensing Applications and Demonstrations: Synthetic Aperture Radar

Learning Two-View Stereo Matching

Projection-Based Methods in Optimization

Computational Methods. Constrained Optimization

Variable Selection 6.783, Biomedical Decision Support

A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem

Compressed Sensing Algorithm for Real-Time Doppler Ultrasound Image Reconstruction

The Ascendance of the Dual Simplex Method: A Geometric View

Compressive Sensing based image processing in TrapView pest monitoring system

On the Global Solution of Linear Programs with Linear Complementarity Constraints

SVMs for Structured Output. Andrea Vedaldi

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Optimization for Machine Learning

Total Variation Denoising with Overlapping Group Sparsity

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

Lagrangean Methods bounding through penalty adjustment

Bilinear Programming

Non-Differentiable Image Manifolds

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

Augmented Lagrangian Methods

CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

Unsupervised Learning

3D Surface Recovery via Deterministic Annealing based Piecewise Linear Surface Fitting Algorithm

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Transcription:

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 1 / 14

Introduction Fundamental principal of Compressive Sensing K-sparse signal x R n can be recovered from relatively few incomplete measurements, b = Ax by solving a l 0 -minimization problem: min x R n x 0 s.t. Ax = b (1) Basis Pursuit min x x R n 1 s.t. Ax = b (2) which is proposed by Candes,Romberg,Tao, is more tractable. Under some reasonable conditions on x and A, the sparsest solution x of problem (1) can be found by solving (2). Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 2 / 14

Introduction Greedy algorithm for (1): Orthogonal Matching Pursuit Solve a sequence of subspace optimization problems of the form: min x A T x T b 2 2 s.t. x i = 0, i / T (3) Start from T =, x = 0, and in each iteration,add to T the index of the largest component of the current gradient of Ax b 2 2 Regularization for (2) min ψ µ(x) = µ x 1 + Ax b 2 x R n 2 (4) Theory of penalty functions implies that the solution to (4) goes to solution to (2) when µ goes to 0. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 3 / 14

Introduction Challenges for standard LP and QP Large-scale real-world applications A is usually dense real-time processing is required Special structure Measurement matrix A often corresponds to partial transform matrices(eg, discrete Fourier), so fast matrix-vector multiplication are available. Sparsity feature of the solutions. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 4 / 14

Outline Outline of the algorithm This paper proposed a two-stage algorithm for the regularized l 1 minimization problem, it combines the good features of both greedy algorithms and convex optimization approach: step1 convex optimization Do not require prior information. A first-order method based on shrinkage is applied. obtain an approximation solution and identify a working set T. step2 greedy algorithm takes advantage of the sparsity structure. A second-order method is applied solve a smooth subspace problem. Embed the two-stage algorithm into a continuation approach by assigning a decreasing sequence of values to µ. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 5 / 14

Shrinkage Iterative Shrinkage Algorithm For the problem Iterative Shrinkage algorithm: min φ(x) := 1 x R n 2 Ax y 2 2 + τc(x) (5) x k+1 = ψ τ/α (x k 1 α AT (Ax k y)) (6) where 1 ψ λ (u) := arg min z 2 z u 2 2 + λc(z) (7) Features Products by A and A T are efficiently computable. Ensure convergence under mild conditions. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 6 / 14

Shrinkage Derivation of Iterative Shrinkage Expectation-Maximization [Figueiredo, Nowak, 2001]. Majorization-Minimization [Daubechies,Defrise,DeMol, 2004]. Forward-Backward Splitting [Hale, Yin, Zhang, 2007]. Separable Approximation [Wright,Figueiredo,Nowak,2009]. Benefit of Iterative Shrinkage Iterative shrinkage yields the support and the signs of the optimal solution x of the problem (4) in a finite number of iterations, that is, k such that sgn(x k ) sgn(x ) for all k > k. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 7 / 14

Use shrinkage in step 1 Notation f(x) = Ax b 2 2 g(x) = f(x) x y: component-wise product of x and y Special shrinkage for c(x) = x 1 S(y, γ) sgn(y) max{ y γ, 0} (8) which is the unique minimizer of the function and we use the iteration: γ x 1 + 1 2 x y 2 2 (9) x k+1 S(x k γg k, µλ), λ > 0 (10) Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 8 / 14

Shrinkage Shrinkage phase of the algorithm Select a parameter λ k and compute a direction S(x k λ k g k, µ k λ k ) x k Do a line search on the direction d k,and set the new iteration point x k+1 = x k + α k d k The set of indices corresponding to 0 and nearly 0 components of x k (say, x k ξ k ) is selected as a working set,which during the step 2, the subspace problem is formed by fixing these components to be 0. Note We may not do a subspace optimization problem in every iteration, we only do that if certain conditions hold for the iteration point in shrinkage phase. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 9 / 14

Subspace optimization phase motivation The iterative shrinkage scheme essentially reduces to a gradient projection method for solving a subspace minimization problem after sufficiently many iterations. A second-order method might be faster than the iterative shrinkage to solve the subspace problem. subspace problem, definition The active set A(x) := {i {1,..., n} x i = 0} The inactive set I(x) := {i {1,..., n} x i > 0} When A(x k ) is a good estimate of the true active set, we approximate sgn(xi ) by sgn(xi k ) and replace the original φ µ (x) by the smooth function(suppose the approximation set are A k, I k ) ϕ µ (x) := µsgn(x k I k ) T x Ik + f(x) (11) Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 10 / 14

Subspace optimization phase Smooth subspace problem Now we come up with a nice subspace problem, we can use efficient and fast algorithm in NLP to solve this simply-constrained problem: min ϕ µ (x) s.t. x Ω(x k ) where Ω(x k ) := {x R n sgn(x k i )x i 0, i I k ; x i = 0, i A k } In the paper, they use a limited-memory quasi-newton method with simple bound constraints. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 11 / 14

Alternating strategy of the two phases Idea of stopping criteria for shrinkage phase We want to start subspace optimization as soon as possible we want the active set that defines the subspace optimization problem to be as accurate as possible In all, we want to make both phases work efficiently, if either of them does not, switch to the other Idea of identification of the active set The efficiency of the algorithm depends on how fast and how well the active set is identified. One approach is to replace the active set A(x k ) and the support I(x k ) by the sets: A(x k, ξ k ) := {i {1,..., n} x k i ξ k }, I(x k, ξ k ) := {i {1,..., n} x k i > ξ k } Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 12 / 14

Continuation strategy Idea objective: create a path of solutions that converge to the solution of the original problem A fixed fractional reduction of µ k should be enforced Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 13 / 14

Summary objective: solve a l 1 -regularized minimization problem approach: two-phases shrinkage phase: do not use prior information, partially use sparsity, take some iterations to obtain an estimate subset of the active set and the support. subspace phase: take advantage of the prior information, reconstruct the problem to be a smooth problem, use fast algorithms. Reasonable strategy to switch between the two phases. Yongjia Song (UW-Madison) Project for CS 838 April 22, 2010 14 / 14