Convex or non-convex: which is better?

Size: px
Start display at page:

Download "Convex or non-convex: which is better?"

Transcription

1 Sparsity Amplified Ivan Selesnick Electrical and Computer Engineering Tandon School of Engineering New York University Brooklyn, New York March 7 / 4

2 Convex or non-convex: which is better? (for sparse-regularized linear inverse problems) Benefits of convex optimization:. Absence of suboptimal local minima. Continuity of solution as a function of input data 3. Algorithms guaranteed to converge to a global optimum 4. Regularization parameters easier to set But convex regularization tends to under-estimate signal values. Non-convex regularization often performs better! Can we design non-convex sparsity-inducing penalties that maintain the convexity of the cost function to be minimized? / 4

3 An alternative Conventional sparse-regularized least squares: J(x) = y Ax + λ x Alternatively, use non-convex penalty ψ that maintains convexity of cost function F F (x) = y Ax + λψ(x) I. Selesnick. Total Variation Denoising via the Moreau Envelope. IEEE Signal Processing Letters, 4():6, February 7. I. W. Selesnick and I. Bayram. Enhanced Sparsity by Non-Separable Regularization. IEEE Trans. Signal Process., 64(9):98 33, May 6. A. Parekh and I. W. Selesnick. Enhanced low-rank matrix approximation. IEEE Signal Processing Letters, 3(4): , April 6. Y. Ding and I. W. Selesnick. Artifact-Free Wavelet Denoising: Non-convex Sparse Regularization, Convex Optimization. IEEE Signal Processing Letters, (9): , September 5. I. W. Selesnick, A. Parekh, and I. Bayram. Convex -D Total Variation Denoising with Non-convex Regularization. IEEE Signal Processing Letters, ():4 44, February 5. 3 / 4

4 MC penalty The minimax-concave (MC) penalty φ: R R is defined as x φ(x) := x, x, x 3 x φ(x) 3 3 x C.-H. Zhang. Nearly unbiased variable selection under minimax concave penalty. Statistics:894 94,. The Annals of 4 / 4

5 Huber function The MC penalty can be expressed as φ(x) = x s(x) where s is the Huber function x The Huber function. 5 / 4

6 Huber function The Huber function s : R R is defined as s(x) := x, x x, x. The Huber function can be written as s(x) = min v R { v + (x v) }, equivalently s = ( ) where denotes infimal convolution. H. H. Bauschke and P. L. Combettes. Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer,. 6 / 4

7 Huber function 3.5 x x +.5 x x The Huber function as the pointwise minimum of three functions. 7 / 4

8 Scaled functions Let b R. The scaled Huber function s b : R R is defined as s b (x) := s(b x)/b, b. For b =, the function is defined as s (x) :=. The scaled MC penalty function φ b : R R is defined as φ b (x) := x s b (x) where s b is the scaled Huber function. 8 / 4

9 Scaled functions Scaled Huber function and MC penalty for several values of the scaling parameter. 3 Scaled Huber function s b (x) b =.4 b =. b = x 3 Scaled MC penalty φ b (x) b =. b =.7 b =. b = x 9 / 4

10 Convexity condition Let λ > and a R. Define f : R R, f (x) = (y ax) + λ φ b (x) where φ b is the scaled MC penalty. If b a /λ, then. f is convex. the minimizer is given by firm thresholding, x opt = firm(y/a; λ/a, /b ). / 4

11 Firm threshold function firm(y; λ, µ) 4 Firm threshold function 3 λ µ y / 4

12 Aims The minimizer of the scalar function f is easily obtained via firm thresholding. However, the situation in the multivariate case is more complicated. To generalize this process to the multivariate case, we aim to:. define a multivariate MC penalty (non-convex),. define a sparse-regularized least squares cost function (convex), 3. generalize the convexity condition, 4. provide a method to calculate a minimizer. / 4

13 Generalized Huber function Let B R M N. We define the generalized Huber function S B : R N R as { S B (x) := min v + v R N B(x } v). In the notation of infimal convolution, we have S B = B. The generalized Huber function satisfies S B (x) x, x R N. 3 / 4

14 Generalized Huber function Generalized Huber function S B (x) 3 B = x x Contours of S B 4 / 4

15 Generalized Huber function Generalized Huber function S B (x) B = [.5] x x Contours of S B 5 / 4

16 Generalized MC penalty Let B R M N. We define the generalized MC (GMC) penalty function ψ B : R N R as ψ B (x) := x S B (x) where S B is the generalized Huber function. The generalized MC penalty satisfies ψ B (x) x for all x R N. 6 / 4

17 Generalized MC penalty x S B (x) ψ B (x) = x S B (x) x x x x Contours Contours Contours 7 / 4

18 Generalized MC penalty x S B (x) ψ B (x) = x S B (x) x x x x Contours Contours Contours 8 / 4

19 Convexity condition Let y R M, A R M N, and λ >. Define F : R N R as F (x) = y Ax + λ ψ B(x) where ψ B is the generalized MC penalty. If B T B λ AT A, then F is a convex function. Hence, for convexity of F, we may simply set B = γ/λ A, γ. () 9 / 4

20 Proximal algorithm Using forward-backward splitting, we have: Let λ > and γ <. Let y R N and A R M N. Then a saddle-point (x opt, v opt ) of F can be obtained by the iterative algorithm: Set ρ = max{, γ/( γ)} A T A Set µ : < µ < /ρ For i =,,,... w (i) = x (i) µa T( A(x (i) + γ(v (i) x (i) )) y ) u (i) = v (i) µγa T A(v (i) x (i) ) x (i+) = soft(w (i), µλ) v (i+) = soft(u (i), µλ) end where i is the iteration counter. / 4

21 Example: frequency-domain sparsity Noise free signal Noisy signal, σ = Denoising via frequency domain sparsity Denoising [L norm] 4 4 RMSE =.43 5 Denoising [GMC penalty] 4 4 RMSE =.49 5 Average RMSE L norm L + debias IPS MUSR GMC 6 4 FFT of noisy signal 5 Optimized Fourier coefficients o GMC x L norm λ Frequency Frequency / 4

22 Example: time-frequency domain sparsity.3 True signal Time frequency representation (db) Frequency (khz) Time (ms) Noisy signal Frequency (khz) 6 4 Time (ms) Time frequency representation (db).3.3 RMSE =.6 Time (ms) Denoising using L norm Frequency (khz) 6 4 Time (ms) Time frequency representation (db).3.3 RMSE =.6 Time (ms) Denoising using GMC penalty Frequency (khz) 6 4 Time (ms) Time frequency representation (db).3 Time (ms) Time (ms) / 4

23 Example: sparsity-assisted signal smoothing (SASS) Biosensor data SASS using L norm (a) SASS using GMC penalty 5 5 (b) Magnified view data L norm GMC (c) (d) 3 / 4

24 Conclusion Convex and the non-convex methods for sparse-regularized least squares are usually mutually exclusive and incompatible. To bridge these two approaches, we introduce a non-convex alternative to the L norm that preserves the convexity of the cost function to be minimized. The proposed penalty leads to optimization problems with no extraneous suboptimal local minima and allows the use of globally convergent, computationally efficient, scalable convex optimization algorithms. The advantages compared to L norm regularization are (i) more accurate estimation of high-amplitude components of sparse solutions, (ii) a higher level of sparsity in a sparse approximation problem. 4 / 4

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song

More information

Sparsity and image processing

Sparsity and image processing Sparsity and image processing Aurélie Boisbunon INRIA-SAM, AYIN March 6, Why sparsity? Main advantages Dimensionality reduction Fast computation Better interpretability Image processing pattern recognition

More information

Total Variation Denoising with Overlapping Group Sparsity

Total Variation Denoising with Overlapping Group Sparsity 1 Total Variation Denoising with Overlapping Group Sparsity Ivan W. Selesnick and Po-Yu Chen Polytechnic Institute of New York University Brooklyn, New York selesi@poly.edu 2 Abstract This paper describes

More information

Collaborative Sparsity and Compressive MRI

Collaborative Sparsity and Compressive MRI Modeling and Computation Seminar February 14, 2013 Table of Contents 1 T2 Estimation 2 Undersampling in MRI 3 Compressed Sensing 4 Model-Based Approach 5 From L1 to L0 6 Spatially Adaptive Sparsity MRI

More information

Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging

Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging Paper ID: 1195 No Institute Given 1 More Experiment Results 1.1 Visual Comparisons We apply all methods on four 2D

More information

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,

More information

Convex optimization algorithms for sparse and low-rank representations

Convex optimization algorithms for sparse and low-rank representations Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,

More information

Iterative CT Reconstruction Using Curvelet-Based Regularization

Iterative CT Reconstruction Using Curvelet-Based Regularization Iterative CT Reconstruction Using Curvelet-Based Regularization Haibo Wu 1,2, Andreas Maier 1, Joachim Hornegger 1,2 1 Pattern Recognition Lab (LME), Department of Computer Science, 2 Graduate School in

More information

ELEG Compressive Sensing and Sparse Signal Representations

ELEG Compressive Sensing and Sparse Signal Representations ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /

More information

G Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing

G Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing G16.4428 Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine Compressed Sensing Ricardo Otazo, PhD ricardo.otazo@nyumc.org Compressed

More information

L1 REGULARIZED STAP ALGORITHM WITH A GENERALIZED SIDELOBE CANCELER ARCHITECTURE FOR AIRBORNE RADAR

L1 REGULARIZED STAP ALGORITHM WITH A GENERALIZED SIDELOBE CANCELER ARCHITECTURE FOR AIRBORNE RADAR L1 REGULARIZED STAP ALGORITHM WITH A GENERALIZED SIDELOBE CANCELER ARCHITECTURE FOR AIRBORNE RADAR Zhaocheng Yang, Rodrigo C. de Lamare and Xiang Li Communications Research Group Department of Electronics

More information

Denoising Using Projection Onto Convex Sets (POCS) Based Framework

Denoising Using Projection Onto Convex Sets (POCS) Based Framework Denoising Using Projection Onto Convex Sets (POCS) Based Framework Mohammad Tofighi, Kivanc Kose, A. Enis Cetin Dept. of Electrical and Electronic Engineering, Bilkent University, Ankara, Turkey Dermatology

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

Polynomial Smoothing of Time Series with Additive Step Discontinuities

Polynomial Smoothing of Time Series with Additive Step Discontinuities JUNE 5, Polynomial Smoothing of Time Series with Additive Step Discontinuities Ivan W. Selesnick, Stephen Arnold, and Venkata R. Dantham Abstract This paper addresses the problem of estimating simultaneously

More information

Divide and Conquer Kernel Ridge Regression

Divide and Conquer Kernel Ridge Regression Divide and Conquer Kernel Ridge Regression Yuchen Zhang John Duchi Martin Wainwright University of California, Berkeley COLT 2013 Yuchen Zhang (UC Berkeley) Divide and Conquer KRR COLT 2013 1 / 15 Problem

More information

Bandwidth Selection for Kernel Density Estimation Using Total Variation with Fourier Domain Constraints

Bandwidth Selection for Kernel Density Estimation Using Total Variation with Fourier Domain Constraints IEEE SIGNAL PROCESSING LETTERS 1 Bandwidth Selection for Kernel Density Estimation Using Total Variation with Fourier Domain Constraints Alexander Suhre, Orhan Arikan, Member, IEEE, and A. Enis Cetin,

More information

Iteratively Re-weighted Least Squares for Sums of Convex Functions

Iteratively Re-weighted Least Squares for Sums of Convex Functions Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He

More information

Recent Developments in Model-based Derivative-free Optimization

Recent Developments in Model-based Derivative-free Optimization Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:

More information

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

Signal Reconstruction from Sparse Representations: An Introdu. Sensing Signal Reconstruction from Sparse Representations: An Introduction to Compressed Sensing December 18, 2009 Digital Data Acquisition Suppose we want to acquire some real world signal digitally. Applications

More information

Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm

Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm Manya V. Afonso, José M. Bioucas-Dias, Mario A. T. Figueiredo To cite this version: Manya V. Afonso, José M. Bioucas-Dias,

More information

Blind Compressed Sensing Using Sparsifying Transforms

Blind Compressed Sensing Using Sparsifying Transforms Blind Compressed Sensing Using Sparsifying Transforms Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laboratory University of Illinois

More information

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff Compressed Sensing David L Donoho Presented by: Nitesh Shroff University of Maryland Outline 1 Introduction Compressed Sensing 2 Problem Formulation Sparse Signal Problem Statement 3 Proposed Solution

More information

A primal-dual framework for mixtures of regularizers

A primal-dual framework for mixtures of regularizers A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland

More information

Optimal Sampling Geometries for TV-Norm Reconstruction of fmri Data

Optimal Sampling Geometries for TV-Norm Reconstruction of fmri Data Optimal Sampling Geometries for TV-Norm Reconstruction of fmri Data Oliver M. Jeromin, Student Member, IEEE, Vince D. Calhoun, Senior Member, IEEE, and Marios S. Pattichis, Senior Member, IEEE Abstract

More information

Picasso: A Sparse Learning Library for High Dimensional Data Analysis in R and Python

Picasso: A Sparse Learning Library for High Dimensional Data Analysis in R and Python Picasso: A Sparse Learning Library for High Dimensional Data Analysis in R and Python J. Ge, X. Li, H. Jiang, H. Liu, T. Zhang, M. Wang and T. Zhao Abstract We describe a new library named picasso, which

More information

The Benefit of Tree Sparsity in Accelerated MRI

The Benefit of Tree Sparsity in Accelerated MRI The Benefit of Tree Sparsity in Accelerated MRI Chen Chen and Junzhou Huang Department of Computer Science and Engineering, The University of Texas at Arlington, TX, USA 76019 Abstract. The wavelet coefficients

More information

Image Denoising Using Sparse Representations

Image Denoising Using Sparse Representations Image Denoising Using Sparse Representations SeyyedMajid Valiollahzadeh 1,,HamedFirouzi 1, Massoud Babaie-Zadeh 1, and Christian Jutten 2 1 Department of Electrical Engineering, Sharif University of Technology,

More information

Adaptive Reconstruction Methods for Low-Dose Computed Tomography

Adaptive Reconstruction Methods for Low-Dose Computed Tomography Adaptive Reconstruction Methods for Low-Dose Computed Tomography Joseph Shtok Ph.D. supervisors: Prof. Michael Elad, Dr. Michael Zibulevsky. Technion IIT, Israel, 011 Ph.D. Talk, Apr. 01 Contents of this

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex

More information

Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs

Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs Benjamin Ricaud Signal Processing Lab 2, EPFL Lausanne, Switzerland SPLab, Brno, 10/2013 B. Ricaud

More information

A Relationship between the Robust Statistics Theory and Sparse Compressive Sensed Signals Reconstruction

A Relationship between the Robust Statistics Theory and Sparse Compressive Sensed Signals Reconstruction THIS PAPER IS A POSTPRINT OF A PAPER SUBMITTED TO AND ACCEPTED FOR PUBLICATION IN IET SIGNAL PROCESSING AND IS SUBJECT TO INSTITUTION OF ENGINEERING AND TECHNOLOGY COPYRIGHT. THE COPY OF RECORD IS AVAILABLE

More information

Convexization in Markov Chain Monte Carlo

Convexization in Markov Chain Monte Carlo in Markov Chain Monte Carlo 1 IBM T. J. Watson Yorktown Heights, NY 2 Department of Aerospace Engineering Technion, Israel August 23, 2011 Problem Statement MCMC processes in general are governed by non

More information

COMPRESSIVE VIDEO SAMPLING

COMPRESSIVE VIDEO SAMPLING COMPRESSIVE VIDEO SAMPLING Vladimir Stanković and Lina Stanković Dept of Electronic and Electrical Engineering University of Strathclyde, Glasgow, UK phone: +44-141-548-2679 email: {vladimir,lina}.stankovic@eee.strath.ac.uk

More information

Compressive Sensing Algorithms for Fast and Accurate Imaging

Compressive Sensing Algorithms for Fast and Accurate Imaging Compressive Sensing Algorithms for Fast and Accurate Imaging Wotao Yin Department of Computational and Applied Mathematics, Rice University SCIMM 10 ASU, Tempe, AZ Acknowledgements: results come in part

More information

Learning with infinitely many features

Learning with infinitely many features Learning with infinitely many features R. Flamary, Joint work with A. Rakotomamonjy F. Yger, M. Volpi, M. Dalla Mura, D. Tuia Laboratoire Lagrange, Université de Nice Sophia Antipolis December 2012 Example

More information

Curvelet-based non-linear adaptive subtraction with sparseness constraints. Felix J Herrmann, Peyman P Moghaddam (EOS-UBC)

Curvelet-based non-linear adaptive subtraction with sparseness constraints. Felix J Herrmann, Peyman P Moghaddam (EOS-UBC) Curvelet-based non-linear adaptive subtraction with sparseness constraints Felix J Herrmann, Peyman P Moghaddam (EOS-UBC) Context Spiky/minimal structure deconvolution (Claerbout, Ulrych, Oldenburg, Sacchi,

More information

SUMMARY THEORY. L q Norm Reflectivity Inversion

SUMMARY THEORY. L q Norm Reflectivity Inversion Optimal L q Norm Regularization for Sparse Reflectivity Inversion Fangyu Li, University of Oklahoma & University of Georgia; Rui Xie, University of Georgia; WenZhan Song, University of Georgia & Intelligent

More information

ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A.

ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A. ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS Donghwan Kim and Jeffrey A. Fessler University of Michigan Dept. of Electrical Engineering and Computer Science

More information

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING Fei Yang 1 Hong Jiang 2 Zuowei Shen 3 Wei Deng 4 Dimitris Metaxas 1 1 Rutgers University 2 Bell Labs 3 National University

More information

Generalized Tree-Based Wavelet Transform and Applications to Patch-Based Image Processing

Generalized Tree-Based Wavelet Transform and Applications to Patch-Based Image Processing Generalized Tree-Based Wavelet Transform and * Michael Elad The Computer Science Department The Technion Israel Institute of technology Haifa 32000, Israel *Joint work with A Seminar in the Hebrew University

More information

Efficient MR Image Reconstruction for Compressed MR Imaging

Efficient MR Image Reconstruction for Compressed MR Imaging Efficient MR Image Reconstruction for Compressed MR Imaging Junzhou Huang, Shaoting Zhang, and Dimitris Metaxas Division of Computer and Information Sciences, Rutgers University, NJ, USA 08854 Abstract.

More information

Kernels and representation

Kernels and representation Kernels and representation Corso di AA, anno 2017/18, Padova Fabio Aiolli 20 Dicembre 2017 Fabio Aiolli Kernels and representation 20 Dicembre 2017 1 / 19 (Hierarchical) Representation Learning Hierarchical

More information

Lecture 25 Nonlinear Programming. November 9, 2009

Lecture 25 Nonlinear Programming. November 9, 2009 Nonlinear Programming November 9, 2009 Outline Nonlinear Programming Another example of NLP problem What makes these problems complex Scalar Function Unconstrained Problem Local and global optima: definition,

More information

IMA Preprint Series # 2211

IMA Preprint Series # 2211 LEARNING TO SENSE SPARSE SIGNALS: SIMULTANEOUS SENSING MATRIX AND SPARSIFYING DICTIONARY OPTIMIZATION By Julio Martin Duarte-Carvajalino and Guillermo Sapiro IMA Preprint Series # 2211 ( May 2008 ) INSTITUTE

More information

Blur Space Iterative De-blurring

Blur Space Iterative De-blurring Blur Space Iterative De-blurring RADU CIPRIAN BILCU 1, MEJDI TRIMECHE 2, SAKARI ALENIUS 3, MARKKU VEHVILAINEN 4 1,2,3,4 Multimedia Technologies Laboratory, Nokia Research Center Visiokatu 1, FIN-33720,

More information

PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization

PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization Marie-Caroline Corbineau 1, Emilie Chouzenoux 1,2, Jean-Christophe Pesquet 1 1 CVN, CentraleSupélec, Université Paris-Saclay,

More information

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1 Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients

More information

Learning Splines for Sparse Tomographic Reconstruction. Elham Sakhaee and Alireza Entezari University of Florida

Learning Splines for Sparse Tomographic Reconstruction. Elham Sakhaee and Alireza Entezari University of Florida Learning Splines for Sparse Tomographic Reconstruction Elham Sakhaee and Alireza Entezari University of Florida esakhaee@cise.ufl.edu 2 Tomographic Reconstruction Recover the image given X-ray measurements

More information

Compressed Sensing Reconstructions for Dynamic Contrast Enhanced MRI

Compressed Sensing Reconstructions for Dynamic Contrast Enhanced MRI 1 Compressed Sensing Reconstructions for Dynamic Contrast Enhanced MRI Kevin T. Looby klooby@stanford.edu ABSTRACT The temporal resolution necessary for dynamic contrast enhanced (DCE) magnetic resonance

More information

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal

More information

An Improved Approach For Mixed Noise Removal In Color Images

An Improved Approach For Mixed Noise Removal In Color Images An Improved Approach For Mixed Noise Removal In Color Images Ancy Mariam Thomas 1, Dr. Deepa J 2, Rijo Sam 3 1P.G. student, College of Engineering, Chengannur, Kerala, India. 2Associate Professor, Electronics

More information

The picasso Package for High Dimensional Regularized Sparse Learning in R

The picasso Package for High Dimensional Regularized Sparse Learning in R The picasso Package for High Dimensional Regularized Sparse Learning in R X. Li, J. Ge, T. Zhang, M. Wang, H. Liu, and T. Zhao Abstract We introduce an R package named picasso, which implements a unified

More information

SEQUENTIAL IMAGE COMPLETION FOR HIGH-SPEED LARGE-PIXEL NUMBER SENSING

SEQUENTIAL IMAGE COMPLETION FOR HIGH-SPEED LARGE-PIXEL NUMBER SENSING SEQUENTIAL IMAGE COMPLETION FOR HIGH-SPEED LARGE-PIXEL NUMBER SENSING Akira Hirabayashi Naoki Nogami Takashi Ijiri Laurent Condat Ritsumeikan University College Info. Science & Eng. Kusatsu, Shiga 525-8577,

More information

CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT

CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT 9.1 Introduction In the previous chapters the inpainting was considered as an iterative algorithm. PDE based method uses iterations to converge

More information

Detection Performance of Radar Compressive Sensing in Noisy Environments

Detection Performance of Radar Compressive Sensing in Noisy Environments Detection Performance of Radar Compressive Sensing in Noisy Environments Asmita Korde a,damon Bradley b and Tinoosh Mohsenin a a Department of Computer Science and Electrical Engineering, University of

More information

Main Menu. Summary. Introduction

Main Menu. Summary. Introduction Noise-thresholding sparse-spike inversion with global convergence: calibration and applications Douglas S. Sassen* and Michael Glinsky, Ion Geophysical Corp. Summary We demonstrate an innovative inversion

More information

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University

DS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University September 20 2018 Review Solution for multiple linear regression can be computed in closed form

More information

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics 1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!

More information

Sparse Reconstruction / Compressive Sensing

Sparse Reconstruction / Compressive Sensing Sparse Reconstruction / Compressive Sensing Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani Sparse Reconstruction / Compressive Sensing 1/ 20 The

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

Lecture 2: Seismic Inversion and Imaging

Lecture 2: Seismic Inversion and Imaging Lecture 2: Seismic Inversion and Imaging Youzuo Lin 1 Joint work with: Lianjie Huang 1 Monica Maceira 2 Ellen M. Syracuse 1 Carene Larmat 1 1: Los Alamos National Laboratory 2: Oak Ridge National Laboratory

More information

Recent Advances inelectrical & Electronic Engineering

Recent Advances inelectrical & Electronic Engineering 4 Send Orders for Reprints to reprints@benthamscience.ae Recent Advances in Electrical & Electronic Engineering, 017, 10, 4-47 RESEARCH ARTICLE ISSN: 35-0965 eissn: 35-0973 Image Inpainting Method Based

More information

Sparse wavelet expansions for seismic tomography: Methods and algorithms

Sparse wavelet expansions for seismic tomography: Methods and algorithms Sparse wavelet expansions for seismic tomography: Methods and algorithms Ignace Loris Université Libre de Bruxelles International symposium on geophysical imaging with localized waves 24 28 July 2011 (Joint

More information

TOTAL VARIATION DENOISING WITH OVERLAPPING GROUP SPARSITY. Ivan W. Selesnick and Po-Yu Chen

TOTAL VARIATION DENOISING WITH OVERLAPPING GROUP SPARSITY. Ivan W. Selesnick and Po-Yu Chen TOTAL VARIATION DENOISING WITH OVERLAPPING GROUP SPARSITY Ivan W. Selesnick and Po-Yu Chen Polytechnic Institute of New York University, Brooklyn, NY 2, USA ABSTRACT This paper describes an extension to

More information

Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy

Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Paroma Varma Stanford University paroma@stanford.edu Abstract In microscopy, shot noise dominates image formation, which

More information

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING Idan Ram, Michael Elad and Israel Cohen Department of Electrical Engineering Department of Computer Science Technion - Israel Institute of Technology

More information

The Trainable Alternating Gradient Shrinkage method

The Trainable Alternating Gradient Shrinkage method The Trainable Alternating Gradient Shrinkage method Carlos Malanche, supervised by Ha Nguyen April 24, 2018 LIB (prof. Michael Unser), EPFL Quick example 1 The reconstruction problem Challenges in image

More information

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Moritz Baecher May 15, 29 1 Introduction Edge-preserving smoothing and super-resolution are classic and important

More information

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust

More information

Image Restoration using Accelerated Proximal Gradient method

Image Restoration using Accelerated Proximal Gradient method Image Restoration using Accelerated Proximal Gradient method Alluri.Samuyl Department of computer science, KIET Engineering College. D.Srinivas Asso.prof, Department of computer science, KIET Engineering

More information

SPARSE REPRESENTATION FOR IMAGE PREDICTION. Aurélie Martin, Jean-Jacques Fuchs, Christine Guillemot and Dominique Thoreau

SPARSE REPRESENTATION FOR IMAGE PREDICTION. Aurélie Martin, Jean-Jacques Fuchs, Christine Guillemot and Dominique Thoreau 15th European Signal Processing Conference (EUSIPCO 27), Poznan, Poland, September 3-7, 27, copyright by EURASIP SPARSE REPRESENTATION FOR IMAGE PREDICTION Aurélie Martin, Jean-Jacques Fuchs, Christine

More information

Algorithms for sparse X-ray CT image reconstruction of objects with known contour

Algorithms for sparse X-ray CT image reconstruction of objects with known contour Center for Nondestructive Evaluation Conference Papers, Posters and Presentations Center for Nondestructive Evaluation 7-2011 Algorithms for sparse X-ray CT image reconstruction of objects with known contour

More information

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments

A Patch Prior for Dense 3D Reconstruction in Man-Made Environments A Patch Prior for Dense 3D Reconstruction in Man-Made Environments Christian Häne 1, Christopher Zach 2, Bernhard Zeisl 1, Marc Pollefeys 1 1 ETH Zürich 2 MSR Cambridge October 14, 2012 A Patch Prior for

More information

Image Inpainting Using Sparsity of the Transform Domain

Image Inpainting Using Sparsity of the Transform Domain Image Inpainting Using Sparsity of the Transform Domain H. Hosseini*, N.B. Marvasti, Student Member, IEEE, F. Marvasti, Senior Member, IEEE Advanced Communication Research Institute (ACRI) Department of

More information

Notes on Robust Estimation David J. Fleet Allan Jepson March 30, 005 Robust Estimataion. The field of robust statistics [3, 4] is concerned with estimation problems in which the data contains gross errors,

More information

Convex Optimization MLSS 2015

Convex Optimization MLSS 2015 Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :

More information

Computational Methods. Constrained Optimization

Computational Methods. Constrained Optimization Computational Methods Constrained Optimization Manfred Huber 2010 1 Constrained Optimization Unconstrained Optimization finds a minimum of a function under the assumption that the parameters can take on

More information

Word Matching of handwritten scripts

Word Matching of handwritten scripts Word Matching of handwritten scripts Seminar about ancient document analysis Introduction Contour extraction Contour matching Other methods Conclusion Questions Problem Text recognition in handwritten

More information

Problem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation

Problem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation 6.098/6.882 Computational Photography 1 Problem Set 4 Assigned: March 23, 2006 Due: April 17, 2006 Problem 1 (6.882) Belief Propagation for Segmentation In this problem you will set-up a Markov Random

More information

Introduction to Topics in Machine Learning

Introduction to Topics in Machine Learning Introduction to Topics in Machine Learning Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani 1/ 27 Compressed Sensing / Sparse Recovery: Given y :=

More information

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi Lagrangian methods for the regularization of discrete ill-posed problems G. Landi Abstract In many science and engineering applications, the discretization of linear illposed problems gives rise to large

More information

Robust Seismic Image Amplitude Recovery Using Curvelets

Robust Seismic Image Amplitude Recovery Using Curvelets Robust Seismic Image Amplitude Recovery Using Curvelets Peyman P. Moghaddam Joint work with Felix J. Herrmann and Christiaan C. Stolk UNIVERSITY OF BRITISH COLUMBIA University of Twente References Curvelets

More information

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks

Compressive Sensing for Multimedia. Communications in Wireless Sensor Networks Compressive Sensing for Multimedia 1 Communications in Wireless Sensor Networks Wael Barakat & Rabih Saliba MDDSP Project Final Report Prof. Brian L. Evans May 9, 2008 Abstract Compressive Sensing is an

More information

SpicyMKL Efficient multiple kernel learning method using dual augmented Lagrangian

SpicyMKL Efficient multiple kernel learning method using dual augmented Lagrangian SpicyMKL Efficient multiple kernel learning method using dual augmented Lagrangian Taiji Suzuki Ryota Tomioka The University of Tokyo Graduate School of Information Science and Technology Department of

More information

Synthesis and Analysis Sparse Representation Models for Image Restoration. Shuhang Gu 顾舒航. Dept. of Computing The Hong Kong Polytechnic University

Synthesis and Analysis Sparse Representation Models for Image Restoration. Shuhang Gu 顾舒航. Dept. of Computing The Hong Kong Polytechnic University Synthesis and Analysis Sparse Representation Models for Image Restoration Shuhang Gu 顾舒航 Dept. of Computing The Hong Kong Polytechnic University Outline Sparse representation models for image modeling

More information

A fast iterative thresholding algorithm for wavelet-regularized deconvolution

A fast iterative thresholding algorithm for wavelet-regularized deconvolution A fast iterative thresholding algorithm for wavelet-regularized deconvolution Cédric Vonesch and Michael Unser Biomedical Imaging Group, EPFL, Lausanne, Switzerland ABSTRACT We present an iterative deconvolution

More information

Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey. Chapter 4 : Optimization for Machine Learning

Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey. Chapter 4 : Optimization for Machine Learning Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey Chapter 4 : Optimization for Machine Learning Summary of Chapter 2 Chapter 2: Convex Optimization with Sparsity

More information

A New Soft-Thresholding Image Denoising Method

A New Soft-Thresholding Image Denoising Method Available online at www.sciencedirect.com Procedia Technology 6 (2012 ) 10 15 2nd International Conference on Communication, Computing & Security [ICCCS-2012] A New Soft-Thresholding Image Denoising Method

More information

Compressive Sensing MRI with Wavelet Tree Sparsity

Compressive Sensing MRI with Wavelet Tree Sparsity Compressive Sensing MRI with Wavelet Tree Sparsity Chen Chen and Junzhou Huang Department of Computer Science and Engineering University of Texas at Arlington cchen@mavs.uta.edu jzhuang@uta.edu Abstract

More information

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007 : Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials

More information

ENHANCED RADAR IMAGING VIA SPARSITY REGULARIZED 2D LINEAR PREDICTION

ENHANCED RADAR IMAGING VIA SPARSITY REGULARIZED 2D LINEAR PREDICTION ENHANCED RADAR IMAGING VIA SPARSITY REGULARIZED 2D LINEAR PREDICTION I.Erer 1, K. Sarikaya 1,2, H.Bozkurt 1 1 Department of Electronics and Telecommunications Engineering Electrics and Electronics Faculty,

More information

arxiv: v1 [cs.lg] 3 Jan 2018

arxiv: v1 [cs.lg] 3 Jan 2018 CLUSTERING OF DATA WITH MISSING ENTRIES Sunrita Poddar, Mathews Jacob Department of Electrical and Computer Engineering, University of Iowa, IA, USA arxiv:1801.01455v1 [cs.lg] 3 Jan 018 ABSTRACT The analysis

More information

Optimization for Machine Learning

Optimization for Machine Learning with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex

More information

A MAP Algorithm for AVO Seismic Inversion Based on the Mixed (L 2, non-l 2 ) Norms to Separate Primary and Multiple Signals in Slowness Space

A MAP Algorithm for AVO Seismic Inversion Based on the Mixed (L 2, non-l 2 ) Norms to Separate Primary and Multiple Signals in Slowness Space A MAP Algorithm for AVO Seismic Inversion Based on the Mixed (L 2, non-l 2 ) Norms to Separate Primary and Multiple Signals in Slowness Space Harold Ivan Angulo Bustos Rio Grande do Norte State University

More information

Proximal operator and methods

Proximal operator and methods Proximal operator and methods Master 2 Data Science, Univ. Paris Saclay Robert M. Gower Optimization Sum of Terms A Datum Function Finite Sum Training Problem The Training Problem Convergence GD I Theorem

More information

CPSC 340: Machine Learning and Data Mining. Robust Regression Fall 2015

CPSC 340: Machine Learning and Data Mining. Robust Regression Fall 2015 CPSC 340: Machine Learning and Data Mining Robust Regression Fall 2015 Admin Can you see Assignment 1 grades on UBC connect? Auditors, don t worry about it. You should already be working on Assignment

More information

Blind Image Restoration Combined Wavelet Transform and RBF Network

Blind Image Restoration Combined Wavelet Transform and RBF Network Blind Image Restoration Combined Wavelet Transform and RBF Network Hongzhai Li and Ping Guo Department of Computer Science Beijing Normal University, Beijing, 100875, P.R.China lihongzhai@163.com; pguo@ieee.org

More information

Image Interpolation Using Multiscale Geometric Representations

Image Interpolation Using Multiscale Geometric Representations Image Interpolation Using Multiscale Geometric Representations Nickolaus Mueller, Yue Lu and Minh N. Do Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign ABSTRACT

More information

RESOLUTION is an unavoidable topic in seismic interpretation,

RESOLUTION is an unavoidable topic in seismic interpretation, IEEE GEOSCIENCE AND REMOTE SENSING LETTERS 1 Optimal Seismic Reflectivity Inversion: Data-driven l p -loss-l q -regularization Sparse Regression Fangyu Li, Rui Xie, Student Member, IEEE, Wen-Zhan Song,

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider

More information