Convex or non-convex: which is better?
|
|
- Vivien Nash
- 5 years ago
- Views:
Transcription
1 Sparsity Amplified Ivan Selesnick Electrical and Computer Engineering Tandon School of Engineering New York University Brooklyn, New York March 7 / 4
2 Convex or non-convex: which is better? (for sparse-regularized linear inverse problems) Benefits of convex optimization:. Absence of suboptimal local minima. Continuity of solution as a function of input data 3. Algorithms guaranteed to converge to a global optimum 4. Regularization parameters easier to set But convex regularization tends to under-estimate signal values. Non-convex regularization often performs better! Can we design non-convex sparsity-inducing penalties that maintain the convexity of the cost function to be minimized? / 4
3 An alternative Conventional sparse-regularized least squares: J(x) = y Ax + λ x Alternatively, use non-convex penalty ψ that maintains convexity of cost function F F (x) = y Ax + λψ(x) I. Selesnick. Total Variation Denoising via the Moreau Envelope. IEEE Signal Processing Letters, 4():6, February 7. I. W. Selesnick and I. Bayram. Enhanced Sparsity by Non-Separable Regularization. IEEE Trans. Signal Process., 64(9):98 33, May 6. A. Parekh and I. W. Selesnick. Enhanced low-rank matrix approximation. IEEE Signal Processing Letters, 3(4): , April 6. Y. Ding and I. W. Selesnick. Artifact-Free Wavelet Denoising: Non-convex Sparse Regularization, Convex Optimization. IEEE Signal Processing Letters, (9): , September 5. I. W. Selesnick, A. Parekh, and I. Bayram. Convex -D Total Variation Denoising with Non-convex Regularization. IEEE Signal Processing Letters, ():4 44, February 5. 3 / 4
4 MC penalty The minimax-concave (MC) penalty φ: R R is defined as x φ(x) := x, x, x 3 x φ(x) 3 3 x C.-H. Zhang. Nearly unbiased variable selection under minimax concave penalty. Statistics:894 94,. The Annals of 4 / 4
5 Huber function The MC penalty can be expressed as φ(x) = x s(x) where s is the Huber function x The Huber function. 5 / 4
6 Huber function The Huber function s : R R is defined as s(x) := x, x x, x. The Huber function can be written as s(x) = min v R { v + (x v) }, equivalently s = ( ) where denotes infimal convolution. H. H. Bauschke and P. L. Combettes. Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer,. 6 / 4
7 Huber function 3.5 x x +.5 x x The Huber function as the pointwise minimum of three functions. 7 / 4
8 Scaled functions Let b R. The scaled Huber function s b : R R is defined as s b (x) := s(b x)/b, b. For b =, the function is defined as s (x) :=. The scaled MC penalty function φ b : R R is defined as φ b (x) := x s b (x) where s b is the scaled Huber function. 8 / 4
9 Scaled functions Scaled Huber function and MC penalty for several values of the scaling parameter. 3 Scaled Huber function s b (x) b =.4 b =. b = x 3 Scaled MC penalty φ b (x) b =. b =.7 b =. b = x 9 / 4
10 Convexity condition Let λ > and a R. Define f : R R, f (x) = (y ax) + λ φ b (x) where φ b is the scaled MC penalty. If b a /λ, then. f is convex. the minimizer is given by firm thresholding, x opt = firm(y/a; λ/a, /b ). / 4
11 Firm threshold function firm(y; λ, µ) 4 Firm threshold function 3 λ µ y / 4
12 Aims The minimizer of the scalar function f is easily obtained via firm thresholding. However, the situation in the multivariate case is more complicated. To generalize this process to the multivariate case, we aim to:. define a multivariate MC penalty (non-convex),. define a sparse-regularized least squares cost function (convex), 3. generalize the convexity condition, 4. provide a method to calculate a minimizer. / 4
13 Generalized Huber function Let B R M N. We define the generalized Huber function S B : R N R as { S B (x) := min v + v R N B(x } v). In the notation of infimal convolution, we have S B = B. The generalized Huber function satisfies S B (x) x, x R N. 3 / 4
14 Generalized Huber function Generalized Huber function S B (x) 3 B = x x Contours of S B 4 / 4
15 Generalized Huber function Generalized Huber function S B (x) B = [.5] x x Contours of S B 5 / 4
16 Generalized MC penalty Let B R M N. We define the generalized MC (GMC) penalty function ψ B : R N R as ψ B (x) := x S B (x) where S B is the generalized Huber function. The generalized MC penalty satisfies ψ B (x) x for all x R N. 6 / 4
17 Generalized MC penalty x S B (x) ψ B (x) = x S B (x) x x x x Contours Contours Contours 7 / 4
18 Generalized MC penalty x S B (x) ψ B (x) = x S B (x) x x x x Contours Contours Contours 8 / 4
19 Convexity condition Let y R M, A R M N, and λ >. Define F : R N R as F (x) = y Ax + λ ψ B(x) where ψ B is the generalized MC penalty. If B T B λ AT A, then F is a convex function. Hence, for convexity of F, we may simply set B = γ/λ A, γ. () 9 / 4
20 Proximal algorithm Using forward-backward splitting, we have: Let λ > and γ <. Let y R N and A R M N. Then a saddle-point (x opt, v opt ) of F can be obtained by the iterative algorithm: Set ρ = max{, γ/( γ)} A T A Set µ : < µ < /ρ For i =,,,... w (i) = x (i) µa T( A(x (i) + γ(v (i) x (i) )) y ) u (i) = v (i) µγa T A(v (i) x (i) ) x (i+) = soft(w (i), µλ) v (i+) = soft(u (i), µλ) end where i is the iteration counter. / 4
21 Example: frequency-domain sparsity Noise free signal Noisy signal, σ = Denoising via frequency domain sparsity Denoising [L norm] 4 4 RMSE =.43 5 Denoising [GMC penalty] 4 4 RMSE =.49 5 Average RMSE L norm L + debias IPS MUSR GMC 6 4 FFT of noisy signal 5 Optimized Fourier coefficients o GMC x L norm λ Frequency Frequency / 4
22 Example: time-frequency domain sparsity.3 True signal Time frequency representation (db) Frequency (khz) Time (ms) Noisy signal Frequency (khz) 6 4 Time (ms) Time frequency representation (db).3.3 RMSE =.6 Time (ms) Denoising using L norm Frequency (khz) 6 4 Time (ms) Time frequency representation (db).3.3 RMSE =.6 Time (ms) Denoising using GMC penalty Frequency (khz) 6 4 Time (ms) Time frequency representation (db).3 Time (ms) Time (ms) / 4
23 Example: sparsity-assisted signal smoothing (SASS) Biosensor data SASS using L norm (a) SASS using GMC penalty 5 5 (b) Magnified view data L norm GMC (c) (d) 3 / 4
24 Conclusion Convex and the non-convex methods for sparse-regularized least squares are usually mutually exclusive and incompatible. To bridge these two approaches, we introduce a non-convex alternative to the L norm that preserves the convexity of the cost function to be minimized. The proposed penalty leads to optimization problems with no extraneous suboptimal local minima and allows the use of globally convergent, computationally efficient, scalable convex optimization algorithms. The advantages compared to L norm regularization are (i) more accurate estimation of high-amplitude components of sparse solutions, (ii) a higher level of sparsity in a sparse approximation problem. 4 / 4
A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]
A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song
More informationSparsity and image processing
Sparsity and image processing Aurélie Boisbunon INRIA-SAM, AYIN March 6, Why sparsity? Main advantages Dimensionality reduction Fast computation Better interpretability Image processing pattern recognition
More informationTotal Variation Denoising with Overlapping Group Sparsity
1 Total Variation Denoising with Overlapping Group Sparsity Ivan W. Selesnick and Po-Yu Chen Polytechnic Institute of New York University Brooklyn, New York selesi@poly.edu 2 Abstract This paper describes
More informationCollaborative Sparsity and Compressive MRI
Modeling and Computation Seminar February 14, 2013 Table of Contents 1 T2 Estimation 2 Undersampling in MRI 3 Compressed Sensing 4 Model-Based Approach 5 From L1 to L0 6 Spatially Adaptive Sparsity MRI
More informationSupplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging
Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging Paper ID: 1195 No Institute Given 1 More Experiment Results 1.1 Visual Comparisons We apply all methods on four 2D
More informationAdvanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude
Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,
More informationConvex optimization algorithms for sparse and low-rank representations
Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,
More informationIterative CT Reconstruction Using Curvelet-Based Regularization
Iterative CT Reconstruction Using Curvelet-Based Regularization Haibo Wu 1,2, Andreas Maier 1, Joachim Hornegger 1,2 1 Pattern Recognition Lab (LME), Department of Computer Science, 2 Graduate School in
More informationELEG Compressive Sensing and Sparse Signal Representations
ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /
More informationG Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine. Compressed Sensing
G16.4428 Practical Magnetic Resonance Imaging II Sackler Institute of Biomedical Sciences New York University School of Medicine Compressed Sensing Ricardo Otazo, PhD ricardo.otazo@nyumc.org Compressed
More informationL1 REGULARIZED STAP ALGORITHM WITH A GENERALIZED SIDELOBE CANCELER ARCHITECTURE FOR AIRBORNE RADAR
L1 REGULARIZED STAP ALGORITHM WITH A GENERALIZED SIDELOBE CANCELER ARCHITECTURE FOR AIRBORNE RADAR Zhaocheng Yang, Rodrigo C. de Lamare and Xiang Li Communications Research Group Department of Electronics
More informationDenoising Using Projection Onto Convex Sets (POCS) Based Framework
Denoising Using Projection Onto Convex Sets (POCS) Based Framework Mohammad Tofighi, Kivanc Kose, A. Enis Cetin Dept. of Electrical and Electronic Engineering, Bilkent University, Ankara, Turkey Dermatology
More informationOptimization for Machine Learning
Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html
More informationPolynomial Smoothing of Time Series with Additive Step Discontinuities
JUNE 5, Polynomial Smoothing of Time Series with Additive Step Discontinuities Ivan W. Selesnick, Stephen Arnold, and Venkata R. Dantham Abstract This paper addresses the problem of estimating simultaneously
More informationDivide and Conquer Kernel Ridge Regression
Divide and Conquer Kernel Ridge Regression Yuchen Zhang John Duchi Martin Wainwright University of California, Berkeley COLT 2013 Yuchen Zhang (UC Berkeley) Divide and Conquer KRR COLT 2013 1 / 15 Problem
More informationBandwidth Selection for Kernel Density Estimation Using Total Variation with Fourier Domain Constraints
IEEE SIGNAL PROCESSING LETTERS 1 Bandwidth Selection for Kernel Density Estimation Using Total Variation with Fourier Domain Constraints Alexander Suhre, Orhan Arikan, Member, IEEE, and A. Enis Cetin,
More informationIteratively Re-weighted Least Squares for Sums of Convex Functions
Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He
More informationRecent Developments in Model-based Derivative-free Optimization
Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:
More informationSignal Reconstruction from Sparse Representations: An Introdu. Sensing
Signal Reconstruction from Sparse Representations: An Introduction to Compressed Sensing December 18, 2009 Digital Data Acquisition Suppose we want to acquire some real world signal digitally. Applications
More informationImage Restoration with Compound Regularization Using a Bregman Iterative Algorithm
Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm Manya V. Afonso, José M. Bioucas-Dias, Mario A. T. Figueiredo To cite this version: Manya V. Afonso, José M. Bioucas-Dias,
More informationBlind Compressed Sensing Using Sparsifying Transforms
Blind Compressed Sensing Using Sparsifying Transforms Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laboratory University of Illinois
More informationOutline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff
Compressed Sensing David L Donoho Presented by: Nitesh Shroff University of Maryland Outline 1 Introduction Compressed Sensing 2 Problem Formulation Sparse Signal Problem Statement 3 Proposed Solution
More informationA primal-dual framework for mixtures of regularizers
A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland
More informationOptimal Sampling Geometries for TV-Norm Reconstruction of fmri Data
Optimal Sampling Geometries for TV-Norm Reconstruction of fmri Data Oliver M. Jeromin, Student Member, IEEE, Vince D. Calhoun, Senior Member, IEEE, and Marios S. Pattichis, Senior Member, IEEE Abstract
More informationPicasso: A Sparse Learning Library for High Dimensional Data Analysis in R and Python
Picasso: A Sparse Learning Library for High Dimensional Data Analysis in R and Python J. Ge, X. Li, H. Jiang, H. Liu, T. Zhang, M. Wang and T. Zhao Abstract We describe a new library named picasso, which
More informationThe Benefit of Tree Sparsity in Accelerated MRI
The Benefit of Tree Sparsity in Accelerated MRI Chen Chen and Junzhou Huang Department of Computer Science and Engineering, The University of Texas at Arlington, TX, USA 76019 Abstract. The wavelet coefficients
More informationImage Denoising Using Sparse Representations
Image Denoising Using Sparse Representations SeyyedMajid Valiollahzadeh 1,,HamedFirouzi 1, Massoud Babaie-Zadeh 1, and Christian Jutten 2 1 Department of Electrical Engineering, Sharif University of Technology,
More informationAdaptive Reconstruction Methods for Low-Dose Computed Tomography
Adaptive Reconstruction Methods for Low-Dose Computed Tomography Joseph Shtok Ph.D. supervisors: Prof. Michael Elad, Dr. Michael Zibulevsky. Technion IIT, Israel, 011 Ph.D. Talk, Apr. 01 Contents of this
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex
More informationGabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs
Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs Benjamin Ricaud Signal Processing Lab 2, EPFL Lausanne, Switzerland SPLab, Brno, 10/2013 B. Ricaud
More informationA Relationship between the Robust Statistics Theory and Sparse Compressive Sensed Signals Reconstruction
THIS PAPER IS A POSTPRINT OF A PAPER SUBMITTED TO AND ACCEPTED FOR PUBLICATION IN IET SIGNAL PROCESSING AND IS SUBJECT TO INSTITUTION OF ENGINEERING AND TECHNOLOGY COPYRIGHT. THE COPY OF RECORD IS AVAILABLE
More informationConvexization in Markov Chain Monte Carlo
in Markov Chain Monte Carlo 1 IBM T. J. Watson Yorktown Heights, NY 2 Department of Aerospace Engineering Technion, Israel August 23, 2011 Problem Statement MCMC processes in general are governed by non
More informationCOMPRESSIVE VIDEO SAMPLING
COMPRESSIVE VIDEO SAMPLING Vladimir Stanković and Lina Stanković Dept of Electronic and Electrical Engineering University of Strathclyde, Glasgow, UK phone: +44-141-548-2679 email: {vladimir,lina}.stankovic@eee.strath.ac.uk
More informationCompressive Sensing Algorithms for Fast and Accurate Imaging
Compressive Sensing Algorithms for Fast and Accurate Imaging Wotao Yin Department of Computational and Applied Mathematics, Rice University SCIMM 10 ASU, Tempe, AZ Acknowledgements: results come in part
More informationLearning with infinitely many features
Learning with infinitely many features R. Flamary, Joint work with A. Rakotomamonjy F. Yger, M. Volpi, M. Dalla Mura, D. Tuia Laboratoire Lagrange, Université de Nice Sophia Antipolis December 2012 Example
More informationCurvelet-based non-linear adaptive subtraction with sparseness constraints. Felix J Herrmann, Peyman P Moghaddam (EOS-UBC)
Curvelet-based non-linear adaptive subtraction with sparseness constraints Felix J Herrmann, Peyman P Moghaddam (EOS-UBC) Context Spiky/minimal structure deconvolution (Claerbout, Ulrych, Oldenburg, Sacchi,
More informationSUMMARY THEORY. L q Norm Reflectivity Inversion
Optimal L q Norm Regularization for Sparse Reflectivity Inversion Fangyu Li, University of Oklahoma & University of Georgia; Rui Xie, University of Georgia; WenZhan Song, University of Georgia & Intelligent
More informationACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A.
ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS Donghwan Kim and Jeffrey A. Fessler University of Michigan Dept. of Electrical Engineering and Computer Science
More informationADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING
ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING Fei Yang 1 Hong Jiang 2 Zuowei Shen 3 Wei Deng 4 Dimitris Metaxas 1 1 Rutgers University 2 Bell Labs 3 National University
More informationGeneralized Tree-Based Wavelet Transform and Applications to Patch-Based Image Processing
Generalized Tree-Based Wavelet Transform and * Michael Elad The Computer Science Department The Technion Israel Institute of technology Haifa 32000, Israel *Joint work with A Seminar in the Hebrew University
More informationEfficient MR Image Reconstruction for Compressed MR Imaging
Efficient MR Image Reconstruction for Compressed MR Imaging Junzhou Huang, Shaoting Zhang, and Dimitris Metaxas Division of Computer and Information Sciences, Rutgers University, NJ, USA 08854 Abstract.
More informationKernels and representation
Kernels and representation Corso di AA, anno 2017/18, Padova Fabio Aiolli 20 Dicembre 2017 Fabio Aiolli Kernels and representation 20 Dicembre 2017 1 / 19 (Hierarchical) Representation Learning Hierarchical
More informationLecture 25 Nonlinear Programming. November 9, 2009
Nonlinear Programming November 9, 2009 Outline Nonlinear Programming Another example of NLP problem What makes these problems complex Scalar Function Unconstrained Problem Local and global optima: definition,
More informationIMA Preprint Series # 2211
LEARNING TO SENSE SPARSE SIGNALS: SIMULTANEOUS SENSING MATRIX AND SPARSIFYING DICTIONARY OPTIMIZATION By Julio Martin Duarte-Carvajalino and Guillermo Sapiro IMA Preprint Series # 2211 ( May 2008 ) INSTITUTE
More informationBlur Space Iterative De-blurring
Blur Space Iterative De-blurring RADU CIPRIAN BILCU 1, MEJDI TRIMECHE 2, SAKARI ALENIUS 3, MARKKU VEHVILAINEN 4 1,2,3,4 Multimedia Technologies Laboratory, Nokia Research Center Visiokatu 1, FIN-33720,
More informationPIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization
PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization Marie-Caroline Corbineau 1, Emilie Chouzenoux 1,2, Jean-Christophe Pesquet 1 1 CVN, CentraleSupélec, Université Paris-Saclay,
More informationSection 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1
Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients
More informationLearning Splines for Sparse Tomographic Reconstruction. Elham Sakhaee and Alireza Entezari University of Florida
Learning Splines for Sparse Tomographic Reconstruction Elham Sakhaee and Alireza Entezari University of Florida esakhaee@cise.ufl.edu 2 Tomographic Reconstruction Recover the image given X-ray measurements
More informationCompressed Sensing Reconstructions for Dynamic Contrast Enhanced MRI
1 Compressed Sensing Reconstructions for Dynamic Contrast Enhanced MRI Kevin T. Looby klooby@stanford.edu ABSTRACT The temporal resolution necessary for dynamic contrast enhanced (DCE) magnetic resonance
More informationSparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual
Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal
More informationAn Improved Approach For Mixed Noise Removal In Color Images
An Improved Approach For Mixed Noise Removal In Color Images Ancy Mariam Thomas 1, Dr. Deepa J 2, Rijo Sam 3 1P.G. student, College of Engineering, Chengannur, Kerala, India. 2Associate Professor, Electronics
More informationThe picasso Package for High Dimensional Regularized Sparse Learning in R
The picasso Package for High Dimensional Regularized Sparse Learning in R X. Li, J. Ge, T. Zhang, M. Wang, H. Liu, and T. Zhao Abstract We introduce an R package named picasso, which implements a unified
More informationSEQUENTIAL IMAGE COMPLETION FOR HIGH-SPEED LARGE-PIXEL NUMBER SENSING
SEQUENTIAL IMAGE COMPLETION FOR HIGH-SPEED LARGE-PIXEL NUMBER SENSING Akira Hirabayashi Naoki Nogami Takashi Ijiri Laurent Condat Ritsumeikan University College Info. Science & Eng. Kusatsu, Shiga 525-8577,
More informationCHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT
CHAPTER 9 INPAINTING USING SPARSE REPRESENTATION AND INVERSE DCT 9.1 Introduction In the previous chapters the inpainting was considered as an iterative algorithm. PDE based method uses iterations to converge
More informationDetection Performance of Radar Compressive Sensing in Noisy Environments
Detection Performance of Radar Compressive Sensing in Noisy Environments Asmita Korde a,damon Bradley b and Tinoosh Mohsenin a a Department of Computer Science and Electrical Engineering, University of
More informationMain Menu. Summary. Introduction
Noise-thresholding sparse-spike inversion with global convergence: calibration and applications Douglas S. Sassen* and Michael Glinsky, Ion Geophysical Corp. Summary We demonstrate an innovative inversion
More informationDS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University
DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University September 20 2018 Review Solution for multiple linear regression can be computed in closed form
More information1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics
1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!
More informationSparse Reconstruction / Compressive Sensing
Sparse Reconstruction / Compressive Sensing Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani Sparse Reconstruction / Compressive Sensing 1/ 20 The
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June
More informationLecture 2: Seismic Inversion and Imaging
Lecture 2: Seismic Inversion and Imaging Youzuo Lin 1 Joint work with: Lianjie Huang 1 Monica Maceira 2 Ellen M. Syracuse 1 Carene Larmat 1 1: Los Alamos National Laboratory 2: Oak Ridge National Laboratory
More informationRecent Advances inelectrical & Electronic Engineering
4 Send Orders for Reprints to reprints@benthamscience.ae Recent Advances in Electrical & Electronic Engineering, 017, 10, 4-47 RESEARCH ARTICLE ISSN: 35-0965 eissn: 35-0973 Image Inpainting Method Based
More informationSparse wavelet expansions for seismic tomography: Methods and algorithms
Sparse wavelet expansions for seismic tomography: Methods and algorithms Ignace Loris Université Libre de Bruxelles International symposium on geophysical imaging with localized waves 24 28 July 2011 (Joint
More informationTOTAL VARIATION DENOISING WITH OVERLAPPING GROUP SPARSITY. Ivan W. Selesnick and Po-Yu Chen
TOTAL VARIATION DENOISING WITH OVERLAPPING GROUP SPARSITY Ivan W. Selesnick and Po-Yu Chen Polytechnic Institute of New York University, Brooklyn, NY 2, USA ABSTRACT This paper describes an extension to
More informationApplication of Proximal Algorithms to Three Dimensional Deconvolution Microscopy
Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Paroma Varma Stanford University paroma@stanford.edu Abstract In microscopy, shot noise dominates image formation, which
More informationIMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING
IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING Idan Ram, Michael Elad and Israel Cohen Department of Electrical Engineering Department of Computer Science Technion - Israel Institute of Technology
More informationThe Trainable Alternating Gradient Shrinkage method
The Trainable Alternating Gradient Shrinkage method Carlos Malanche, supervised by Ha Nguyen April 24, 2018 LIB (prof. Michael Unser), EPFL Quick example 1 The reconstruction problem Challenges in image
More informationLocally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling
Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Moritz Baecher May 15, 29 1 Introduction Edge-preserving smoothing and super-resolution are classic and important
More informationRobust Kernel Methods in Clustering and Dimensionality Reduction Problems
Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust
More informationImage Restoration using Accelerated Proximal Gradient method
Image Restoration using Accelerated Proximal Gradient method Alluri.Samuyl Department of computer science, KIET Engineering College. D.Srinivas Asso.prof, Department of computer science, KIET Engineering
More informationSPARSE REPRESENTATION FOR IMAGE PREDICTION. Aurélie Martin, Jean-Jacques Fuchs, Christine Guillemot and Dominique Thoreau
15th European Signal Processing Conference (EUSIPCO 27), Poznan, Poland, September 3-7, 27, copyright by EURASIP SPARSE REPRESENTATION FOR IMAGE PREDICTION Aurélie Martin, Jean-Jacques Fuchs, Christine
More informationAlgorithms for sparse X-ray CT image reconstruction of objects with known contour
Center for Nondestructive Evaluation Conference Papers, Posters and Presentations Center for Nondestructive Evaluation 7-2011 Algorithms for sparse X-ray CT image reconstruction of objects with known contour
More informationA Patch Prior for Dense 3D Reconstruction in Man-Made Environments
A Patch Prior for Dense 3D Reconstruction in Man-Made Environments Christian Häne 1, Christopher Zach 2, Bernhard Zeisl 1, Marc Pollefeys 1 1 ETH Zürich 2 MSR Cambridge October 14, 2012 A Patch Prior for
More informationImage Inpainting Using Sparsity of the Transform Domain
Image Inpainting Using Sparsity of the Transform Domain H. Hosseini*, N.B. Marvasti, Student Member, IEEE, F. Marvasti, Senior Member, IEEE Advanced Communication Research Institute (ACRI) Department of
More informationNotes on Robust Estimation David J. Fleet Allan Jepson March 30, 005 Robust Estimataion. The field of robust statistics [3, 4] is concerned with estimation problems in which the data contains gross errors,
More informationConvex Optimization MLSS 2015
Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :
More informationComputational Methods. Constrained Optimization
Computational Methods Constrained Optimization Manfred Huber 2010 1 Constrained Optimization Unconstrained Optimization finds a minimum of a function under the assumption that the parameters can take on
More informationWord Matching of handwritten scripts
Word Matching of handwritten scripts Seminar about ancient document analysis Introduction Contour extraction Contour matching Other methods Conclusion Questions Problem Text recognition in handwritten
More informationProblem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation
6.098/6.882 Computational Photography 1 Problem Set 4 Assigned: March 23, 2006 Due: April 17, 2006 Problem 1 (6.882) Belief Propagation for Segmentation In this problem you will set-up a Markov Random
More informationIntroduction to Topics in Machine Learning
Introduction to Topics in Machine Learning Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University Namrata Vaswani 1/ 27 Compressed Sensing / Sparse Recovery: Given y :=
More informationLagrangian methods for the regularization of discrete ill-posed problems. G. Landi
Lagrangian methods for the regularization of discrete ill-posed problems G. Landi Abstract In many science and engineering applications, the discretization of linear illposed problems gives rise to large
More informationRobust Seismic Image Amplitude Recovery Using Curvelets
Robust Seismic Image Amplitude Recovery Using Curvelets Peyman P. Moghaddam Joint work with Felix J. Herrmann and Christiaan C. Stolk UNIVERSITY OF BRITISH COLUMBIA University of Twente References Curvelets
More informationCompressive Sensing for Multimedia. Communications in Wireless Sensor Networks
Compressive Sensing for Multimedia 1 Communications in Wireless Sensor Networks Wael Barakat & Rabih Saliba MDDSP Project Final Report Prof. Brian L. Evans May 9, 2008 Abstract Compressive Sensing is an
More informationSpicyMKL Efficient multiple kernel learning method using dual augmented Lagrangian
SpicyMKL Efficient multiple kernel learning method using dual augmented Lagrangian Taiji Suzuki Ryota Tomioka The University of Tokyo Graduate School of Information Science and Technology Department of
More informationSynthesis and Analysis Sparse Representation Models for Image Restoration. Shuhang Gu 顾舒航. Dept. of Computing The Hong Kong Polytechnic University
Synthesis and Analysis Sparse Representation Models for Image Restoration Shuhang Gu 顾舒航 Dept. of Computing The Hong Kong Polytechnic University Outline Sparse representation models for image modeling
More informationA fast iterative thresholding algorithm for wavelet-regularized deconvolution
A fast iterative thresholding algorithm for wavelet-regularized deconvolution Cédric Vonesch and Michael Unser Biomedical Imaging Group, EPFL, Lausanne, Switzerland ABSTRACT We present an iterative deconvolution
More informationIncremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey. Chapter 4 : Optimization for Machine Learning
Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey Chapter 4 : Optimization for Machine Learning Summary of Chapter 2 Chapter 2: Convex Optimization with Sparsity
More informationA New Soft-Thresholding Image Denoising Method
Available online at www.sciencedirect.com Procedia Technology 6 (2012 ) 10 15 2nd International Conference on Communication, Computing & Security [ICCCS-2012] A New Soft-Thresholding Image Denoising Method
More informationCompressive Sensing MRI with Wavelet Tree Sparsity
Compressive Sensing MRI with Wavelet Tree Sparsity Chen Chen and Junzhou Huang Department of Computer Science and Engineering University of Texas at Arlington cchen@mavs.uta.edu jzhuang@uta.edu Abstract
More informationLecture 19: Convex Non-Smooth Optimization. April 2, 2007
: Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials
More informationENHANCED RADAR IMAGING VIA SPARSITY REGULARIZED 2D LINEAR PREDICTION
ENHANCED RADAR IMAGING VIA SPARSITY REGULARIZED 2D LINEAR PREDICTION I.Erer 1, K. Sarikaya 1,2, H.Bozkurt 1 1 Department of Electronics and Telecommunications Engineering Electrics and Electronics Faculty,
More informationarxiv: v1 [cs.lg] 3 Jan 2018
CLUSTERING OF DATA WITH MISSING ENTRIES Sunrita Poddar, Mathews Jacob Department of Electrical and Computer Engineering, University of Iowa, IA, USA arxiv:1801.01455v1 [cs.lg] 3 Jan 018 ABSTRACT The analysis
More informationOptimization for Machine Learning
with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex
More informationA MAP Algorithm for AVO Seismic Inversion Based on the Mixed (L 2, non-l 2 ) Norms to Separate Primary and Multiple Signals in Slowness Space
A MAP Algorithm for AVO Seismic Inversion Based on the Mixed (L 2, non-l 2 ) Norms to Separate Primary and Multiple Signals in Slowness Space Harold Ivan Angulo Bustos Rio Grande do Norte State University
More informationProximal operator and methods
Proximal operator and methods Master 2 Data Science, Univ. Paris Saclay Robert M. Gower Optimization Sum of Terms A Datum Function Finite Sum Training Problem The Training Problem Convergence GD I Theorem
More informationCPSC 340: Machine Learning and Data Mining. Robust Regression Fall 2015
CPSC 340: Machine Learning and Data Mining Robust Regression Fall 2015 Admin Can you see Assignment 1 grades on UBC connect? Auditors, don t worry about it. You should already be working on Assignment
More informationBlind Image Restoration Combined Wavelet Transform and RBF Network
Blind Image Restoration Combined Wavelet Transform and RBF Network Hongzhai Li and Ping Guo Department of Computer Science Beijing Normal University, Beijing, 100875, P.R.China lihongzhai@163.com; pguo@ieee.org
More informationImage Interpolation Using Multiscale Geometric Representations
Image Interpolation Using Multiscale Geometric Representations Nickolaus Mueller, Yue Lu and Minh N. Do Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign ABSTRACT
More informationRESOLUTION is an unavoidable topic in seismic interpretation,
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS 1 Optimal Seismic Reflectivity Inversion: Data-driven l p -loss-l q -regularization Sparse Regression Fangyu Li, Rui Xie, Student Member, IEEE, Wen-Zhan Song,
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider
More information