A primal-dual framework for mixtures of regularizers

Size: px
Start display at page:

Download "A primal-dual framework for mixtures of regularizers"

Transcription

1 A primal-dual framework for mixtures of regularizers Baran Gözcü Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland EPFL [March 25, 2015] Joint work with Luca Baldassarre, Quoc Tran Dinh, Cosimo Aprile and Volkan LIONS

2 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 2/ 28

3 Overview of compressive imaging System model y = Mx + ω (1) M is the measurement matrix (Fourier, Gaussian etc.) x is the image that is in vectorized form ω is the image that is in vectorized form b is the measurement vector Solution Then we solve min x W x 1 subject to Mx = y (2) where W is the sparsity basis. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 3/ 28

4 Why a mixture model? Signals can posseses various structures at the same time. For example the reconstruction with TV norm x TV = i,j ( (x)) i,j 2 will introduce flat regions. min x subject to x TV Mx = y What happens if we solve with a mixture of regularizers? min x α x TV + (1 α) x 1 subject to Mx = y (3) (4) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 4/ 28

5 Illustration: Mixture model performs better SNR (db) Original (2048x2048,%15 of DCT) Original BP TV TV&BP tv-bp max SNR , A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 5/ 28

6 General Problem min x subject to p f (x) := i=0 f i(x) Ax = b (5) How to solve with computational efficiency guarantee on objective function guarantee on feasibility A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 6/ 28

7 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 7/ 28

8 Swiss army knife of convex formulations Our primal problem prototype: A simple mathematical formulation 1 { } f := min f (x) : Ax = b, x X, (6) x R p f is a proper, closed and convex function, and X is a nonempty, closed convex set. A R n p and b R n are known. An optimal solution x to (6) satisfies f (x ) = f, Ax = b and x X. 1 We can simply replace Ax = b with Ax b C for a convex cone C without any fundamental change. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 8/ 28

9 Swiss army knife of convex formulations Our primal problem prototype: A simple mathematical formulation 1 { } f := min f (x) : Ax = b, x X, (6) x R p f is a proper, closed and convex function, and X is a nonempty, closed convex set. A R n p and b R n are known. An optimal solution x to (6) satisfies f (x ) = f, Ax = b and x X. Example to keep in mind in the sequel { } x := arg min x 1 : Ax = b, x 1 x R p 1 We can simply replace Ax = b with Ax b C for a convex cone C without any fundamental change. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 8/ 28

10 Swiss army knife of convex formulations Our primal problem prototype: A simple mathematical formulation 1 { } f := min f (x) : Ax = b, x X, (6) x R p f is a proper, closed and convex function, and X is a nonempty, closed convex set. A R n p and b R n are known. An optimal solution x to (6) satisfies f (x ) = f, Ax = b and x X. Example to keep in mind in the sequel { } x := arg min x 1 : Ax = b, x 1 x R p Broader context for (6): Standard convex optimization formulations: linear programming, convex quadratic programming, second order cone programming, semidefinite programming and interior point algorithms. Reformulations of existing unconstrained problems via convex splitting: composite convex minimization, consensus optimization,... 1 We can simply replace Ax = b with Ax b C for a convex cone C without any fundamental change. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 8/ 28

11 Numerical ɛ-accuracy Exact vs. approximate solutions Computing an exact solution x to (6) is impracticable unless problem has a closed form solution, which is extremely limited in reality. Numerical optimization algorithms result in x ɛ that approximates x up to a given accuracy ɛ in some sense. In the sequel, by ɛ-accurate solutions x ɛ of (6), we mean the following Definition (ɛ-accurate solutions) Given a numerical tolerance ɛ 0, a point xɛ Rp is called an ɛ-solution of (6) if f (xɛ ) f ɛ (objective residual), Axɛ b ɛ (feasibility gap), xɛ X (exact simple set feasibility). Indeed, ɛ can be different for the objective, feasibility gap, or the iterate residual. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 9/ 28

12 The optimal solution set Before we talk about algorithms, we must first characterize what we are looking for! Optimality condition The optimality condition of min x R p {f (x) : Ax = b} can be written as { 0 A T λ + f (x ), 0 = Ax b. (Subdifferential) f (x) := {v R p : f (y) f (x) + v T (y x), y R p }. This is the well-known KKT (Karush-Kuhn-Tucker) condition. Any point (x, λ ) satisfying (7) is called a KKT point. x is called a stationary point and λ is the corresponding multipliers. Lagrange function and the minimax formulation We can naturally interpret the optimality condition via a minimax formulation max λ min L(x, λ), x dom(f ) where λ R n is the vector of Lagrange multipliers or dual variables w.r.t. Ax = b associated with the Lagrange function: L(x, λ) := f (x) + λ T (Ax b) (7) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 10/ 28

13 Finding an optimal solution A plausible strategy: To solve the constrained problem (6), we therefore seek the solutions (x, λ ) arg max min L(x, λ), λ x X which we can naively brake down into two in general nonsmooth problems: Lagrangian subproblem: x (λ) arg min x X {L(x, λ) := f (x) + λ, Ax b } Dual problem: λ arg max λ {d(λ) := L(x (λ), λ)} The function d(λ) is called the dual function. The optimal dual objective value is d = d(λ ). The dual function d(λ) is concave. Hence, we can attempt the following strategy: 1. Find the optimal solution λ of the convex dual problem. 2. Obtain the optimal primal solution x = x (λ ) via the convex primal problem. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 11/ 28

14 Finding an optimal solution A plausible strategy: To solve the constrained problem (6), we therefore seek the solutions (x, λ ) arg max min L(x, λ), λ x X which we can naively brake down into two in general nonsmooth problems: Lagrangian subproblem: x (λ) arg min x X {L(x, λ) := f (x) + λ, Ax b } Dual problem: λ arg max λ {d(λ) := L(x (λ), λ)} The function d(λ) is called the dual function. The optimal dual objective value is d = d(λ ). The dual function d(λ) is concave. Hence, we can attempt the following strategy: 1. Find the optimal solution λ of the convex dual problem. 2. Obtain the optimal primal solution x = x (λ ) via the convex primal problem. Challenges for the plausible strategy above 1. Establishing its correctness 2. Computational efficiency of finding an ɛ-approximate optimal dual solution λ ɛ 3. Mapping λ ɛ x ɛ (i.e., ɛ(ɛ)), where ɛ is for the original constrained problem (6) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 11/ 28

15 Finding an optimal solution A plausible strategy: To solve the constrained problem (6), we therefore seek the solutions (x, λ ) arg max min L(x, λ), λ x X which we can naively brake down into two in general nonsmooth problems: Lagrangian subproblem: x (λ) arg min x X {L(x, λ) := f (x) + λ, Ax b } Dual problem: λ arg max λ {d(λ) := L(x (λ), λ)} The function d(λ) is called the dual function. The optimal dual objective value is d = d(λ ). The dual function d(λ) is concave. Hence, we can attempt the following strategy: 1. Find the optimal solution λ of the convex dual problem. 2. Obtain the optimal primal solution x = x (λ ) via the convex primal problem. Challenges for the plausible strategy above 1. Establishing its correctness: Assume f > and Slater s condition for f = d 2. Computational efficiency of finding an ɛ-approximate optimal dual solution λ ɛ 3. Mapping λ ɛ x ɛ (i.e., ɛ(ɛ)), where ɛ is for the original constrained problem (6) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 11/ 28

16 Nesterov s smoothing idea: From O ( ) ( ) 1 ɛ to O 1 ɛ 2 When can the dual function have Lipschitz gradient? When f (x) is γ-strongly convex, the dual function d(λ) is A 2 -Lipschitz gradient. γ (Strong convexity) f (x) is γ-strongly convex iff f (x) γ 2 x 2 2 is convex. d(λ) = min f (x) γ x:x X 2 x 2 2 } {{ } convex & possibly nonsmooth + λ, Ax b + γ 2 x 2 2 } {{ } leads to d F L ( ) AGM automatically obtains d d(x k 1 ) ɛ with k = O ɛ A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 12/ 28

17 Nesterov s smoothing idea: From O ( ) ( ) 1 ɛ to O 1 ɛ 2 When can the dual function have Lipschitz gradient? When f (x) is γ-strongly convex, the dual function d(λ) is A 2 -Lipschitz gradient. γ (Strong convexity) f (x) is γ-strongly convex iff f (x) γ 2 x 2 2 is convex. d(λ) = min f (x) γ x:x X 2 x 2 2 } {{ } convex & possibly nonsmooth + λ, Ax b + γ 2 x 2 2 } {{ } leads to d F L Nesterov s smoother [3] We add a strongly convex term to Lagrange subproblem so that the dual is smooth! d γ(λ) = min f (x) + λ, Ax b + γ x:x X 2 x xc 2 2, with a center point xc X d γ(λ) = Ax γ(λ) b (x γ(λ): the γ-lagrangian subproblem solution) 1. d γ(λ) γd X d(λ) d γ(λ), where D X = max x X 1 2 x xc x k of AGM on d γ(λ) has d d(x k ) γd X + d γ dγ(xk ) γd X + 2 A 2 R 2 γ(k+2) We minimize the upperbound wrt γ and obtain d d(x k ) ɛ with k = O ( 1 ɛ ). A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 12/ 28

18 Computational efficiency: The key role of the prox-operator Smoothed dual: d γ (λ) = min x:x X f (x) + λ, Ax b + γ 2 x x c 2 2 x (λ) = prox f /γ (x c 1 ) γ AT λ Definition (Prox-operator) Key properties: prox g (x) := arg min z R p{g(z) + (1/2) z x 2 }. distributes when the primal problem has decomposable structure: f (x) := i=1 where m 1 is the number of components. m f i (x i ), and X := X 1 X m. often efficient & has closed form expression. For instance, if g(z) = z 1, then the prox-operator performs coordinate-wise soft-thresholding by 1. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 13/ 28

19 Going from the dual ɛ to the primal ɛ I Optimality condition (revisted) Two equivalent ways of viewing the optimality condition of the primal problem (6) mixed variational inequality (MVIP) inclusion { f (x) f (x ) + M(z ) T (z z ) 0, z X R n 0 A T λ + f (x ), = 0 = Ax b. [ ] A where M(z) := T λ and z Ax b := (x, λ ) is a primal-dual solution of (6). A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 14/ 28

20 Going from the dual ɛ to the primal ɛ I Optimality condition (revisted) Two equivalent ways of viewing the optimality condition of the primal problem (6) mixed variational inequality (MVIP) inclusion { f (x) f (x ) + M(z ) T (z z ) 0, z X R n 0 A T λ + f (x ), = 0 = Ax b. [ ] A where M(z) := T λ and z Ax b := (x, λ ) is a primal-dual solution of (6). Measuring progress via the gap function Unfortunately, measuring progress with the inclusion formulation is hard. However, associated with MVIP, we can define a gap function to measure our progress { } G(z) := max f (x) f (ˆx) + M(z) T (z ẑ). (8) ẑ X R n Key observations: G(z) = max f (x) + ˆλ, Ax b min f (ˆx) + λ, Aˆx b 0, z X R n ˆλ R } n ˆx X {{ }} {{ } =f (x) if Ax=b, o/w =d(λ) G(z ) = 0 iff z := (x, λ ) is a primal-dual solution of (6). Primal accuracy ɛ and the dual accuracy ɛ can be related via the gap function. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 14/ 28

21 Going from the dual ɛ to the primal ɛ II A smoothed gap function measuring the excessive primal-dual gap We define a smoothed version of the gap function G γβ (z) = max f (x) + ˆλ, Ax b β ˆλ R n 2 ˆλ ˆλ c 2 2 min f (ˆx) + λ, Aˆx b + γ ˆx X 2 ˆx ˆxc 2 2 } {{ }} {{ } =f β (x)=f (x)+ ˆλ c,ax b + 2β 1 Ax b 2 =d γ (λ) 2 where (ˆx c, ˆλ c) X R n are primal-dual center points. In the sequel, they are 0. The primal accuracy ɛ is related to our primal model estimate f β (x) The dual accuracy ɛ is related to our smoothed dual function d γ(λ) We must relate G γβ (z) to G(z) so that we can tie ɛ to ɛ A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 15/ 28

22 Going from the dual ɛ to the primal ɛ II A smoothed gap function measuring the excessive primal-dual gap We define a smoothed version of the gap function G γβ (z) = max f (x) + ˆλ, Ax b β ˆλ R n 2 ˆλ ˆλ c 2 2 min f (ˆx) + λ, Aˆx b + γ ˆx X 2 ˆx ˆxc 2 2 } {{ }} {{ } =f β (x)=f (x)+ ˆλ c,ax b + 2β 1 Ax b 2 =d γ (λ) 2 where (ˆx c, ˆλ c) X R n are primal-dual center points. In the sequel, they are 0. The primal accuracy ɛ is related to our primal model estimate f β (x) The dual accuracy ɛ is related to our smoothed dual function d γ(λ) We must relate G γβ (z) to G(z) so that we can tie ɛ to ɛ Our algorithm via MEG: model-based excessive gap (cf., [4]) Let G k ( ) := G γk β k ( ). We generate a sequence { z k, γ k, β k } k 0 such that G k+1 ( z k+1 ) (1 τ k )G k ( z k ) + ψ k (MEG) for ψ k 0, rate τ k (0, 1) ( k τ k = ), γ k β k+1 < γ k β k so that G γk β k ( ) G( ). Consequence: G( z k ) 0 + z k z = (x, λ ) (primal-dual solution). A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 15/ 28

23 Going from the dual ɛ to the primal ɛ III An uncertainty relation via MEG The product of the primal and dual convergence rates is lowerbounded by MEG: γ k β k τ 2 k A 2 Note that τ 2 k = Ω ( 1 k 2 ) due to Nesterov s lowerbound. The rate of γ k controls the primal residual: f (x k ) f O (γ k ) The rate of β k controls the feasibility: Ax k b 2 O (β k + τ k ) = O (β k ) They cannot be simultaneously O ( 1 k 2 )! A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 16/ 28

24 Convergence guarantee Theorem [4, 5] 1. When f is strongly convex with µ > 0, we can take γ k = µ and β k = O ( 1 k 2 ) : D Λ Ax k b f (x k ) f 0 Ax k b 4 A 2 (k+2) 2 µ D Λ x k x 4 A (k+2)µ D Λ ( ) ( ) 1 2. When f is non-smooth, the best we can do is γ k = O k and 1 βk = O k : { D Λ Ax k b f (x k ) f 2 2 A D X, k+1 Ax k b 2 2 A (D Λ + D X ) k+1. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 17/ 28

25 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 18/ 28

26 An application: Magnetic Resonance Imaging (MRI) Mixture Model: min x 1 2 Mx y α x TV + µ W x 1 + β W x tree (9) x tree := s x gi 2 (10) With z = Gx we can define the tree norm with non-overlapping groups g i s x tree = (Gx) gi 2 (11) i=1 i=1 x_1 g_1 ğ_1 g_2 x_2 G z = ğ_2 ğ_3 x_3 g_3 x_4 A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 19/ 28

27 Wavelet Tree Sparsity Algorithm (WaTMRI) [1] Mixture Model 1 min x,z 2 Mx y α x TV + µ W x 1 + β s (z) gi 2 + λ 2 z GW x 2 2 (12) i=1 Two subproblems: min z gi β (z) gi 2 + λ 2 z g i (GW x) gi 2 2 is solved by proximity operator. min x 1 2 Mx y λ 2 z GW x α x TV + µ W x 1 is solved by FISTA Proximal operator of α x TV + µ W x 1 is solved with an iterative algorithm Fast empirical convergence but does not allow parallelization No guarantee and does not solve the original problem nor the augmented problem A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 20/ 28

28 Solving with the Primal-Dual Framework [2] Mixture Model min x 1 2 Mx } {{ y } x α x }{{} x 1 TV + µ W x }{{} 1 + β x 2 s i=1 (GW x } {{ } x 3 ) g i 2 (13) f 0 (x 0 ) = x 0 2 2, f 1(x 1 ) = α x 1 TV, f 2 (x 2 ) = µ x 2 1, f 3 (x 3 ) = β s i=1 (x 3) gi 2 Decomposable form where 3 min x=[x T 0,...,x T f (x) := p ]T i=0 f i(x i ) (14) subject to Ax = b W I A = 0 G I 0 and b = 0 (15) M 0 0 I y A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 21/ 28

29 Experimental Setup Original image Subsampling map N = MRI brain image sampled via a partial Fourier operator at a subsampling ratio of 0.2 Note that although we use the same coefficient values for α, β, µ, WaTMRI addresses the augmented problem without the constraint. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 22/ 28

30 Results 10 7 Objective Decopt 1P2D WaTMRI 10 5 Feasibility Decopt 1P2D WaTMRI SNR (db) Decopt 1P2D WaTMRI iteration iteration iteration (a) (b) (c) Figure : MRI experiment. (a) Objective function vs iterations. (b) Feasibility gap z GW x 2 vs iterations. (c) Signal-to-noise ratio of the iterates vs iterations. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 23/ 28

31 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 24/ 28

32 Conclusion Reliable solver for mixture of regularizers Convergence guarantee on both objective and feasibility gap Can handle as many regularizers as we want Requires only proximal operator computations and parallelizable A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 25/ 28

33 References [1] C. Chen and J. Huang. Compressive sensing mri with wavelet tree sparsity. In Advances in neural information processing systems, [2] Baran Gözcü, Luca Baldassarre, Quoc Tran-Dinh, Cosimo Aprile, and Volkan Cevher. A primal-dual framework for mixtures of regularizers [3] Yu. Nesterov. Smooth minimization of non-smooth functions. Math. Program., Ser. A, 103: , [4] Quoc Tran-Dinh and Volkan Cevher. Constrained convex minimization via model-based excessive gap. In Conference of Neural Information Processing Systems (NIPS), [5] Quoc Tran-Dinh and Volkan Cevher. A primal-dual algorithmic framework for constrained convex minimization. Technical report, EPFL, A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 26/ 28

34 A primal-dual framework for mixtures of regularizers Baran Gözcü, Slide 27/ 28

35 1P2D Algorithm Update the primal-dual sequence { z k } We can design different strategies to update {z k }. For instance: ˆλ k := (1 τ k ) λ k + τ k λ β ( x k ) k x k+1 := (1 τ k ) x k + τ k xγ k+1 (ˆλ k ) λ k+1 := ˆλ k + α k (Axγ k+1 (ˆλ k ) b) (1P2D) where α k := γ k+1 A 2 (Bregman), or α k := γ k+1 (augmented Lagrangian). Update parameters The parameters β k and γ k are updated as (c k ( 1, 1] given): γ k+1 := (1 c k τ k )γ k and β k+1 = (1 τ k )β k (16) The parameter τ k is updated as: ( a k+1 := 1 + c k+1 + 4ak 2 + (1 c k+1) 2) /2, and τ k+1 = a 1 k+1. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 28/ 28

Composite Self-concordant Minimization

Composite Self-concordant Minimization Composite Self-concordant Minimization Volkan Cevher Laboratory for Information and Inference Systems-LIONS Ecole Polytechnique Federale de Lausanne (EPFL) volkan.cevher@epfl.ch Paris 6 Dec 11, 2013 joint

More information

A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS. LIONS - EPFL Lausanne - Switzerland

A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS. LIONS - EPFL Lausanne - Switzerland A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS Baran Gözcü, Luca Baldassarre, Quoc Tran-Dinh, Cosimo Aprile and Volkan Cevher LIONS - EPFL Lausanne - Switzerland ABSTRACT Effectively solving many

More information

Convex Optimization and Machine Learning

Convex Optimization and Machine Learning Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

The Benefit of Tree Sparsity in Accelerated MRI

The Benefit of Tree Sparsity in Accelerated MRI The Benefit of Tree Sparsity in Accelerated MRI Chen Chen and Junzhou Huang Department of Computer Science and Engineering, The University of Texas at Arlington, TX, USA 76019 Abstract. The wavelet coefficients

More information

Convex Optimization MLSS 2015

Convex Optimization MLSS 2015 Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :

More information

Lecture 18: March 23

Lecture 18: March 23 0-725/36-725: Convex Optimization Spring 205 Lecturer: Ryan Tibshirani Lecture 8: March 23 Scribes: James Duyck Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal

More information

Efficient MR Image Reconstruction for Compressed MR Imaging

Efficient MR Image Reconstruction for Compressed MR Imaging Efficient MR Image Reconstruction for Compressed MR Imaging Junzhou Huang, Shaoting Zhang, and Dimitris Metaxas Division of Computer and Information Sciences, Rutgers University, NJ, USA 08854 Abstract.

More information

Compressive Sensing MRI with Wavelet Tree Sparsity

Compressive Sensing MRI with Wavelet Tree Sparsity Compressive Sensing MRI with Wavelet Tree Sparsity Chen Chen and Junzhou Huang Department of Computer Science and Engineering University of Texas at Arlington cchen@mavs.uta.edu jzhuang@uta.edu Abstract

More information

Convexity Theory and Gradient Methods

Convexity Theory and Gradient Methods Convexity Theory and Gradient Methods Angelia Nedić angelia@illinois.edu ISE Department and Coordinated Science Laboratory University of Illinois at Urbana-Champaign Outline Convex Functions Optimality

More information

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007 : Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials

More information

Gate Sizing by Lagrangian Relaxation Revisited

Gate Sizing by Lagrangian Relaxation Revisited Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007

More information

Convex optimization algorithms for sparse and low-rank representations

Convex optimization algorithms for sparse and low-rank representations Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,

More information

COMS 4771 Support Vector Machines. Nakul Verma

COMS 4771 Support Vector Machines. Nakul Verma COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex

More information

Characterizing Improving Directions Unconstrained Optimization

Characterizing Improving Directions Unconstrained Optimization Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not

More information

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem

More information

Solution Methods Numerical Algorithms

Solution Methods Numerical Algorithms Solution Methods Numerical Algorithms Evelien van der Hurk DTU Managment Engineering Class Exercises From Last Time 2 DTU Management Engineering 42111: Static and Dynamic Optimization (6) 09/10/2017 Class

More information

Convex Optimization. Lijun Zhang Modification of

Convex Optimization. Lijun Zhang   Modification of Convex Optimization Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Modification of http://stanford.edu/~boyd/cvxbook/bv_cvxslides.pdf Outline Introduction Convex Sets & Functions Convex Optimization

More information

Introduction to Constrained Optimization

Introduction to Constrained Optimization Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR

More information

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction

More information

ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A.

ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A. ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS Donghwan Kim and Jeffrey A. Fessler University of Michigan Dept. of Electrical Engineering and Computer Science

More information

PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization

PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization Marie-Caroline Corbineau 1, Emilie Chouzenoux 1,2, Jean-Christophe Pesquet 1 1 CVN, CentraleSupélec, Université Paris-Saclay,

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider

More information

A proximal center-based decomposition method for multi-agent convex optimization

A proximal center-based decomposition method for multi-agent convex optimization Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 2008 A proximal center-based decomposition method for multi-agent convex optimization Ion Necoara and Johan A.K.

More information

On a first-order primal-dual algorithm

On a first-order primal-dual algorithm On a first-order primal-dual algorithm Thomas Pock 1 and Antonin Chambolle 2 1 Institute for Computer Graphics and Vision, Graz University of Technology, 8010 Graz, Austria 2 Centre de Mathématiques Appliquées,

More information

Parallel and Distributed Sparse Optimization Algorithms

Parallel and Distributed Sparse Optimization Algorithms Parallel and Distributed Sparse Optimization Algorithms Part I Ruoyu Li 1 1 Department of Computer Science and Engineering University of Texas at Arlington March 19, 2015 Ruoyu Li (UTA) Parallel and Distributed

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,

More information

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics 1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!

More information

Applied Lagrange Duality for Constrained Optimization

Applied Lagrange Duality for Constrained Optimization Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity

More information

Templates for Convex Cone Problems with Applications to Sparse Signal Recovery

Templates for Convex Cone Problems with Applications to Sparse Signal Recovery Mathematical Programming Computation manuscript No. (will be inserted by the editor) Templates for Convex Cone Problems with Applications to Sparse Signal Recovery Stephen R. Becer Emmanuel J. Candès Michael

More information

Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging

Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging Paper ID: 1195 No Institute Given 1 More Experiment Results 1.1 Visual Comparisons We apply all methods on four 2D

More information

Augmented Lagrangian Methods

Augmented Lagrangian Methods Augmented Lagrangian Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Augmented Lagrangian IMA, August 2016 1 /

More information

Exploiting Low-Rank Structure in Semidenite Programming by Approximate Operator Splitting

Exploiting Low-Rank Structure in Semidenite Programming by Approximate Operator Splitting Exploiting Low-Rank Structure in Semidenite Programming by Approximate Operator Splitting Mario Souto, Joaquim D. Garcia and Álvaro Veiga June 26, 2018 Outline Introduction Algorithm Exploiting low-rank

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i

More information

Linear methods for supervised learning

Linear methods for supervised learning Linear methods for supervised learning LDA Logistic regression Naïve Bayes PLA Maximum margin hyperplanes Soft-margin hyperplanes Least squares resgression Ridge regression Nonlinear feature maps Sometimes

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

Lecture 7: Support Vector Machine

Lecture 7: Support Vector Machine Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each

More information

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma Support Vector Machines James McInerney Adapted from slides by Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake

More information

Programs. Introduction

Programs. Introduction 16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen

More information

Lecture 2 September 3

Lecture 2 September 3 EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give

More information

CME307/MS&E311 Theory Summary

CME307/MS&E311 Theory Summary CME307/MS&E311 Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/

More information

Optimization III: Constrained Optimization

Optimization III: Constrained Optimization Optimization III: Constrained Optimization CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization III: Constrained Optimization

More information

Theoretical Concepts of Machine Learning

Theoretical Concepts of Machine Learning Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5

More information

Augmented Lagrangian Methods

Augmented Lagrangian Methods Augmented Lagrangian Methods Mário A. T. Figueiredo 1 and Stephen J. Wright 2 1 Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal 2 Computer Sciences Department, University of

More information

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi Lagrangian methods for the regularization of discrete ill-posed problems G. Landi Abstract In many science and engineering applications, the discretization of linear illposed problems gives rise to large

More information

A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem

A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem Presented by: Ted Ralphs Joint work with: Leo Kopman Les Trotter Bill Pulleyblank 1 Outline of Talk Introduction Description

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Constrained Optimization Marc Toussaint U Stuttgart Constrained Optimization General constrained optimization problem: Let R n, f : R n R, g : R n R m, h : R n R l find min

More information

Projection-Based Methods in Optimization

Projection-Based Methods in Optimization Projection-Based Methods in Optimization Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences University of Massachusetts Lowell Lowell, MA

More information

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs 15.082J and 6.855J Lagrangian Relaxation 2 Algorithms Application to LPs 1 The Constrained Shortest Path Problem (1,10) 2 (1,1) 4 (2,3) (1,7) 1 (10,3) (1,2) (10,1) (5,7) 3 (12,3) 5 (2,2) 6 Find the shortest

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods 1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization

More information

Templates for convex cone problems with applications to sparse signal recovery

Templates for convex cone problems with applications to sparse signal recovery Math. Prog. Comp. (2011) 3:165 218 DOI 10.1007/s12532-011-0029-5 FULL LENGTH PAPER Templates for convex cone problems with applications to sparse signal recovery Stephen R. Becker Emmanuel J. Candès Michael

More information

Mathematical Programming and Research Methods (Part II)

Mathematical Programming and Research Methods (Part II) Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types

More information

Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology

Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology Hausdorff Institute for Mathematics (HIM) Trimester: Mathematics of Signal Processing

More information

ML Detection via SDP Relaxation

ML Detection via SDP Relaxation ML Detection via SDP Relaxation Junxiao Song and Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) ELEC 5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture

More information

Optimization under uncertainty: modeling and solution methods

Optimization under uncertainty: modeling and solution methods Optimization under uncertainty: modeling and solution methods Paolo Brandimarte Dipartimento di Scienze Matematiche Politecnico di Torino e-mail: paolo.brandimarte@polito.it URL: http://staff.polito.it/paolo.brandimarte

More information

Lagrangian Relaxation: An overview

Lagrangian Relaxation: An overview Discrete Math for Bioinformatics WS 11/12:, by A. Bockmayr/K. Reinert, 22. Januar 2013, 13:27 4001 Lagrangian Relaxation: An overview Sources for this lecture: D. Bertsimas and J. Tsitsiklis: Introduction

More information

A Short SVM (Support Vector Machine) Tutorial

A Short SVM (Support Vector Machine) Tutorial A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange

More information

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous

More information

Learning with infinitely many features

Learning with infinitely many features Learning with infinitely many features R. Flamary, Joint work with A. Rakotomamonjy F. Yger, M. Volpi, M. Dalla Mura, D. Tuia Laboratoire Lagrange, Université de Nice Sophia Antipolis December 2012 Example

More information

9. Support Vector Machines. The linearly separable case: hard-margin SVMs. The linearly separable case: hard-margin SVMs. Learning objectives

9. Support Vector Machines. The linearly separable case: hard-margin SVMs. The linearly separable case: hard-margin SVMs. Learning objectives Foundations of Machine Learning École Centrale Paris Fall 25 9. Support Vector Machines Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech Learning objectives chloe agathe.azencott@mines

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs

Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs Benjamin Ricaud Signal Processing Lab 2, EPFL Lausanne, Switzerland SPLab, Brno, 10/2013 B. Ricaud

More information

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms

More information

Convex or non-convex: which is better?

Convex or non-convex: which is better? Sparsity Amplified Ivan Selesnick Electrical and Computer Engineering Tandon School of Engineering New York University Brooklyn, New York March 7 / 4 Convex or non-convex: which is better? (for sparse-regularized

More information

IE598 Big Data Optimization Summary Nonconvex Optimization

IE598 Big Data Optimization Summary Nonconvex Optimization IE598 Big Data Optimization Summary Nonconvex Optimization Instructor: Niao He April 16, 2018 1 This Course Big Data Optimization Explore modern optimization theories, algorithms, and big data applications

More information

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report

More information

Collaborative Sparsity and Compressive MRI

Collaborative Sparsity and Compressive MRI Modeling and Computation Seminar February 14, 2013 Table of Contents 1 T2 Estimation 2 Undersampling in MRI 3 Compressed Sensing 4 Model-Based Approach 5 From L1 to L0 6 Spatially Adaptive Sparsity MRI

More information

Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging

Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging Florin C. Ghesu 1, Thomas Köhler 1,2, Sven Haase 1, Joachim Hornegger 1,2 04.09.2014 1 Pattern

More information

Lecture 19: November 5

Lecture 19: November 5 0-725/36-725: Convex Optimization Fall 205 Lecturer: Ryan Tibshirani Lecture 9: November 5 Scribes: Hyun Ah Song Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information

Convex Optimization. Erick Delage, and Ashutosh Saxena. October 20, (a) (b) (c)

Convex Optimization. Erick Delage, and Ashutosh Saxena. October 20, (a) (b) (c) Convex Optimization (for CS229) Erick Delage, and Ashutosh Saxena October 20, 2006 1 Convex Sets Definition: A set G R n is convex if every pair of point (x, y) G, the segment beteen x and y is in A. More

More information

Iteratively Re-weighted Least Squares for Sums of Convex Functions

Iteratively Re-weighted Least Squares for Sums of Convex Functions Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He

More information

Local and Global Minimum

Local and Global Minimum Local and Global Minimum Stationary Point. From elementary calculus, a single variable function has a stationary point at if the derivative vanishes at, i.e., 0. Graphically, the slope of the function

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Maximum Margin Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574

More information

A primal-dual Dikin affine scaling method for symmetric conic optimization

A primal-dual Dikin affine scaling method for symmetric conic optimization A primal-dual Dikin affine scaling method for symmetric conic optimization Ali Mohammad-Nezhad Tamás Terlaky Department of Industrial and Systems Engineering Lehigh University July 15, 2015 A primal-dual

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,

More information

Optimization for Machine Learning

Optimization for Machine Learning with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex

More information

Blind Compressed Sensing Using Sparsifying Transforms

Blind Compressed Sensing Using Sparsifying Transforms Blind Compressed Sensing Using Sparsifying Transforms Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laboratory University of Illinois

More information

A fast weighted stochastic gradient descent algorithm for image reconstruction in 3D computed tomography

A fast weighted stochastic gradient descent algorithm for image reconstruction in 3D computed tomography A fast weighted stochastic gradient descent algorithm for image reconstruction in 3D computed tomography Davood Karimi, Rabab Ward Department of Electrical and Computer Engineering University of British

More information

A More Efficient Approach to Large Scale Matrix Completion Problems

A More Efficient Approach to Large Scale Matrix Completion Problems A More Efficient Approach to Large Scale Matrix Completion Problems Matthew Olson August 25, 2014 Abstract This paper investigates a scalable optimization procedure to the low-rank matrix completion problem

More information

Compressive Sensing Algorithms for Fast and Accurate Imaging

Compressive Sensing Algorithms for Fast and Accurate Imaging Compressive Sensing Algorithms for Fast and Accurate Imaging Wotao Yin Department of Computational and Applied Mathematics, Rice University SCIMM 10 ASU, Tempe, AZ Acknowledgements: results come in part

More information

Fast-Lipschitz Optimization

Fast-Lipschitz Optimization Fast-Lipschitz Optimization DREAM Seminar Series University of California at Berkeley September 11, 2012 Carlo Fischione ACCESS Linnaeus Center, Electrical Engineering KTH Royal Institute of Technology

More information

Recent Developments in Model-based Derivative-free Optimization

Recent Developments in Model-based Derivative-free Optimization Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:

More information

Lecture 4 Duality and Decomposition Techniques

Lecture 4 Duality and Decomposition Techniques Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian

More information

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song

More information

CME307/MS&E311 Optimization Theory Summary

CME307/MS&E311 Optimization Theory Summary CME307/MS&E311 Optimization Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/

More information

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator

Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Volkan Cevher Laboratory for Information and Inference Systems LIONS / EPFL http://lions.epfl.ch & Idiap Research Institute joint

More information

A Multilevel Proximal Gradient Algorithm for a Class of Composite Optimization Problems

A Multilevel Proximal Gradient Algorithm for a Class of Composite Optimization Problems A Multilevel Proximal Gradient Algorithm for a Class of Composite Optimization Problems Panos Parpas May 9, 2017 Abstract Composite optimization models consist of the minimization of the sum of a smooth

More information

Kernels and Constrained Optimization

Kernels and Constrained Optimization Machine Learning 1 WS2014 Module IN2064 Sheet 8 Page 1 Machine Learning Worksheet 8 Kernels and Constrained Optimization 1 Kernelized k-nearest neighbours To classify the point x the k-nearest neighbours

More information

PRIMAL-DUAL SOLUTION PERTURBATIONS IN CONVEX OPTIMIZATION

PRIMAL-DUAL SOLUTION PERTURBATIONS IN CONVEX OPTIMIZATION Set-Valued Analysis (2001) PRIMAL-DUAL SOLUTION PERTURBATIONS IN CONVEX OPTIMIZATION A. L. Dontchev 1 and R. T. Rockafellar 2 3 Abstract. Solutions to optimization problems of convex type are typically

More information

Interior Point I. Lab 21. Introduction

Interior Point I. Lab 21. Introduction Lab 21 Interior Point I Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen the discovery

More information

LECTURE 18 LECTURE OUTLINE

LECTURE 18 LECTURE OUTLINE LECTURE 18 LECTURE OUTLINE Generalized polyhedral approximation methods Combined cutting plane and simplicial decomposition methods Lecture based on the paper D. P. Bertsekas and H. Yu, A Unifying Polyhedral

More information

Convex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21

Convex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21 Convex Programs COMPSCI 371D Machine Learning COMPSCI 371D Machine Learning Convex Programs 1 / 21 Logistic Regression! Support Vector Machines Support Vector Machines (SVMs) and Convex Programs SVMs are

More information

PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY

PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. An algorithm is developed for projecting a point onto a polyhedron. The algorithm solves a dual version

More information

Introduction to optimization

Introduction to optimization Introduction to optimization G. Ferrari Trecate Dipartimento di Ingegneria Industriale e dell Informazione Università degli Studi di Pavia Industrial Automation Ferrari Trecate (DIS) Optimization Industrial

More information

Demo 1: KKT conditions with inequality constraints

Demo 1: KKT conditions with inequality constraints MS-C5 Introduction to Optimization Solutions 9 Ehtamo Demo : KKT conditions with inequality constraints Using the Karush-Kuhn-Tucker conditions, see if the points x (x, x ) (, 4) or x (x, x ) (6, ) are

More information

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1 Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients

More information

x 1 SpaceNet: Multivariate brain decoding and segmentation Elvis DOHMATOB (Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX) 75 x y=20

x 1 SpaceNet: Multivariate brain decoding and segmentation Elvis DOHMATOB (Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX) 75 x y=20 SpaceNet: Multivariate brain decoding and segmentation Elvis DOHMATOB (Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX) L R 75 x 2 38 0 y=20-38 -75 x 1 1 Introducing the model 2 1 Brain decoding

More information