A primal-dual framework for mixtures of regularizers
|
|
- Derrick Wheeler
- 5 years ago
- Views:
Transcription
1 A primal-dual framework for mixtures of regularizers Baran Gözcü Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland EPFL [March 25, 2015] Joint work with Luca Baldassarre, Quoc Tran Dinh, Cosimo Aprile and Volkan LIONS
2 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 2/ 28
3 Overview of compressive imaging System model y = Mx + ω (1) M is the measurement matrix (Fourier, Gaussian etc.) x is the image that is in vectorized form ω is the image that is in vectorized form b is the measurement vector Solution Then we solve min x W x 1 subject to Mx = y (2) where W is the sparsity basis. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 3/ 28
4 Why a mixture model? Signals can posseses various structures at the same time. For example the reconstruction with TV norm x TV = i,j ( (x)) i,j 2 will introduce flat regions. min x subject to x TV Mx = y What happens if we solve with a mixture of regularizers? min x α x TV + (1 α) x 1 subject to Mx = y (3) (4) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 4/ 28
5 Illustration: Mixture model performs better SNR (db) Original (2048x2048,%15 of DCT) Original BP TV TV&BP tv-bp max SNR , A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 5/ 28
6 General Problem min x subject to p f (x) := i=0 f i(x) Ax = b (5) How to solve with computational efficiency guarantee on objective function guarantee on feasibility A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 6/ 28
7 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 7/ 28
8 Swiss army knife of convex formulations Our primal problem prototype: A simple mathematical formulation 1 { } f := min f (x) : Ax = b, x X, (6) x R p f is a proper, closed and convex function, and X is a nonempty, closed convex set. A R n p and b R n are known. An optimal solution x to (6) satisfies f (x ) = f, Ax = b and x X. 1 We can simply replace Ax = b with Ax b C for a convex cone C without any fundamental change. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 8/ 28
9 Swiss army knife of convex formulations Our primal problem prototype: A simple mathematical formulation 1 { } f := min f (x) : Ax = b, x X, (6) x R p f is a proper, closed and convex function, and X is a nonempty, closed convex set. A R n p and b R n are known. An optimal solution x to (6) satisfies f (x ) = f, Ax = b and x X. Example to keep in mind in the sequel { } x := arg min x 1 : Ax = b, x 1 x R p 1 We can simply replace Ax = b with Ax b C for a convex cone C without any fundamental change. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 8/ 28
10 Swiss army knife of convex formulations Our primal problem prototype: A simple mathematical formulation 1 { } f := min f (x) : Ax = b, x X, (6) x R p f is a proper, closed and convex function, and X is a nonempty, closed convex set. A R n p and b R n are known. An optimal solution x to (6) satisfies f (x ) = f, Ax = b and x X. Example to keep in mind in the sequel { } x := arg min x 1 : Ax = b, x 1 x R p Broader context for (6): Standard convex optimization formulations: linear programming, convex quadratic programming, second order cone programming, semidefinite programming and interior point algorithms. Reformulations of existing unconstrained problems via convex splitting: composite convex minimization, consensus optimization,... 1 We can simply replace Ax = b with Ax b C for a convex cone C without any fundamental change. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 8/ 28
11 Numerical ɛ-accuracy Exact vs. approximate solutions Computing an exact solution x to (6) is impracticable unless problem has a closed form solution, which is extremely limited in reality. Numerical optimization algorithms result in x ɛ that approximates x up to a given accuracy ɛ in some sense. In the sequel, by ɛ-accurate solutions x ɛ of (6), we mean the following Definition (ɛ-accurate solutions) Given a numerical tolerance ɛ 0, a point xɛ Rp is called an ɛ-solution of (6) if f (xɛ ) f ɛ (objective residual), Axɛ b ɛ (feasibility gap), xɛ X (exact simple set feasibility). Indeed, ɛ can be different for the objective, feasibility gap, or the iterate residual. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 9/ 28
12 The optimal solution set Before we talk about algorithms, we must first characterize what we are looking for! Optimality condition The optimality condition of min x R p {f (x) : Ax = b} can be written as { 0 A T λ + f (x ), 0 = Ax b. (Subdifferential) f (x) := {v R p : f (y) f (x) + v T (y x), y R p }. This is the well-known KKT (Karush-Kuhn-Tucker) condition. Any point (x, λ ) satisfying (7) is called a KKT point. x is called a stationary point and λ is the corresponding multipliers. Lagrange function and the minimax formulation We can naturally interpret the optimality condition via a minimax formulation max λ min L(x, λ), x dom(f ) where λ R n is the vector of Lagrange multipliers or dual variables w.r.t. Ax = b associated with the Lagrange function: L(x, λ) := f (x) + λ T (Ax b) (7) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 10/ 28
13 Finding an optimal solution A plausible strategy: To solve the constrained problem (6), we therefore seek the solutions (x, λ ) arg max min L(x, λ), λ x X which we can naively brake down into two in general nonsmooth problems: Lagrangian subproblem: x (λ) arg min x X {L(x, λ) := f (x) + λ, Ax b } Dual problem: λ arg max λ {d(λ) := L(x (λ), λ)} The function d(λ) is called the dual function. The optimal dual objective value is d = d(λ ). The dual function d(λ) is concave. Hence, we can attempt the following strategy: 1. Find the optimal solution λ of the convex dual problem. 2. Obtain the optimal primal solution x = x (λ ) via the convex primal problem. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 11/ 28
14 Finding an optimal solution A plausible strategy: To solve the constrained problem (6), we therefore seek the solutions (x, λ ) arg max min L(x, λ), λ x X which we can naively brake down into two in general nonsmooth problems: Lagrangian subproblem: x (λ) arg min x X {L(x, λ) := f (x) + λ, Ax b } Dual problem: λ arg max λ {d(λ) := L(x (λ), λ)} The function d(λ) is called the dual function. The optimal dual objective value is d = d(λ ). The dual function d(λ) is concave. Hence, we can attempt the following strategy: 1. Find the optimal solution λ of the convex dual problem. 2. Obtain the optimal primal solution x = x (λ ) via the convex primal problem. Challenges for the plausible strategy above 1. Establishing its correctness 2. Computational efficiency of finding an ɛ-approximate optimal dual solution λ ɛ 3. Mapping λ ɛ x ɛ (i.e., ɛ(ɛ)), where ɛ is for the original constrained problem (6) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 11/ 28
15 Finding an optimal solution A plausible strategy: To solve the constrained problem (6), we therefore seek the solutions (x, λ ) arg max min L(x, λ), λ x X which we can naively brake down into two in general nonsmooth problems: Lagrangian subproblem: x (λ) arg min x X {L(x, λ) := f (x) + λ, Ax b } Dual problem: λ arg max λ {d(λ) := L(x (λ), λ)} The function d(λ) is called the dual function. The optimal dual objective value is d = d(λ ). The dual function d(λ) is concave. Hence, we can attempt the following strategy: 1. Find the optimal solution λ of the convex dual problem. 2. Obtain the optimal primal solution x = x (λ ) via the convex primal problem. Challenges for the plausible strategy above 1. Establishing its correctness: Assume f > and Slater s condition for f = d 2. Computational efficiency of finding an ɛ-approximate optimal dual solution λ ɛ 3. Mapping λ ɛ x ɛ (i.e., ɛ(ɛ)), where ɛ is for the original constrained problem (6) A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 11/ 28
16 Nesterov s smoothing idea: From O ( ) ( ) 1 ɛ to O 1 ɛ 2 When can the dual function have Lipschitz gradient? When f (x) is γ-strongly convex, the dual function d(λ) is A 2 -Lipschitz gradient. γ (Strong convexity) f (x) is γ-strongly convex iff f (x) γ 2 x 2 2 is convex. d(λ) = min f (x) γ x:x X 2 x 2 2 } {{ } convex & possibly nonsmooth + λ, Ax b + γ 2 x 2 2 } {{ } leads to d F L ( ) AGM automatically obtains d d(x k 1 ) ɛ with k = O ɛ A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 12/ 28
17 Nesterov s smoothing idea: From O ( ) ( ) 1 ɛ to O 1 ɛ 2 When can the dual function have Lipschitz gradient? When f (x) is γ-strongly convex, the dual function d(λ) is A 2 -Lipschitz gradient. γ (Strong convexity) f (x) is γ-strongly convex iff f (x) γ 2 x 2 2 is convex. d(λ) = min f (x) γ x:x X 2 x 2 2 } {{ } convex & possibly nonsmooth + λ, Ax b + γ 2 x 2 2 } {{ } leads to d F L Nesterov s smoother [3] We add a strongly convex term to Lagrange subproblem so that the dual is smooth! d γ(λ) = min f (x) + λ, Ax b + γ x:x X 2 x xc 2 2, with a center point xc X d γ(λ) = Ax γ(λ) b (x γ(λ): the γ-lagrangian subproblem solution) 1. d γ(λ) γd X d(λ) d γ(λ), where D X = max x X 1 2 x xc x k of AGM on d γ(λ) has d d(x k ) γd X + d γ dγ(xk ) γd X + 2 A 2 R 2 γ(k+2) We minimize the upperbound wrt γ and obtain d d(x k ) ɛ with k = O ( 1 ɛ ). A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 12/ 28
18 Computational efficiency: The key role of the prox-operator Smoothed dual: d γ (λ) = min x:x X f (x) + λ, Ax b + γ 2 x x c 2 2 x (λ) = prox f /γ (x c 1 ) γ AT λ Definition (Prox-operator) Key properties: prox g (x) := arg min z R p{g(z) + (1/2) z x 2 }. distributes when the primal problem has decomposable structure: f (x) := i=1 where m 1 is the number of components. m f i (x i ), and X := X 1 X m. often efficient & has closed form expression. For instance, if g(z) = z 1, then the prox-operator performs coordinate-wise soft-thresholding by 1. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 13/ 28
19 Going from the dual ɛ to the primal ɛ I Optimality condition (revisted) Two equivalent ways of viewing the optimality condition of the primal problem (6) mixed variational inequality (MVIP) inclusion { f (x) f (x ) + M(z ) T (z z ) 0, z X R n 0 A T λ + f (x ), = 0 = Ax b. [ ] A where M(z) := T λ and z Ax b := (x, λ ) is a primal-dual solution of (6). A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 14/ 28
20 Going from the dual ɛ to the primal ɛ I Optimality condition (revisted) Two equivalent ways of viewing the optimality condition of the primal problem (6) mixed variational inequality (MVIP) inclusion { f (x) f (x ) + M(z ) T (z z ) 0, z X R n 0 A T λ + f (x ), = 0 = Ax b. [ ] A where M(z) := T λ and z Ax b := (x, λ ) is a primal-dual solution of (6). Measuring progress via the gap function Unfortunately, measuring progress with the inclusion formulation is hard. However, associated with MVIP, we can define a gap function to measure our progress { } G(z) := max f (x) f (ˆx) + M(z) T (z ẑ). (8) ẑ X R n Key observations: G(z) = max f (x) + ˆλ, Ax b min f (ˆx) + λ, Aˆx b 0, z X R n ˆλ R } n ˆx X {{ }} {{ } =f (x) if Ax=b, o/w =d(λ) G(z ) = 0 iff z := (x, λ ) is a primal-dual solution of (6). Primal accuracy ɛ and the dual accuracy ɛ can be related via the gap function. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 14/ 28
21 Going from the dual ɛ to the primal ɛ II A smoothed gap function measuring the excessive primal-dual gap We define a smoothed version of the gap function G γβ (z) = max f (x) + ˆλ, Ax b β ˆλ R n 2 ˆλ ˆλ c 2 2 min f (ˆx) + λ, Aˆx b + γ ˆx X 2 ˆx ˆxc 2 2 } {{ }} {{ } =f β (x)=f (x)+ ˆλ c,ax b + 2β 1 Ax b 2 =d γ (λ) 2 where (ˆx c, ˆλ c) X R n are primal-dual center points. In the sequel, they are 0. The primal accuracy ɛ is related to our primal model estimate f β (x) The dual accuracy ɛ is related to our smoothed dual function d γ(λ) We must relate G γβ (z) to G(z) so that we can tie ɛ to ɛ A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 15/ 28
22 Going from the dual ɛ to the primal ɛ II A smoothed gap function measuring the excessive primal-dual gap We define a smoothed version of the gap function G γβ (z) = max f (x) + ˆλ, Ax b β ˆλ R n 2 ˆλ ˆλ c 2 2 min f (ˆx) + λ, Aˆx b + γ ˆx X 2 ˆx ˆxc 2 2 } {{ }} {{ } =f β (x)=f (x)+ ˆλ c,ax b + 2β 1 Ax b 2 =d γ (λ) 2 where (ˆx c, ˆλ c) X R n are primal-dual center points. In the sequel, they are 0. The primal accuracy ɛ is related to our primal model estimate f β (x) The dual accuracy ɛ is related to our smoothed dual function d γ(λ) We must relate G γβ (z) to G(z) so that we can tie ɛ to ɛ Our algorithm via MEG: model-based excessive gap (cf., [4]) Let G k ( ) := G γk β k ( ). We generate a sequence { z k, γ k, β k } k 0 such that G k+1 ( z k+1 ) (1 τ k )G k ( z k ) + ψ k (MEG) for ψ k 0, rate τ k (0, 1) ( k τ k = ), γ k β k+1 < γ k β k so that G γk β k ( ) G( ). Consequence: G( z k ) 0 + z k z = (x, λ ) (primal-dual solution). A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 15/ 28
23 Going from the dual ɛ to the primal ɛ III An uncertainty relation via MEG The product of the primal and dual convergence rates is lowerbounded by MEG: γ k β k τ 2 k A 2 Note that τ 2 k = Ω ( 1 k 2 ) due to Nesterov s lowerbound. The rate of γ k controls the primal residual: f (x k ) f O (γ k ) The rate of β k controls the feasibility: Ax k b 2 O (β k + τ k ) = O (β k ) They cannot be simultaneously O ( 1 k 2 )! A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 16/ 28
24 Convergence guarantee Theorem [4, 5] 1. When f is strongly convex with µ > 0, we can take γ k = µ and β k = O ( 1 k 2 ) : D Λ Ax k b f (x k ) f 0 Ax k b 4 A 2 (k+2) 2 µ D Λ x k x 4 A (k+2)µ D Λ ( ) ( ) 1 2. When f is non-smooth, the best we can do is γ k = O k and 1 βk = O k : { D Λ Ax k b f (x k ) f 2 2 A D X, k+1 Ax k b 2 2 A (D Λ + D X ) k+1. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 17/ 28
25 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 18/ 28
26 An application: Magnetic Resonance Imaging (MRI) Mixture Model: min x 1 2 Mx y α x TV + µ W x 1 + β W x tree (9) x tree := s x gi 2 (10) With z = Gx we can define the tree norm with non-overlapping groups g i s x tree = (Gx) gi 2 (11) i=1 i=1 x_1 g_1 ğ_1 g_2 x_2 G z = ğ_2 ğ_3 x_3 g_3 x_4 A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 19/ 28
27 Wavelet Tree Sparsity Algorithm (WaTMRI) [1] Mixture Model 1 min x,z 2 Mx y α x TV + µ W x 1 + β s (z) gi 2 + λ 2 z GW x 2 2 (12) i=1 Two subproblems: min z gi β (z) gi 2 + λ 2 z g i (GW x) gi 2 2 is solved by proximity operator. min x 1 2 Mx y λ 2 z GW x α x TV + µ W x 1 is solved by FISTA Proximal operator of α x TV + µ W x 1 is solved with an iterative algorithm Fast empirical convergence but does not allow parallelization No guarantee and does not solve the original problem nor the augmented problem A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 20/ 28
28 Solving with the Primal-Dual Framework [2] Mixture Model min x 1 2 Mx } {{ y } x α x }{{} x 1 TV + µ W x }{{} 1 + β x 2 s i=1 (GW x } {{ } x 3 ) g i 2 (13) f 0 (x 0 ) = x 0 2 2, f 1(x 1 ) = α x 1 TV, f 2 (x 2 ) = µ x 2 1, f 3 (x 3 ) = β s i=1 (x 3) gi 2 Decomposable form where 3 min x=[x T 0,...,x T f (x) := p ]T i=0 f i(x i ) (14) subject to Ax = b W I A = 0 G I 0 and b = 0 (15) M 0 0 I y A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 21/ 28
29 Experimental Setup Original image Subsampling map N = MRI brain image sampled via a partial Fourier operator at a subsampling ratio of 0.2 Note that although we use the same coefficient values for α, β, µ, WaTMRI addresses the augmented problem without the constraint. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 22/ 28
30 Results 10 7 Objective Decopt 1P2D WaTMRI 10 5 Feasibility Decopt 1P2D WaTMRI SNR (db) Decopt 1P2D WaTMRI iteration iteration iteration (a) (b) (c) Figure : MRI experiment. (a) Objective function vs iterations. (b) Feasibility gap z GW x 2 vs iterations. (c) Signal-to-noise ratio of the iterates vs iterations. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 23/ 28
31 Outline Mixture of regularizers Constrained convex minimization: A primal-dual framework Application Conclusion A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 24/ 28
32 Conclusion Reliable solver for mixture of regularizers Convergence guarantee on both objective and feasibility gap Can handle as many regularizers as we want Requires only proximal operator computations and parallelizable A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 25/ 28
33 References [1] C. Chen and J. Huang. Compressive sensing mri with wavelet tree sparsity. In Advances in neural information processing systems, [2] Baran Gözcü, Luca Baldassarre, Quoc Tran-Dinh, Cosimo Aprile, and Volkan Cevher. A primal-dual framework for mixtures of regularizers [3] Yu. Nesterov. Smooth minimization of non-smooth functions. Math. Program., Ser. A, 103: , [4] Quoc Tran-Dinh and Volkan Cevher. Constrained convex minimization via model-based excessive gap. In Conference of Neural Information Processing Systems (NIPS), [5] Quoc Tran-Dinh and Volkan Cevher. A primal-dual algorithmic framework for constrained convex minimization. Technical report, EPFL, A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 26/ 28
34 A primal-dual framework for mixtures of regularizers Baran Gözcü, Slide 27/ 28
35 1P2D Algorithm Update the primal-dual sequence { z k } We can design different strategies to update {z k }. For instance: ˆλ k := (1 τ k ) λ k + τ k λ β ( x k ) k x k+1 := (1 τ k ) x k + τ k xγ k+1 (ˆλ k ) λ k+1 := ˆλ k + α k (Axγ k+1 (ˆλ k ) b) (1P2D) where α k := γ k+1 A 2 (Bregman), or α k := γ k+1 (augmented Lagrangian). Update parameters The parameters β k and γ k are updated as (c k ( 1, 1] given): γ k+1 := (1 c k τ k )γ k and β k+1 = (1 τ k )β k (16) The parameter τ k is updated as: ( a k+1 := 1 + c k+1 + 4ak 2 + (1 c k+1) 2) /2, and τ k+1 = a 1 k+1. A primal-dual framework for mixtures of regularizers Baran Gözcü, baran.goezcue@epfl.ch Slide 28/ 28
Composite Self-concordant Minimization
Composite Self-concordant Minimization Volkan Cevher Laboratory for Information and Inference Systems-LIONS Ecole Polytechnique Federale de Lausanne (EPFL) volkan.cevher@epfl.ch Paris 6 Dec 11, 2013 joint
More informationA PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS. LIONS - EPFL Lausanne - Switzerland
A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS Baran Gözcü, Luca Baldassarre, Quoc Tran-Dinh, Cosimo Aprile and Volkan Cevher LIONS - EPFL Lausanne - Switzerland ABSTRACT Effectively solving many
More informationConvex Optimization and Machine Learning
Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June
More informationThe Benefit of Tree Sparsity in Accelerated MRI
The Benefit of Tree Sparsity in Accelerated MRI Chen Chen and Junzhou Huang Department of Computer Science and Engineering, The University of Texas at Arlington, TX, USA 76019 Abstract. The wavelet coefficients
More informationConvex Optimization MLSS 2015
Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :
More informationLecture 18: March 23
0-725/36-725: Convex Optimization Spring 205 Lecturer: Ryan Tibshirani Lecture 8: March 23 Scribes: James Duyck Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More informationSparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual
Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal
More informationEfficient MR Image Reconstruction for Compressed MR Imaging
Efficient MR Image Reconstruction for Compressed MR Imaging Junzhou Huang, Shaoting Zhang, and Dimitris Metaxas Division of Computer and Information Sciences, Rutgers University, NJ, USA 08854 Abstract.
More informationCompressive Sensing MRI with Wavelet Tree Sparsity
Compressive Sensing MRI with Wavelet Tree Sparsity Chen Chen and Junzhou Huang Department of Computer Science and Engineering University of Texas at Arlington cchen@mavs.uta.edu jzhuang@uta.edu Abstract
More informationConvexity Theory and Gradient Methods
Convexity Theory and Gradient Methods Angelia Nedić angelia@illinois.edu ISE Department and Coordinated Science Laboratory University of Illinois at Urbana-Champaign Outline Convex Functions Optimality
More informationLecture 19: Convex Non-Smooth Optimization. April 2, 2007
: Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials
More informationGate Sizing by Lagrangian Relaxation Revisited
Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007
More informationConvex optimization algorithms for sparse and low-rank representations
Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,
More informationCOMS 4771 Support Vector Machines. Nakul Verma
COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex
More informationCharacterizing Improving Directions Unconstrained Optimization
Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not
More informationPRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction
PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem
More informationSolution Methods Numerical Algorithms
Solution Methods Numerical Algorithms Evelien van der Hurk DTU Managment Engineering Class Exercises From Last Time 2 DTU Management Engineering 42111: Static and Dynamic Optimization (6) 09/10/2017 Class
More informationConvex Optimization. Lijun Zhang Modification of
Convex Optimization Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Modification of http://stanford.edu/~boyd/cvxbook/bv_cvxslides.pdf Outline Introduction Convex Sets & Functions Convex Optimization
More informationIntroduction to Constrained Optimization
Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR
More informationLECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach
LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction
More informationACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A.
ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS Donghwan Kim and Jeffrey A. Fessler University of Michigan Dept. of Electrical Engineering and Computer Science
More informationPIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization
PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization Marie-Caroline Corbineau 1, Emilie Chouzenoux 1,2, Jean-Christophe Pesquet 1 1 CVN, CentraleSupélec, Université Paris-Saclay,
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider
More informationA proximal center-based decomposition method for multi-agent convex optimization
Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 2008 A proximal center-based decomposition method for multi-agent convex optimization Ion Necoara and Johan A.K.
More informationOn a first-order primal-dual algorithm
On a first-order primal-dual algorithm Thomas Pock 1 and Antonin Chambolle 2 1 Institute for Computer Graphics and Vision, Graz University of Technology, 8010 Graz, Austria 2 Centre de Mathématiques Appliquées,
More informationParallel and Distributed Sparse Optimization Algorithms
Parallel and Distributed Sparse Optimization Algorithms Part I Ruoyu Li 1 1 Department of Computer Science and Engineering University of Texas at Arlington March 19, 2015 Ruoyu Li (UTA) Parallel and Distributed
More informationNonlinear Programming
Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,
More information1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics
1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!
More informationApplied Lagrange Duality for Constrained Optimization
Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity
More informationTemplates for Convex Cone Problems with Applications to Sparse Signal Recovery
Mathematical Programming Computation manuscript No. (will be inserted by the editor) Templates for Convex Cone Problems with Applications to Sparse Signal Recovery Stephen R. Becer Emmanuel J. Candès Michael
More informationSupplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging
Supplemental Material for Efficient MR Image Reconstruction for Compressed MR Imaging Paper ID: 1195 No Institute Given 1 More Experiment Results 1.1 Visual Comparisons We apply all methods on four 2D
More informationAugmented Lagrangian Methods
Augmented Lagrangian Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Augmented Lagrangian IMA, August 2016 1 /
More informationExploiting Low-Rank Structure in Semidenite Programming by Approximate Operator Splitting
Exploiting Low-Rank Structure in Semidenite Programming by Approximate Operator Splitting Mario Souto, Joaquim D. Garcia and Álvaro Veiga June 26, 2018 Outline Introduction Algorithm Exploiting low-rank
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i
More informationLinear methods for supervised learning
Linear methods for supervised learning LDA Logistic regression Naïve Bayes PLA Maximum margin hyperplanes Soft-margin hyperplanes Least squares resgression Ridge regression Nonlinear feature maps Sometimes
More informationOptimization for Machine Learning
Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html
More informationLecture 7: Support Vector Machine
Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each
More informationSupport Vector Machines. James McInerney Adapted from slides by Nakul Verma
Support Vector Machines James McInerney Adapted from slides by Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake
More informationPrograms. Introduction
16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen
More informationLecture 2 September 3
EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give
More informationCME307/MS&E311 Theory Summary
CME307/MS&E311 Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/
More informationOptimization III: Constrained Optimization
Optimization III: Constrained Optimization CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization III: Constrained Optimization
More informationTheoretical Concepts of Machine Learning
Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5
More informationAugmented Lagrangian Methods
Augmented Lagrangian Methods Mário A. T. Figueiredo 1 and Stephen J. Wright 2 1 Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal 2 Computer Sciences Department, University of
More informationLagrangian methods for the regularization of discrete ill-posed problems. G. Landi
Lagrangian methods for the regularization of discrete ill-posed problems G. Landi Abstract In many science and engineering applications, the discretization of linear illposed problems gives rise to large
More informationA Generic Separation Algorithm and Its Application to the Vehicle Routing Problem
A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem Presented by: Ted Ralphs Joint work with: Leo Kopman Les Trotter Bill Pulleyblank 1 Outline of Talk Introduction Description
More informationIntroduction to Optimization
Introduction to Optimization Constrained Optimization Marc Toussaint U Stuttgart Constrained Optimization General constrained optimization problem: Let R n, f : R n R, g : R n R m, h : R n R l find min
More informationProjection-Based Methods in Optimization
Projection-Based Methods in Optimization Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences University of Massachusetts Lowell Lowell, MA
More information15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs
15.082J and 6.855J Lagrangian Relaxation 2 Algorithms Application to LPs 1 The Constrained Shortest Path Problem (1,10) 2 (1,1) 4 (2,3) (1,7) 1 (10,3) (1,2) (10,1) (5,7) 3 (12,3) 5 (2,2) 6 Find the shortest
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationUnconstrained Optimization Principles of Unconstrained Optimization Search Methods
1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization
More informationTemplates for convex cone problems with applications to sparse signal recovery
Math. Prog. Comp. (2011) 3:165 218 DOI 10.1007/s12532-011-0029-5 FULL LENGTH PAPER Templates for convex cone problems with applications to sparse signal recovery Stephen R. Becker Emmanuel J. Candès Michael
More informationMathematical Programming and Research Methods (Part II)
Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types
More informationAspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology
Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology Hausdorff Institute for Mathematics (HIM) Trimester: Mathematics of Signal Processing
More informationML Detection via SDP Relaxation
ML Detection via SDP Relaxation Junxiao Song and Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) ELEC 5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture
More informationOptimization under uncertainty: modeling and solution methods
Optimization under uncertainty: modeling and solution methods Paolo Brandimarte Dipartimento di Scienze Matematiche Politecnico di Torino e-mail: paolo.brandimarte@polito.it URL: http://staff.polito.it/paolo.brandimarte
More informationLagrangian Relaxation: An overview
Discrete Math for Bioinformatics WS 11/12:, by A. Bockmayr/K. Reinert, 22. Januar 2013, 13:27 4001 Lagrangian Relaxation: An overview Sources for this lecture: D. Bertsimas and J. Tsitsiklis: Introduction
More informationA Short SVM (Support Vector Machine) Tutorial
A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange
More informationConic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding
Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous
More informationLearning with infinitely many features
Learning with infinitely many features R. Flamary, Joint work with A. Rakotomamonjy F. Yger, M. Volpi, M. Dalla Mura, D. Tuia Laboratoire Lagrange, Université de Nice Sophia Antipolis December 2012 Example
More information9. Support Vector Machines. The linearly separable case: hard-margin SVMs. The linearly separable case: hard-margin SVMs. Learning objectives
Foundations of Machine Learning École Centrale Paris Fall 25 9. Support Vector Machines Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech Learning objectives chloe agathe.azencott@mines
More informationContents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.
page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5
More informationGabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs
Gabor and Wavelet transforms for signals defined on graphs An invitation to signal processing on graphs Benjamin Ricaud Signal Processing Lab 2, EPFL Lausanne, Switzerland SPLab, Brno, 10/2013 B. Ricaud
More informationDavid G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer
David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms
More informationConvex or non-convex: which is better?
Sparsity Amplified Ivan Selesnick Electrical and Computer Engineering Tandon School of Engineering New York University Brooklyn, New York March 7 / 4 Convex or non-convex: which is better? (for sparse-regularized
More informationIE598 Big Data Optimization Summary Nonconvex Optimization
IE598 Big Data Optimization Summary Nonconvex Optimization Instructor: Niao He April 16, 2018 1 This Course Big Data Optimization Explore modern optimization theories, algorithms, and big data applications
More informationDistance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form
Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report
More informationCollaborative Sparsity and Compressive MRI
Modeling and Computation Seminar February 14, 2013 Table of Contents 1 T2 Estimation 2 Undersampling in MRI 3 Compressed Sensing 4 Model-Based Approach 5 From L1 to L0 6 Spatially Adaptive Sparsity MRI
More informationGuided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging
Guided Image Super-Resolution: A New Technique for Photogeometric Super-Resolution in Hybrid 3-D Range Imaging Florin C. Ghesu 1, Thomas Köhler 1,2, Sven Haase 1, Joachim Hornegger 1,2 04.09.2014 1 Pattern
More informationLecture 19: November 5
0-725/36-725: Convex Optimization Fall 205 Lecturer: Ryan Tibshirani Lecture 9: November 5 Scribes: Hyun Ah Song Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More informationConvex Optimization. Erick Delage, and Ashutosh Saxena. October 20, (a) (b) (c)
Convex Optimization (for CS229) Erick Delage, and Ashutosh Saxena October 20, 2006 1 Convex Sets Definition: A set G R n is convex if every pair of point (x, y) G, the segment beteen x and y is in A. More
More informationIteratively Re-weighted Least Squares for Sums of Convex Functions
Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He
More informationLocal and Global Minimum
Local and Global Minimum Stationary Point. From elementary calculus, a single variable function has a stationary point at if the derivative vanishes at, i.e., 0. Graphically, the slope of the function
More informationIntroduction to Machine Learning
Introduction to Machine Learning Maximum Margin Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574
More informationA primal-dual Dikin affine scaling method for symmetric conic optimization
A primal-dual Dikin affine scaling method for symmetric conic optimization Ali Mohammad-Nezhad Tamás Terlaky Department of Industrial and Systems Engineering Lehigh University July 15, 2015 A primal-dual
More informationIntroduction to Optimization
Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,
More informationOptimization for Machine Learning
with a focus on proximal gradient descent algorithm Department of Computer Science and Engineering Outline 1 History & Trends 2 Proximal Gradient Descent 3 Three Applications A Brief History A. Convex
More informationBlind Compressed Sensing Using Sparsifying Transforms
Blind Compressed Sensing Using Sparsifying Transforms Saiprasad Ravishankar and Yoram Bresler Department of Electrical and Computer Engineering and Coordinated Science Laboratory University of Illinois
More informationA fast weighted stochastic gradient descent algorithm for image reconstruction in 3D computed tomography
A fast weighted stochastic gradient descent algorithm for image reconstruction in 3D computed tomography Davood Karimi, Rabab Ward Department of Electrical and Computer Engineering University of British
More informationA More Efficient Approach to Large Scale Matrix Completion Problems
A More Efficient Approach to Large Scale Matrix Completion Problems Matthew Olson August 25, 2014 Abstract This paper investigates a scalable optimization procedure to the low-rank matrix completion problem
More informationCompressive Sensing Algorithms for Fast and Accurate Imaging
Compressive Sensing Algorithms for Fast and Accurate Imaging Wotao Yin Department of Computational and Applied Mathematics, Rice University SCIMM 10 ASU, Tempe, AZ Acknowledgements: results come in part
More informationFast-Lipschitz Optimization
Fast-Lipschitz Optimization DREAM Seminar Series University of California at Berkeley September 11, 2012 Carlo Fischione ACCESS Linnaeus Center, Electrical Engineering KTH Royal Institute of Technology
More informationRecent Developments in Model-based Derivative-free Optimization
Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:
More informationLecture 4 Duality and Decomposition Techniques
Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian
More informationA fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]
A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song
More informationCME307/MS&E311 Optimization Theory Summary
CME307/MS&E311 Optimization Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/
More informationCombinatorial Selection and Least Absolute Shrinkage via The CLASH Operator
Combinatorial Selection and Least Absolute Shrinkage via The CLASH Operator Volkan Cevher Laboratory for Information and Inference Systems LIONS / EPFL http://lions.epfl.ch & Idiap Research Institute joint
More informationA Multilevel Proximal Gradient Algorithm for a Class of Composite Optimization Problems
A Multilevel Proximal Gradient Algorithm for a Class of Composite Optimization Problems Panos Parpas May 9, 2017 Abstract Composite optimization models consist of the minimization of the sum of a smooth
More informationKernels and Constrained Optimization
Machine Learning 1 WS2014 Module IN2064 Sheet 8 Page 1 Machine Learning Worksheet 8 Kernels and Constrained Optimization 1 Kernelized k-nearest neighbours To classify the point x the k-nearest neighbours
More informationPRIMAL-DUAL SOLUTION PERTURBATIONS IN CONVEX OPTIMIZATION
Set-Valued Analysis (2001) PRIMAL-DUAL SOLUTION PERTURBATIONS IN CONVEX OPTIMIZATION A. L. Dontchev 1 and R. T. Rockafellar 2 3 Abstract. Solutions to optimization problems of convex type are typically
More informationInterior Point I. Lab 21. Introduction
Lab 21 Interior Point I Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen the discovery
More informationLECTURE 18 LECTURE OUTLINE
LECTURE 18 LECTURE OUTLINE Generalized polyhedral approximation methods Combined cutting plane and simplicial decomposition methods Lecture based on the paper D. P. Bertsekas and H. Yu, A Unifying Polyhedral
More informationConvex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21
Convex Programs COMPSCI 371D Machine Learning COMPSCI 371D Machine Learning Convex Programs 1 / 21 Logistic Regression! Support Vector Machines Support Vector Machines (SVMs) and Convex Programs SVMs are
More informationPROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY
PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. An algorithm is developed for projecting a point onto a polyhedron. The algorithm solves a dual version
More informationIntroduction to optimization
Introduction to optimization G. Ferrari Trecate Dipartimento di Ingegneria Industriale e dell Informazione Università degli Studi di Pavia Industrial Automation Ferrari Trecate (DIS) Optimization Industrial
More informationDemo 1: KKT conditions with inequality constraints
MS-C5 Introduction to Optimization Solutions 9 Ehtamo Demo : KKT conditions with inequality constraints Using the Karush-Kuhn-Tucker conditions, see if the points x (x, x ) (, 4) or x (x, x ) (6, ) are
More informationSection 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1
Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients
More informationx 1 SpaceNet: Multivariate brain decoding and segmentation Elvis DOHMATOB (Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX) 75 x y=20
SpaceNet: Multivariate brain decoding and segmentation Elvis DOHMATOB (Joint work with: M. EICKENBERG, B. THIRION, & G. VAROQUAUX) L R 75 x 2 38 0 y=20-38 -75 x 1 1 Introducing the model 2 1 Brain decoding
More information